Preview only show first 10 pages with watermark. For full document please download

Version 4.0.1 Reference Guide

   EMBED


Share

Transcript

Reference Guide MapR Administrator Training April 2012 Version 4.0.1 1. Reference Guide . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.1 Documentation for Previous Releases . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.2 Release Notes . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.3 MapR Control System . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.3.1 Cluster Views . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.3.2 MapR-FS Views . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.3.3 NFS HA Views . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.3.4 Alarms Views . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.3.5 System Settings Views . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.3.6 Other Views . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.3.6.1 CLDB View . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.3.6.2 HBase View . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.3.6.2.1 HBase Local Logs View . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.3.6.2.2 HBase Log Level View . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.3.6.2.3 HBase Thread Dump View . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.3.6.3 JobTracker View . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.3.6.3.1 JobTracker Configuration View . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.3.6.4 Nagios View . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.3.7 Node-Related Dialog Boxes . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.4 Hadoop Commands . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.4.1 hadoop archive . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.4.2 hadoop classpath . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.4.3 hadoop daemonlog . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.4.4 hadoop distcp . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.4.5 hadoop fs . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.4.6 hadoop jar . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.4.7 hadoop job . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.4.8 hadoop jobtracker . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.4.9 hadoop mfs . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.4.10 hadoop mradmin . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.4.11 hadoop pipes . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.4.12 hadoop queue . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.4.13 hadoop tasktracker . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.4.14 hadoop version . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.4.15 hadoop conf . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.4.16 Hadoop 2 Commands . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.4.17 Hadoop 1 Commands . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.5 YARN commands . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.5.1 yarn application . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.5.2 yarn classpath . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.5.3 yarn daemonlog . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.5.4 yarn jar . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.5.5 yarn logs . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.5.6 yarn node . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.5.7 yarn rmadmin . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.5.8 yarn version . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.6 API Reference . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.6.1 acl . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.6.1.1 acl edit . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.6.1.2 acl set . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.6.1.3 acl show . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.6.2 alarm . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.6.2.1 alarm clear . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.6.2.2 alarm clearall . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.6.2.3 alarm config load . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.6.2.4 alarm config save . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.6.2.5 alarm list . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.6.2.6 alarm names . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.6.2.7 alarm raise . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.6.3 cluster . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.6.3.1 cluster mapreduce get . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.6.3.2 cluster mapreduce set . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.6.4 config . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.6.4.1 config load . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.6.4.2 config save . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.6.5 dashboard . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.6.5.1 dashboard info . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.6.6 dialhome . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.6.6.1 dialhome ackdial . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.6.6.2 dialhome enable . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.6.6.3 dialhome lastdialed . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.6.6.4 dialhome metrics . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 5 5 5 5 7 32 40 41 44 49 50 51 52 52 53 54 56 61 64 65 68 69 71 72 74 77 78 80 81 83 84 85 85 89 89 90 91 92 92 93 93 94 94 94 95 96 96 99 99 100 102 103 103 104 104 106 107 109 109 110 110 110 111 114 115 116 116 120 120 120 121 122 1.6.6.5 dialhome status . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.6.7 disk . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.6.7.1 disk add . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.6.7.2 disk list . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.6.7.3 disk listall . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.6.7.4 disk remove . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.6.8 dump . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.6.8.1 dump balancerinfo . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.6.8.2 dump balancermetrics . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.6.8.3 dump cldbnodes . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.6.8.4 dump containerinfo . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.6.8.5 dump replicationmanagerinfo . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.6.8.6 dump replicationmanagerqueueinfo . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.6.8.7 dump rereplicationinfo . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.6.8.8 dump rolebalancerinfo . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.6.8.9 dump rolebalancermetrics . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.6.8.10 dump volumeinfo . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.6.8.11 dump volumenodes . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.6.8.12 dump zkinfo . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.6.9 entity . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.6.9.1 entity info . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.6.9.2 entity list . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.6.9.3 entity modify . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.6.10 job . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.6.10.1 job changepriority . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.6.10.2 job kill . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.6.10.3 job linklogs . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.6.10.4 job table . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.6.11 license . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.6.11.1 license add . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.6.11.2 license addcrl . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.6.11.3 license apps . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.6.11.4 license list . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.6.11.5 license listcrl . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.6.11.6 license remove . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.6.11.7 license showid . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.6.12 Metrics API . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.6.13 nagios . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.6.13.1 nagios generate . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.6.14 nfsmgmt . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.6.14.1 nfsmgmt refreshexports . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.6.15 node . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.6.15.1 add-to-cluster . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.6.15.2 node allow-into-cluster . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.6.15.3 node cldbmaster . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.6.15.4 node heatmap . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.6.15.5 node list . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.6.15.6 node listcldbs . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.6.15.7 node listcldbzks . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.6.15.8 node listzookeepers . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.6.15.9 node maintenance . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.6.15.10 node metrics . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.6.15.11 node move . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.6.15.12 node remove . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.6.15.13 node services . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.6.15.14 node topo . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.6.16 rlimit . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.6.16.1 rlimit get . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.6.16.2 rlimit set . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.6.17 schedule . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.6.17.1 schedule create . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.6.17.2 schedule list . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.6.17.3 schedule modify . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.6.17.4 schedule remove . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.6.18 service list . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.6.19 setloglevel . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.6.19.1 setloglevel cldb . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.6.19.2 setloglevel fileserver . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.6.19.3 setloglevel hbmaster . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.6.19.4 setloglevel hbregionserver . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.6.19.5 setloglevel jobtracker . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.6.19.6 setloglevel nfs . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 123 124 124 125 126 127 128 129 132 133 134 137 139 142 144 145 146 148 149 151 151 152 153 154 154 155 156 157 164 164 164 165 165 166 166 166 167 167 167 170 170 170 174 175 175 176 178 180 180 181 182 182 187 187 188 189 190 191 191 192 193 194 195 196 196 197 197 198 198 199 199 200 1.6.19.7 setloglevel tasktracker . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.6.20 table . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.6.20.1 table attr . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.6.20.1.1 table attr edit . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.6.20.1.2 table attr list . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.6.20.2 table cf . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.6.20.2.1 table cf create . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.6.20.2.2 table cf delete . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.6.20.2.3 table cf edit . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.6.20.2.4 table cf list . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.6.20.3 table create . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.6.20.4 table delete . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.6.20.5 table listrecent . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.6.20.6 table region . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.6.20.6.1 table region list . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.6.20.6.2 table region merge . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.6.20.6.3 table region split . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.6.21 task . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.6.21.1 task failattempt . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.6.21.2 task killattempt . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.6.21.3 task table . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.6.22 trace . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.6.22.1 trace dump . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.6.22.2 trace info . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.6.22.3 trace print . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.6.22.4 trace reset . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.6.22.5 trace resize . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.6.22.6 trace setlevel . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.6.22.7 trace setmode . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.6.23 urls . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.6.24 userconfig . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.6.24.1 userconfig load . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.6.25 virtualip . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.6.25.1 virtualip add . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.6.25.2 virtualip edit . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.6.25.3 virtualip list . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.6.25.4 virtualip move . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.6.25.5 virtualip remove . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.6.26 volume . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.6.26.1 volume container move . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.6.26.2 volume create . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.6.26.3 volume dump create . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.6.26.4 volume dump restore . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.6.26.5 volume fixmountpath . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.6.26.6 volume info . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.6.26.7 volume link create . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.6.26.8 volume link remove . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.6.26.9 volume list . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.6.26.10 volume mirror push . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.6.26.11 volume mirror start . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.6.26.12 volume mirror stop . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.6.26.13 volume modify . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.6.26.14 volume mount . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.6.26.15 volume move . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.6.26.16 volume remove . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.6.26.17 volume rename . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.6.26.18 volume showmounts . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.6.26.19 volume snapshot create . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.6.26.20 volume snapshot list . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.6.26.21 volume snapshot preserve . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.6.26.22 volume snapshot remove . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.6.26.23 volume unmount . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.6.27 blacklist user . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.6.27.1 blacklist listusers . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.7 Alarms Reference . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.8 Utilities . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.8.1 configure.sh . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.8.2 disksetup . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.8.3 fsck . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.8.4 gfsck . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.8.5 mapr-support-collect.sh . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.8.6 mapr-support-dump.sh . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 201 201 201 202 203 204 204 205 205 206 207 209 209 210 210 212 212 213 213 213 214 220 220 220 223 223 224 225 225 226 227 227 228 228 229 230 230 231 231 233 233 236 237 239 239 240 241 241 244 245 246 246 248 248 249 249 250 251 251 253 254 255 256 256 257 268 268 272 273 275 278 280 1.8.7 mrconfig . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.8.7.1 mrconfig dg . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.8.7.1.1 mrconfig dg create . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.8.7.1.2 mrconfig dg help . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.8.7.1.3 mrconfig dg list . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.8.7.2 mrconfig info . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.8.7.2.1 mrconfig info containerchain . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.8.7.2.2 mrconfig info containerlist . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.8.7.2.3 mrconfig info containers . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.8.7.2.4 mrconfig info dumpcontainers . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.8.7.2.5 mrconfig info fsstate . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.8.7.2.6 mrconfig info fsthreads . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.8.7.2.7 mrconfig info orphanlist . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.8.7.2.8 mrconfig info replication . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.8.7.2.9 mrconfig info slabs . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.8.7.2.10 mrconfig info threads . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.8.7.2.11 mrconfig info volume snapshot . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.8.7.3 mrconfig sp . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.8.7.3.1 mrconfig sp help . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.8.7.3.2 mrconfig sp list . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.8.7.3.3 mrconfig sp load . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.8.7.3.4 mrconfig sp make . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.8.7.3.5 mrconfig sp offline . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.8.7.3.6 mrconfig sp online . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.8.7.3.7 mrconfig sp refresh . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.8.7.3.8 mrconfig sp shutdown . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.8.7.3.9 mrconfig sp unload . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.8.7.4 mrconfig disk help . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.8.7.5 mrconfig disk init . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.8.7.6 mrconfig disk list . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.8.7.7 mrconfig disk load . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.8.7.8 mrconfig disk remove . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.8.8 pullcentralconfig . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.8.9 rollingupgrade.sh . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.9 Environment Variables . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.10 Configuration Files . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.10.1 .dfs_attributes . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.10.2 cldb.conf . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.10.3 daemon.conf . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.10.4 disktab . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.10.5 hadoop-metrics.properties . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.10.6 mapr-clusters.conf . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.10.7 mfs.conf . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.10.8 taskcontroller.cfg . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.10.9 warden.conf . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.10.10 exports . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.10.11 zoo.cfg . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.10.12 db.conf . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.10.13 warden..conf . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.10.14 mapr.login.conf . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.10.15 core-site.xml (Hadoop 1.x) . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.10.16 mapred-default.xml (MapReduce v1) . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.10.17 mapred-site.xml (MapReduce v1) . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.10.18 yarn-site.xml . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.10.19 yarn-default.xml . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.10.20 mapred-site.xml . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.10.21 mapred-default.xml . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.11 MapR Environment . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.12 MapR Parameters . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.13 Ports Used by MapR . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.14 Best Practices . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.15 Glossary . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 1.16 Source Code for MapR Software . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 281 281 281 283 283 283 283 284 285 286 288 288 289 289 290 291 291 292 292 293 294 294 296 297 297 298 298 299 299 300 301 301 302 302 303 304 304 305 308 308 308 311 312 313 313 316 317 318 319 320 321 325 338 354 356 364 366 380 380 384 389 392 394 Reference Guide The MapR Reference Guide contains in-depth reference information for MapR Software. Choose a subtopic below for more detail. Release Notes - Known issues and new features, by release MapR Control System - User interface reference API Reference - Information about the command-line interface and the REST API Utilities - MapR tool and utility reference Environment Variables - Environment variables specific to MapR Configuration Files - Information about MapR settings Ports Used by MapR - List of network ports used by MapR services Glossary - Essential MapR terms and definitions Hadoop Commands - Listing of Hadoop commands and options YARN Commands - Listing of Hadoop YARN commands and options Documentation for Previous Releases Here are links to documentation for all major releases of MapR software: Version 3.x (PDF here) - Latest Release Version 2.x (PDF here) Version 1.x (PDF here) Release Notes Ecosystem Components Release notes for open source Hadoop components in the MapR Distribution for Hadoop are available here. MapR Distribution for Hadoop Release notes for versions of the MapR Distribution for Hadoop are available here. MapR Control System The MapR Control System main screen consists of a navigation pane to the left and a view to the right. Dialogs appear over the main screen to perform certain actions. View this video for an introduction to the MapR Control System dashboard... Logging on to the MapR Control System 1. In a browser, navigate to the node that is running the mapr-webserver service: https://:8443 2. When prompted, enter the username and password of the administrative user. The Dashboard The Navigation pane to the left lets you choose which view to display on the right. The main view groups are: Cluster Views - information about the nodes in the cluster MapR-FS - information about volumes, snapshots and schedules NFS HA Views - NFS nodes and virtual IP addresses Alarms Views - node and volume alarms System Settings Views - configuration of alarm notifications, quotas, users, groups, SMTP, and HTTP Some other views are separate from the main navigation tree: CLDB View - information about the container location database HBase View - information about HBase on the cluster JobTracker View - information about the JobTracker Nagios View - information about the Nagios configuration script JobHistory Server View - information about MapReduce jobs after their ApplicationMaster terminates Views Views display information about the system. As you open views, tabs along the top let you switch between them quickly. Clicking any column name in a view sorts the data in ascending or descending order by that column. Most views contain a Filter toolbar that lets you sort data in the view, so you can quickly find the information you want. Some views contain collapsible panes that provide different types of detailed information. Each collapsible pane has a control at the top left that expands and collapses the pane. The control changes to show the state of the pane: - pane is collapsed; click to expand - pane is expanded; click to collapse The Filter Toolbar The Filter toolbar lets you build search expressions to provide sophisticated filtering capabilities for locating specific data on views that display a large number of nodes. Expressions are implicitly connected by the AND operator; any search results satisfy the criteria specified in all expressions. The Filter toolbar has two controls: The Minus button ( The Plus button ( ) removes the expression. ) adds a new expression. Expressions Each expression specifies a semantic statement that consists of a field, an operator, and a value. The first dropdown menu specifies the field to match. The second dropdown menu specifies the type of match to perform. The text field specifies a value to match or exclude in the field. You can use a wildcard to substitute for any part of the string. Cluster Views This section provides reference for the following views in the MapR Control System: Dashboard - Summary of cluster health, activity, and usage Cluster Heatmap Alarms Cluster Utilization Yarn Classic MapReduce (v1) Services Volumes Nodes - Summary of node information Overview Services Performance Disks MapReduce NFS Nodes Alarm Status Node Properties View - Details about a node Alarms Machine Performance MapR-FS and Available Disks System Disks Manage Node Services MapReduce DB Gets, Puts, Scans Node Heatmap Jobs The Job Pane The Task Table The Task Attempt Pane Dashboard - Summary of cluster health, activity, and usage The Dashboard displays a summary of information about the cluster in six panes. Panes include: Cluster Heatmap - the alarms and health for each node, by rack Alarms - a summary of alarms for the cluster Cluster Utilization - CPU, Memory, and Disk Space usage Yarn - the number of running and queued applications, number of Node Managers, used memory, total memory, percent of memory used, CPU's used, CPU's total, percent of CPU's used MapReduce - the number of running and queued jobs, running tasks, running map tasks, running reduce tasks, map task capacity, reduce task capacity, map task prefetch capacity, and blacklisted nodes Services - the number of instances of each service Volumes - the number of available, under-replicated, and unavailable volumes Links in each pane provide shortcuts to more detailed information. The following sections provide information about each pane. Cluster Heatmap The Cluster Heatmap pane displays the health of the nodes in the cluster, by rack. Each node appears as a colored square to show its health at a glance. If you click on the small wrench icon at the upper right of the Cluster Heatmap pane, a key to the color-coded heatmap display slides into view. At the top of the display, you can set the refresh rate for the display (measured in seconds), as well as the number of columns to display (for example, 20 nodes are displayed across two rows for a 10-column display). Click the wrench icon again to slide the display back out of view. The left drop-down menu at the top of the pane lets you choose which data is displayed. Some of the choices are shown below. Heatmap legend by category The heatmap legend changes depending on the criteria you select from the drop-down menu. All the criteria and their corresponding legends are shown here. Health Healthy - all services up, MapR-FS and all disks OK, and normal heartbeat Upgrading - upgrade in process Degraded - one or more services down, or no heartbeat for over 1 minute Maintenance - routine maintenance in process Critical - Mapr-FS Inactive/Dead/Replicate, or no heartbeat for over 5 minutes Click to see the legend for all Heatmap displays, such as CPU, memory and disk space... CPU Utilization CPU < 50% CPU < 80% CPU >= 80% Unknown Memory Utilization Memory < 50% Memory < 80% Memory >= 80% Unknown Disk Space Utilization Used < 50% Used < 80% Used >= 80% Unknown Too Many Containers Alarm Containers within limit Containers exceeded limit Duplicate HostId Alarm No duplicate host id detected Duplicate host id detected UID Mismatch Alarm No UID mismatch detected UID mismatch detected No Heartbeat Detected Alarm Node heartbeat detected Node heartbeat not detected TaskTracker Local Dir Full Alarm TaskTracker local directory is not full TaskTracker local directory full PAM Misconfigured Alarm PAM configured PAM misconfigured High FileServer Memory Alarm Fileserver memory OK Fileserver memory high Cores Present Alarm No core files Core files present Installation Directory Full Alarm Installation Directory free Installation Directory full Metrics Write Problem Alarm Metrics writing to Database Metrics unable to write to Database Root Partition Full Alarm Root partition free Root partition full HostStats Down Alarm HostStats running HostStats down Webserver Down Alarm Webserver running Webserver down NFS Gateway Down Alarm NFS Gateway running NFS Gateway down HBase RegionServer Down Alarm HBase RegionServer running HBase RegionServer down HBase Master Down Alarm HBase Master running HBase Master down TaskTracker Down Alarm TaskTracker running TaskTracker down JobTracker Down Alarm JobTracker running JobTracker down FileServer Down Alarm FileServer running FileServer down CLDB Down Alarm CLDB running CLDB down Time Skew Alarm Time OK Time skew alarm(s) Software Installation & Upgrades Alarm Version OK Version alarm(s) Disk Failure(s) Alarm Disks OK Disk alarm(s) Excessive Logging Alarm No debug Debugging Zoomed view You can see a zoomed view of all the nodes in the cluster by moving the zoom slide bar. The zoomed display reveals more details about each node, based on the criteria you chose from the drop-down menu. In this example, CPU Utilization is displayed for each node. Clicking a rack name navigates to the Nodes view, which provides more detailed information about the nodes in the rack. Clicking a colored square navigates to the Node Properties View, which provides detailed information about the node. Alarms The Alarms pane includes these four columns: Alarm - a list of alarms raised on the cluster Last Raised - the most recent time each alarm state changed Summary - how many nodes or volumes have raised each alarm Clear Alarm - clicking on the X clears the corresponding alarm Clicking Alarm, Last Raised, or Summary sorts data in ascending or descending order by that column. Cluster Utilization The Cluster Utilization pane displays a summary of the total usage of the following resources: CPU Memory Disk Space For each resource type, the pane displays the percentage of cluster resources used, the amount used, and the total amount present in the system. A colored dot after the pane's title summarizes the status of the disk and role Configuring Balancer Settings: Green: Both balancers are running. Orange: The replication role balancer is running. Yellow: The disk space balancer is running. Purple: Neither balancer is running. Click the colored dot to bring up the Balancer Configuration dialog. Yarn The Yarn pane shows information about Yarn applications: Running Applications - the number of Yarn applications currently running Queued Applications - the number of Yarn applications queued to run Number of Node Managers - the number of Node Managers (and the number of nodes) in the cluster Used Memory - how much memory has been used to run the applications Total Memory - how much total memory is available for running applications Percent of Memory Used - the percent of memory used compared to the total memory available CPU's Used - the number of CPU cores used CPU's Total - the total number of CPU cores available Percent of CPU's Used - the percent of CPU's used compared to the total number of CPU cores available Classic MapReduce (v1) The Classic MapReduce (v1) pane shows information about MapReduce jobs: Running Jobs - the number of MapReduce jobs currently running Queued Jobs - the number of MapReduce jobs queued to run Running Tasks - the number of Map and Reduce tasks currently running Running Map Tasks - the number of Map tasks currently running Running Reduce Tasks - the number of Reduce tasks currently running Map Task Capacity - the number of map slots available across all nodes in the cluster Reduce Task Capacity - the number of reduce slots available across all nodes in the cluster Map Task Prefetch Capacity - the number of map tasks that can be queued to fill map slots once they become available Blacklisted Nodes - the number of nodes that have been eliminated from the MapReduce pool Services The Services pane shows information about the services running on the cluster. For each service, the pane displays the following information: Actv - the number of running instances of the service Stby - the number of instances of the service that are configured and standing by to provide failover Stop - the number of instances of the service that have been intentionally stopped Fail - the number of instances of the service that have failed, indicated by a corresponsing Service Down alarm Total - the total number of instances of the service configured on the cluster Clicking a service navigates to the Services view. Volumes The Volumes pane displays the total number of volumes, and the number of volumes that are mounted and unmounted. For each category, the Volumes pane displays the number, percent of the total, and total size. Clicking Mounted or Unmounted navigates to the Volumes view. Nodes - Summary of node information The Nodes view displays the nodes in the cluster, by rack. The Nodes view contains two panes: the Topology pane and the Nodes pane. The Topology pane shows the racks in the cluster. Selecting a rack displays that rack's nodes in the Nodes pane to the right. Selecting Cluster displa ys all the nodes in the cluster. Clicking any column name sorts data in ascending or descending order by that column. Selecting the checkbox beside one node makes the following buttons available: Properties - navigates to the Node Properties View, which displays detailed information about a single node. Manage Services - displays the Manage Node Services dialog, which lets you start and stop services on the node. Change Topology - displays the Change Node Topology dialog, which lets you change the topology path for a node. Note: If a node has a No Heartbeat alarm raised, the Forget Node button is also displayed. When you click on Forget Node, the following Message appears: When you click on Manage Services, a dialog is displayed where you can stop, start, or restart the services on the node. When you click on Change Topology, a dialog is displayed where you can choose a different location for the selected node. Selecting the checkboxes beside multiple nodes changes the text on the buttons to reflect the number of nodes affected: The dropdown menu at the top left specifies the type of information to display: Overview - general information about each node Services - services running on each node Performance - information about memory, CPU, I/O and RPC performance on each node Disks - information about disk usage, failed disks, and the MapR-FS heartbeat from each node MapReduce - information about the JobTracker heartbeat and TaskTracker slots on each node NFS Nodes - the IP addresses and Virtual IPs assigned to each NFS node Alarm Status - the status of alarms on each node Clicking a node's Hostname navigates to the Node Properties View, which provides detailed information about the node. Selecting the Filter checkbox displays the Filter toolbar, which provides additional data filtering options. Each time you select a filtering option, the option is displayed in the window below the filter checkbox. You can add more options by clicking on the . Overview The Overview displays the following general information about nodes in the cluster: Hlth - each node's health: healthy, degraded, critical, or maintenance Hostname - the hostname of each node Physical IP(s) - the IP address or addresses associated with each node FS HB - time since each node's last heartbeat to the CLDB Physical Topology - the rack path to each node Services The Services view displays the following information about nodes in the cluster: Hlth - eact node's health: healthy, degraded, critical, or maintenance Hostname - the hostname of each node Configured Services - a list of the services specified in the config file Running Services - a list of the services running on each node Physical Topology - each node's physical topology Performance The Performance view displays the following information about nodes in the cluster, including: Hlth - each node's health: healthy, degraded, critical, or maintenance Hostname - DNS hostname for the nodes in this cluster Memory - percentage of memory used and the total memory % CPU - percentage of CPU usage on the node # CPUs - number of CPUs present on the node Bytes Received - number of bytes received in 1 second, through all network interfaces on the node Click to see all Performance metrics... Bytes Sent - number of bytes sent in 1 second, through all network interfaces on the node # RPCs - number of RPC calls RPC In Bytes - number of RPC bytes received by this node every second RPC Out Bytes - number of RPC bytes sent by this node every second # Disk Reads - number of disk read operations on this node every second # Disk Writes - number of disk write operations on this node every second Disk Read Bytes - number of bytes read from all the disks on this node every second Disk Write Bytes - number of bytes written to all the disks on this node every second # Disks - number of disks on this node Gets - 1m - number of data retrievals (gets) executed on this region's primary node in a 1-minute interval Puts - 1m - number of data writes (puts) executed on this region's primary node in a 1-minute interval Scans - 1m - number of data seeks (scans) executed on this region's primary node in a 1-minute interval Disks The Disks view displays the following information about nodes in the cluster: Hlth - each node's health: healthy, degraded, or critical Hostname - the hostname of each node # Bad Disks - the number of failed disks on each node Disk Space - the amount of disk used and total disk capacity, in gigabytes MapReduce The MapReduce view displays the following information about nodes in the cluster: Hlth - each node's health: healthy, degraded, or critical Hostname - the hostname of each node TT Map Slots - the number of map slots on each node TT Map Slots Used - the number of map slots in use on each node TT Reduce Slots - the number of reduce slots on each node TT Reduce Slots Used - the number of reduce slots in use on each node NFS Nodes The NFS Nodes view displays the following information about nodes in the cluster: Hlth - each node's health: healthy, degraded, or critical Hostname - the hostname of each node Physical IP(s) - the IP address or addresses associated with each node Virtual IP(s) - the virtual IP address or addresses assigned to each node Alarm Status The Alarm Status view displays the following information about nodes in the cluster: Hlth - each node's health: healthy, degraded, critical, or maintenance Hostname - DNS hostname for nodes in this cluster Version Alarm - one or more services on the node are running an unexpected version No Heartbeat Alarm - node is not undergoing maintenance, and no heartbeat is detected for over 5 minutes UID Mismatch Alarm - services in the cluster are being run with different user names (UIDs) Duplicate HostId Alarm - two or more nodes in the cluster have the same host id Click to see all Alarm Status alerts... Too Many Containers Alarm - number of containers on this node reached the maximum limit Excess Logs Alarm - debug logging is enabled on the node (debug logging generates enormous amounts of data and can fill up disk space) Disk Failure Alarm - a disk has failed on the node Time Skew Alarm - the clock on the node is out of sync with the master CLDB by more than 20 seconds Root Partition Full Alarm - the root partition ("/") on the node is running out of space (99% full) Installation Directory Full Alarm - the partition /opt/mapr on the node is running out of space (95% full) Core Present Alarm - a service on the node has crashed and created a core dump file High FileServer Memory Alarm - memory consumed by fileserver service on the node is high Pam Misconfigured Alarm - the PAM authentication on the node is configured incorrectly TaskTracker Local Directory Full Alarm - the local directory used by the TaskTracker on the specified node(s) is full, and the TaskTracker cannot operate as a result CLDB Alarm - the CLDB service on the node has stopped running FileServer Alarm - the FileServer service on the node has stopped running JobTracker Alarm - the JobTracker service on the node has stopped running TaskTracker Alarm - the TaskTracker service on the node has stopped running HBase Master Alarm - the HBase Master service on the node has stopped running HBase RegionServer Alarm - the HBase RegionServer service on the node has stopped running NFS Gateway Alarm - the NFS service on the node has stopped running WebServer Alarm - the WebServer service on the node has stopped running HostStats Alarm - the HostStats service has stopped running Metrics write problem Alarm - metric data was not written to the database Node Properties View - Details about a node The Node Properties view displays detailed information about a single node in seven collapsible panes: Alarms Machine Performance MapR-FS and Available Disks System Disks Manage Node Services MapReduce DB Gets, Puts, Scans Buttons: Forget Node - displays the Forget Node dialog box Alarms The Alarms pane displays a list of alarms that have been raised on the system, and the following information about each alarm: Alarm - the alarm name Last Raised - the most recent time when the alarm was raised Summary - a description of the alarm Clear Alarm - clicking on the X clears the corresponding alarm Machine Performance The Machine Performance pane displays the following information about the node's performance and resource usage since it last reported to the CLDB: Memory Used - the amount of memory in use on the node Disk Used - the amount of disk space used on the node CPU - The number of CPUs and the percentage of CPU used on the node Network I/O - the input and output to the node per second RPC I/O - the number of RPC calls on the node and the amount of RPC input and output Disk I/O - the amount of data read to and written from the disk # Operations - the number of disk reads and writes MapR-FS and Available Disks The MapR-FS and Available Disks pane displays the disks on the node and information about each disk. Information headings include: Status - the status of the disk (healthy, failed, or offline) Mount - whether the disk is mounted (indicated by ) or unmounted Device - the device name File System - the file system on the disk Used - the percentage of memory used out of total memory available on the disk Model # - the model number of the disk Serial # - the serial number of the disk Firmware Version - the version of the firmware being used Add to MAPR-FS - clicking the Remove from MAPR-FS - clicking the adds the disk to MAPR-FS storage displays a dialog that asks you to verify that you want to remove the disk If you confirm by clicking OK, and data on that disk has not been replicated, a warning dialog appears: For more information on disk status, and the proper procedure for adding, removing, and replacing disks, see the Managing Disks page. If you are running MapR 1.2.2 or earlier, do not use the disk add command or the MapR Control System to add disks to MapR-FS. You must either upgrade to MapR 1.2.3 before adding or replacing a disk, or use the following procedure (which avoids the disk add comma nd): 1. Use the MapR Control System to remove the failed disk. All other disks in the same storage pool are removed at the same time. Make a note of which disks have been removed. 2. Create a text file /tmp/disks.txt containing a list of the disks you just removed. See Setting Up Disks for MapR. 3. Add the disks to MapR-FS by typing the following command (as root or with sudo): /opt/mapr/server/disksetup -F /tmp/disks.txt System Disks The System Disks pane displays information about disks present and mounted on the node: Status - the status of the disk (healthy, failed, or offline) Mount - whether the disk is mounted (indicated by ) or unmounted Device - the device name File System - the file system on the disk Used - the percentage of memory used out of total memory available on the disk Model # - the model number of the disk Serial # - the serial number of the disk Firmware Version - the version of the firmware being used Manage Node Services The Manage Node Services pane displays the status of each service on the node. Service - the name of each service State: Configured: the package for the service is installed and the service is configured for all nodes, but it is not enabled for the particular node Not Configured: the package for the service is not installed and/or the service is not configured (configure.sh has not run) Running: the service is installed, has been started by the warden, and is currently executing Stopped: the service is installed and configure.sh has run, but the service is currently not executing StandBy: the service is installed Failed: the service was running, but terminated unexpectedly Log Path - the path to where each service stores its logs Stop/Start: click on to stop the service click on to start the service Restart - click on to restart the service Log Settings - displays the Trace Activity dialog where you can set the level of logging for a service on a particular node. When you select a log level, all the levels listed above it are included in the log. Levels include: ERROR WARN INFO DEBUG TRACE You can also start and stop services in the the Manage Node Services dialog, by clicking Manage Services in the Nodes view. MapReduce The MapReduce pane displays the number of map and reduce slots used, and the total number of map and reduce slots on the node. DB Gets, Puts, Scans The DB Gets, Puts, Scans pane displays the number of gets, puts, and scan operations performed during various time intervals. Node Heatmap The Node Heatmap view provides a graphical summary of node status across the cluster. This view displays the same information as the Node Heatmap pane on the Dashboard, without the other panes that appear on the dashboard. Jobs The Jobs view displays the data collected by the MapR Metrics service. The Jobs view contains two panes: the chart pane and the data grid. The chart pane displays the data corresponding to the selected metric in histogram form. The data grid lists the jobs running on the cluster. Click on the wrench icon to slide out a menu of information to display. Choices include: Cumulative Job Combine Input Records Cumulative Job Map Input Bytes Cumulative Job Map Input Records Cumulative Job Map Output Bytes Cumulative Job Map Output Records Cumulative Job Reduce Input Records Click to see all Job metrics... Cumulative Job Reduce Output Bytes Cumulative Job Reduce Shuffle Bytes Cumulative Physical Memory Current CPU Current Memory Job Average Map Attempt Duration Job Average Reduce Attempt Duration Job Average Task Duration Job Combine Output Records Job Complete Map Task Count Job Complete Reduce Task Count Job Complete Task Count Job Cumulative CPU Job Data-local Map Tasks Job Duration Job End Time Job Error Count Job Failed Map Task Attempt Count Job Failed Map Task Count Job Failed Reduce Task Attempt Count Job Failed Reduce Task Count Job Failed Task Attempt Count Job Failed Task Count Job Id Job Map CPU Job Map Cumulative Memory Bytes Job Map File Bytes Written Job Map GC Time Job Map Input Bytes/Sec Job Map Input Records/Sec Job Map Output Bytes/Sec Job Map Output Records/Sec Job Map Progress Job Map Reserve Slot Wait Job Map Spilled Records Job Map Split Raw Bytes Job Map Task Attempt Count Job Map Task Count Job Map Tasks Duration Job Map Virtual Memory Bytes Job MapR-FS Map Bytes Read Job MapR-FS Map Bytes Written Job MapR-FS Reduce Bytes Read Job MapR-FS Reduce Bytes Written Job MapR-FS Total Bytes Read Job MapR-FS Total Bytes Written Job Maximum Map Attempt Duration Job Maximum Reduce Attempt Duration Job Maximum Task Duration Job Name Job Non-local Map Tasks Job Rack-local Map Tasks Job Reduce CPU Job Reduce Cumulative Memory Bytes Job Reduce File Bytes Written Job Reduce GC Time Job Reduce Input Groups Job Reduce Input Records/Sec Job Reduce Output Records/Sec Job Reduce Progress Job Reduce Reserve Slot Wait Job Reduce Shuffle Bytes/Sec Job Reduce Spilled Records Job Reduce Split Raw Bytes Job Reduce Task Attempt Count Job Reduce Task Count Job Reduce Tasks Duration Job Reduce Virtual Memory Bytes Job Running Map Task Count Job Running Reduce Task Count Job Running Task Count Job Split Raw Bytes Job Start Time Job Submit Time Job Task Attempt Count Job Total File Bytes Written Job Total GC Time Job Total Spilled Records Job Total Task Count Job User Logs Map Tasks Finish Time Map Tasks Start Time Priority Reduce Tasks Finish Time Reduce Tasks Start Time Status Virtual Memory Bytes Select the Filter checkbox to display the Filter toolbar, which provides additional data filtering options. The x-axis: drop-down selector lets you change the display scale of the histogram's X axis between a uniform or logarithmic scale. Hover the cursor over a bar in the histogram to display the Filter and Zoom buttons. Click the Filter button or click the bar to filter the table below the histogram by the data range corresponding to that bar. The selected bar turns yellow. Hover the cursor over the selected bar to display the Clear Filter and Zoom buttons. Click the Clear Filter button to remove the filter from the data range in the table below the histogram. Double-click a bar or click the Zoom button to zoom in and display a new histogram that displays metrics constrained to the data range represented by the bar. The data range applied to the metrics data set displays above the histogram. Click the plus or minus buttons in the filter conditions panel to add or remove filter conditions. Uncheck the Filter checkbox above the histogram to clear the entire filter. Check the box next to a job in the table below the histogram to enable the View Job button. If the job is still running, checking this box also enables the Kill Job button. Clicking Kill Job will display a confirmation dialog to choose whether or not to terminate the job. Click the View Job button or click the job name in the table below the histogram to open the Job tab for that job. The Job Pane From the main Jobs page, select a job from the list below the histogram and click View Job. You can also click directly on the name of the job in the list. The Job Properties pane displays with the Tasks tab selected by default. This pane has three tabs, Tasks, Charts, and Info. If the job is running, the Kill Job button is enabled. The Tasks Tab The Tasks tab has two panes. The upper pane displays histograms of metrics for the tasks and task attempts in the selected job. The lower pane displays a table that lists the tasks and primary task attempts in the selected job. Tasks can be in any of the following states: COMPLETE FAILED KILLED PENDING RUNNING The table of tasks also lists the following information for each task: Task ID. Click the link to display a table with information about the task attempts for this task. Task type: M: Map R: Reduce TC: Task Cleanup JS: Job Setup JC: Job Cleanup Primary task attempt ID. Click the link to display the task attempt pane for this task attempt. Task starting timestamp Task ending timestamp Task duration Host locality Node running the task. Click the link to display the Node Properties pane for this node. You can select the following task histogram metrics for this job from the drop-down selector: Task Duration Task Attempt Duration Task Attempt Local Bytes Read Task Attempt Local Bytes Written Task Attempt MapR-FS Bytes Read Click to see all Task metrics... Task Attempt MapR-FS Bytes Written Task Attempt Garbage Collection Time Task Attempt CPU Time Task Attempt Physical Memory Bytes Task Attempt Virtual Memory Bytes Map Task Attempt Input Records Map Task Attempt Output Records Map Task Attempt Skipped Records Map Task Attempt Input Bytes Map Task Attempt Output Bytes Reduce Task Attempt Input Groups Reduce Task Attempt Shuffle Bytes Reduce Task Attempt Input Records Reduce Task Attempt Output Records Reduce Task Attempt Skipped Records Task Attempt Spilled Records Combined Task Attempt Input Records Combined Task Attempt Output Records Uncheck the Show Map Tasks box to hide map tasks. Uncheck the Show Reduce Tasks box to hide reduce tasks. Check the Show Setup/Cleanup Tasks box to display job and task setup and cleanup tasks. Histogram filtering and zoom work in the same way as the Jobs pane . The Charts Tab Click the Charts tab to display your job's line chart metrics. Click the Add chart button to add a new line chart. You can use the X and minus buttons at the top-left of each chart to dismiss or hide the chart. Line charts can display the following metrics for your job: Cumulative CPU used Cumulative physical memory used Number of failed map tasks Number of failed reduce tasks Number of running map tasks Click to see all available Chart metrics... Number of running reduce tasks Number of map task attempts Number of failed map task attempts Number of failed reduce task attempts Rate of map record input Rate of map record output Rate of map input bytes Rate of map output bytes Rate of reduce record output Rate of reduce shuffle bytes Average duration of map attempts Average duration of reduce attempts Maximum duration of map attempts Maximum duration of reduce attempts The Information Tab The Information tab of the Job Properties pane displays summary information about the job in three collapsible panes: The MapReduce Framework Counters pane displays information about this job's MapReduce activity. The Job Counters pane displays information about the number of this job's map tasks. The File System Counters pane displays information about this job's interactions with the cluster's file system. The Task Table The Task table displays a list of the task attempts for the selected task, along with the following information for each task attempt: Status: RUNNING SUCCEEDED FAILED UNASSIGNED KILLED COMMIT PENDING FAILED UNCLEAN KILLED UNCLEAN Task attempt ID. Click the link to display the task attempt pane for this task attempt. Task attempt type: M: Map R: Reduce TC: Task Cleanup JS: Job Setup JC: Job Cleanup Task attempt starting timestamp Task attempt ending timestamp Task attempt shuffle ending timestamp Task attempt sort ending timestamp Task attempt duration Node running the task attempt. Click the link to display the Node Properties pane for this node. A link to the log file for this task attempt Diagnostic information about this task attempt The Task Attempt Pane The Task Attempt pane has two tabs, Info and Charts. The Task Attempt Info Tab The Info tab displays summary information about this task attempt in three panes: The MapReduce Framework Counters pane displays information about this task attempt's MapReduce activity. The MapReduce Throughput Counters pane displays information about the I/O performance in Bytes/sec and Records/sec. The File System Counters pane displays information about this task attempt's interactions with the cluster's file system. The Task Attempt Charts Tab The Task Attempt Charts tab displays line charts for metrics specific to this task attempt. By default, this tab displays charts for these metrics: Cumulative CPU by Time Physical Memory by Time Virtual Memory by Time Click the Add chart button to add a new line chart. You can use the X and minus buttons at the top-left of each chart to dismiss or hide the chart. Line charts can display the following metrics for your task: Combine Task Attempt Input Records Combine Task Attempt Output Records Map Task Attempt Input Bytes Map Task Attempt Input Records Map Task Attempt Output Bytes Click to see all available Task Attempt metrics... Map Task Attempt Output Records Map Task Attempt Skipped Records Reduce Task Attempt Input Groups Reduce Task Attempt Input Records Reduce Task Attempt Output Records Reduce Task Attempt Shuffle Bytes Reduce Task Attempt Skipped Records Task Attempt CPU Time Task Attempt Local Bytes Read Task Attempt Local Bytes Written Task Attempt MapR-FS Bytes Read Task Attempt MapR-FS Bytes Written Task Attempt Physical Memory Bytes Task Attempt Spilled Records Task Attempt Virtual Memory Bytes MapR-FS Views The MapR-FS group provides the following views: Tables - information about M7 tables in the cluster Volumes - information about volumes in the cluster Mirror Volumes - information about mirrors User Disk Usage - cluster disk usage Snapshots - information about volume snapshots Schedules - information about schedules Tables The Tables view displays a list of tables in the cluster. The New Table button displays a field where you can enter the path to a new table to create from the MCS. Click the name of a table from the Tables view to display the table detail view. From the table detail view, click Delete Table to delete this table. The table detail view has the following tabs: Column Families Regions The Column Families tab displays the following information: Column Family Name Max Versions Min Versions Compression Time-to-Live In Memory Click the Edit Column Family button to change these values. Click the Delete Column Family button to delete the selected column families. The Regions tab displays the following information: Start Key - The first key in the region range. End Key - The last key in the region range. Physical Size - The physical size of the region with compression. Logical Size - The logical size of the region without compression. # Rows - The number of rows stored in the region. Primary Node - The region's original source for storage and computation. Secondary Nodes - The region's replicated sources for storage and computation. Last HB - The time interval since the last data communication with the region's primary node. Region Identifier - The tablet region identifier. Volumes The Volumes view displays the following information about volumes in the cluster: Mnt - Whether the volume is mounted. ( ) Vol Name - The name of the volume. Mount Path - The path where the volume is mounted. Creator - The user or group that owns the volume. Quota - The volume quota. Vol Size - The size of the volume. Data Size - The size of the volume on the disk before compression. Snap Size - The size of the all snapshots for the volume. As the differences between the snapshot and the current state of the volume grow, the amount of data storage taken up by the snapshots increases. Total Size - The size of the volume and all its snapshots. Replication Factor - The number of copies of the volume. Physical Topology - The rack path to the volume. Clicking any column name sorts data in ascending or descending order by that column. The Unmounted checkbox specifies whether to show unmounted volumes: selected - show both mounted and unmounted volumes unselected - show mounted volumes only The System checkbox specifies whether to show system volumes: selected - show both system and user volumes unselected - show user volumes only Selecting the Filter checkbox displays the Filter toolbar, which provides additional data filtering options. Clicking New Volume displays the New Volume dialog. New Volume The New Volume dialog lets you create a new volume. For mirror volumes, the Snapshot Scheduling section is replaced with a section called Mirror Scheduling: The Volume Setup section specifies basic information about the volume using the following fields: Volume Type - a standard volume, or a local or remote mirror volume Volume Name (required) - a name for the new volume Mount Path - a path on which to mount the volume (check the small box at the right to indicate the mount path for the new volume; if the box is not checked, an unmounted volume is created) Topology - the new volume's rack topology Read-only - if checked, prevents writes to the volume The Permissions section lets you grant specific permissions on the volume to certain users or groups: User/Group field - the user or group to which permissions are to be granted (one user or group per row) Permissions field - the permissions to grant to the user or group (see the Permissions table below) Delete button ( ) - deletes the current row [+ Add Permission ] - adds a new row Volume Permissions Code Allowed Action dump Dump/Back up the volume restore Restore/Mirror the volume m Edit volume properties d Delete the volume fc Full control (admin access and permission to change volume ACL) The Usage Tracking section displays cluster usage and sets quotas for the volume using the following fields: Quotas - the volume quotas: Volume Advisory Quota - if selected, enter the advisory quota for the volume expressed as an integer plus the single letter abbreviation for the unit (such as 100G for 100GB). When this quota is reached, an advisory email is sent to the user or group. Volume Hard Quota - if selected, enter the maximum limit for the volume expressed as an integer plus the single letter abbreviation for the unit (such as 128G for 128GB). When this hard limit is reached, no more data is written to the volume. The Replication section contains the following fields: Replication - the requested replication factor for the volume Min Replication - the minimum replication factor for the volume. When the number of replicas drops down to or below this number, the volume is aggressively re-replicated to bring it above the minimum replication factor. Optimize Replication For - the basis for choosing the optimum replication factor (high throughput or low latency) The Snapshot Scheduling section (normal volumes) contains the snapshot schedule, which determines when snapshots will be automatically created. Select an existing schedule from the pop-up menu. The Mirror Scheduling section (local and remote mirror volumes) contains the mirror schedule, which determines when mirror volumes will be automatically created. Select an existing schedule from the pop-up menu. Buttons: OK - creates the new volume Cancel - exits without creating the volume Volume Actions You can modify a volume's properties by selecting the checkbox next to that volume and clicking the Volume Actions button. A dropdown menu of properties you can modify displays. To apply one set of changes to multiple volumes, mark the checkboxes next to each volume. Properties Clicking on a volume name displays the Volume Properties dialog where you can view information about the volume, and check or change various settings. You can also remove the volume. If you click on Remove Volume, the following dialog appears: Buttons: OK - removes the volume or volumes Cancel - exits without removing the volume or volumes For information about the fields in the Volume Properties dialog, see New Volume. The Partly Out of Topology checkbox fills when containers in this volume are outside the volume's main topology. Snapshots The Snapshots dialog displays the following information about snapshots for the specified volume: Snapshot Name - The name of the snapshot. Disk Used - The total amount of logical storage held by the snapshot. Since the current volume and all of its snapshots will often have storage held in common, the total disk usage reported will often exceed the total storage used by the volume. The value reported in this field is the size the snapshot would have if the difference between the snapshot and the volume's current state is 100%. Created - The date and time the snapshot was created. Expires - The snapshot expiration date and time. Buttons: New Snapshot - Displays the Snapshot Name dialog. Remove - When the checkboxes beside one or more snapshots are selected, displays the Remove Snapshots dialog. Preserve - When the checkboxes beside one or more snapshots are selected, prevents the snapshots from expiring. Close - Closes the dialog. New Snapshot The Create New Snapshot dialog lets you specify the name for a new snapshot you are creating. The Snapshot Name dialog creates a new snapshot with the name specified in the following field: Name For New Snapshot(s) - the new snapshot name Buttons: OK - creates a snapshot with the specified name Cancel - exits without creating a snapshot Remove Snapshots The Remove Snapshots dialog prompts you for confirmation before removing the specified snapshot or snapshots. Buttons Yes - removes the snapshot or snapshots No - exits without removing the snapshot or snapshots Assign Cluster-Wide Provisionary Space Limit You can modify standard volume reserve limits by clicking the Assign Cluster-Wide Provisionary Space Limit button. When you set a reserve limit, you provision a certain amount of space to each volume as a percentage of cluster capacity. You may want to set a reserve limit to free up space that could potentially be unused or allocate more space for replication. As data is written to the volume, available space is automatically allocated. The volume reserve increases up to the reserve limit you set. Mirror Volumes The Mirror Volumes pane displays information about mirror volumes in the cluster: Mnt - whether the volume is mounted Vol Name - the name of the volume Src Vol - the source volume Src Clu - the source cluster Orig Vol -the originating volume for the data being mirrored Orig Clu - the originating cluster for the data being mirrored Last Mirrored - the time at which mirroring was most recently completed - status of the last mirroring operation % Done - progress of the mirroring operation Error(s) - any errors that occurred during the last mirroring operation User Disk Usage The User Disk Usage view displays information about disk usage by cluster users: Name - the username Disk Usage - the total disk space used by the user # Vols - the number of volumes Hard Quota - the user's quota Advisory Quota - the user's advisory quota Email - the user's email address Snapshots The Snapshots view displays the following information about volume snapshots in the cluster: Snapshot Name - the name of the snapshot Volume Name - the name of the source volume volume for the snapshot Disk Space used - the disk space occupied by the snapshot Created - the creation date and time of the snapshot Expires - the expiration date and time of the snapshot Clicking any column name sorts data in ascending or descending order by that column. Selecting the Filter checkbox displays the Filter toolbar, which provides additional data filtering options. Buttons: Remove Snapshot - when the checkboxes beside one or more snapshots are selected, displays the Remove Snapshots dialog Preserve Snapshot - when the checkboxes beside one or more snapshots are selected, prevents the snapshots from expiring Schedules The Schedules view lets you view and edit schedules, which can then can be attached to events to create occurrences. A schedule is a named group of rules that describe one or more points of time in the future at which an action can be specified to take place. The left pane of the Schedules view lists the following information about the existing schedules: Schedule Name - the name of the schedule; clicking a name displays the schedule details in the right pane for editing In Use - indicates whether the schedule is in use ( ), or attached to an action The right pane provides the following tools for creating or editing schedules: Schedule Name - the name of the schedule Schedule Rules - specifies schedule rules with the following components: A dropdown that specifies frequency (Once, Yearly, Monthly, Weekly, Daily, Hourly, Every X minutes) Dropdowns that specify the time within the selected frequency Retain For - the time for which the scheduled snapshot or mirror data is to be retained after creation [ +Add Rule ] - adds another rule to the schedule Navigating away from a schedule with unsaved changes displays the Save Schedule dialog. Buttons: New Schedule - starts editing a new schedule Remove Schedule - displays the Remove Schedule dialog Save Schedule - saves changes to the current schedule Cancel - cancels changes to the current schedule Remove Schedule The Remove Schedule dialog prompts you for confirmation before removing the specified schedule. Buttons Yes - removes the schedule No - exits without removing the schedule NFS HA Views The NFS view group provides the following views: NFS Setup - information about NFS nodes in the cluster VIP Assignments - information about virtual IP addresses (VIPs) in the cluster NFS Nodes - information about NFS nodes in the cluster NFS Setup The NFS Setup view displays information about NFS nodes in the cluster and any VIPs assigned to them: Starting VIP - the starting IP of the VIP range Ending VIP - the ending IP of the VIP range Node Name(s) - the names of the NFS nodes IP Address(es) - the IP addresses of the NFS nodes MAC Address(es) - the MAC addresses associated with the IP addresses Buttons: Start NFS - displays the Manage Node Services dialog Add VIP - displays the Add Virtual IPs dialog Edit - when one or more checkboxes are selected, edits the specified VIP ranges Remove- when one or more checkboxes are selected, removes the specified VIP ranges Unconfigured Nodes - displays nodes not running the NFS service (in the Nodes view) VIP Assignments - displays the VIP Assignments view VIP Assignments The VIP Assignments view displays VIP assignments beside the nodes to which they are assigned: Virtual IP Address - each VIP in the range Node Name - the node to which the VIP is assigned IP Address - the IP address of the node MAC Address - the MAC address associated with the IP address Buttons: Start NFS - displays the Manage Node Services dialog Add VIP - displays the Add Virtual IPs dialog Unconfigured Nodes - displays nodes not running the NFS service (in the Nodes view) NFS Nodes The NFS Nodes view displays information about nodes running the NFS service: Hlth - the health of the node Hostname - the hostname of the node Physical IP(s) - physical IP addresses associated with the node Virtual IP(s) - virtual IP addresses associated with the node Buttons: Properties - when one or more nodes are selected, navigates to the Node Properties View Forget Node - navigates to the Remove Node dialog, which lets you remove the node Manage Services - navigates to the Manage Node Services dialog, which lets you start and stop services on the node Change Topology - navigates to the Change Node Topology dialog, which lets you change the rack or switch path for a node Alarms Views The Alarms view group provides the following views: Node Alarms - information about node alarms in the cluster Volume Alarms - information about volume alarms in the cluster User/Group Alarms - information about users or groups that have exceeded quotas Alarm Notifications - configure where notifications are sent when alarms are raised The following controls are available on views: Clicking any column name sorts data in ascending or descending order by that column. Selecting the Filter checkbox displays the Filter toolbar, which provides additional data filtering options. Clicking the Column Controls icon ( ) opens a dialog that lets you select which columns to view. Click any item to toggle its column on or off. You can also specify the refresh rate for updating data on the page. For example: Node Alarms The Node Alarms view displays information about alarms on any node in the cluster that has raised an alarm. The first two columns display Hlth - a color indicating the status of each node (see Cluster Heat Map) Hostname - the hostname of the node The remaining columns are based on alarm type, such as: Version Alarm - one or more services on the node are running an unexpected version No Heartbeat Alarm - no heartbeat has been detected for over 5 minutes, and the node is not undergoing maintenance UID Mismatch Alarm - services in the cluster are being run with different usernames (UIDs) Duplicate HostId Alarm - two or more nodes in the cluster have the same Host ID Too Many Containers Alarm - the number of containers on this node reached the maximum limit Excess Logs Alarm - debug logging is enabled on this node, which can fill up disk space Disk Failure Alarm - a disk has failed on the node (the disk health log indicates which one failed) Time Skew Alarm - the clock on the node is out of sync with the master CLDB by more than 20 seconds Root Partition Full Alarm - the root partition ("/") on the node is 99% full and running out of space Installation Directory Full Alarm - the partition /opt/mapr on the node is running out of space (95% full) Core Present Alarm - a service on the node has crashed and created a core dump file High FileServer Memory Alarm - the FileServer service on the node has high memory consumption Pam Misconfigured Alarm - the PAM authentication on the node is configured incorrectly TaskTracker Local Directory Full Alarm - the local directory used by the TaskTracker is full, and the TaskTracker cannot operate as a result CLDB Alarm - the CLDB service on the node has stopped running FileServer Alarm - the FileServer service on the node has stopped running JobTracker Alarm - the JobTracker service on the node has stopped running TaskTracker Alarm - the TaskTracker service on the node has stopped running HBase Master Alarm - the HBase Master service on the node has stopped running HBase RegionServer Alarm - the HBase RegionServer service on the node has stopped running NFS Gateway Alarm - the NFS Gateway service on the node has stopped running Webserver Alarm - the WebServer service on the node has stopped running HostStats Alarm - the HostStats service on the node has stopped running Metrics write problem Alarm - metric data was not written to the database, or there were issues writing to a logical volume See Alarms Reference. Note the following behavior on the Node Alarms view: Clicking a node's Hostname navigates to the Node Properties View, which provides detailed information about the node. The left pane of the Node Alarms view displays the available topologies. Click a topology name to view only the nodes in that topology. Buttons: Properties - navigates to the Node Properties View Forget Node - opens the Forget Node dialog to remove the node(s) from active management in this cluster. Services on the node must be stopped before the node can be forgotten. Manage Services - opens the Manage Node Services dialog, which lets you start and stop services on the node Change Topology - opens the Change Node Topology dialog, which lets you change the rack or switch path for a node Volume Alarms The Volume Alarms view displays information about volume alarms in the cluster: Mnt - whether the volume is mounted Vol Name - the name of the volume Snapshot Alarm - last Snapshot Failed alarm Mirror Alarm - last Mirror Failed alarm Replication Alarm - last Data Under-Replicated alarm Data Alarm - last Data Unavailable alarm Vol Advisory Quota Alarm - last Volume Advisory Quota Exceeded alarm Vol Quota Alarm- last Volume Quota Exceeded alarm Clicking any column name sorts data in ascending or descending order by that column. Clicking a volume name displays the Volume Properties dialog Selecting the Show Unmounted checkbox shows unmounted volumes as well as mounted volumes. Selecting the Filter checkbox displays the Filter toolbar, which provides additional data filtering options.Buttons: New Volume displays the New Volume Dialog. Properties - if the checkboxes beside one or more volumes is selected,displays the Volume Properties dialog Mount (Unmount) - if an unmounted volume is selected, mounts it; if a mounted volume is selected, unmounts it Remove - if the checkboxes beside one or more volumes is selected, displays the Remove Volume dialog Start Mirroring - if a mirror volume is selected, starts the mirror sync process Snapshots - if the checkboxes beside one or more volumes is selected,displays the Snapshots for Volume dialog New Snapshot - if the checkboxes beside one or more volumes is selected,displays the Snapshot Name dialog User/Group Alarms The User/Group Alarms view displays information about user and group quota alarms in the cluster: Name - the name of the user or group User Advisory Quota Alarm - the last Advisory Quota Exceeded alarm User Quota Alarm - the last Quota Exceeded alarm Buttons: Edit Properties - opens a User Properties dialog box that lets you change user properties and clear alarms. Alerts The Alerts dialog lets you specify which alarms cause a notification event and where email notifications are sent. Fields: Alarm Name - select the alarm to configure Standard Notification - send notification to the default for the alarm type (the cluster administrator or volume creator, for example) Additional Email Address - specify an additional custom email address to receive notifications for the alarm type Buttons: OK - save changes and exit Cancel - exit without saving changes System Settings Views The System Settings view group provides the following views: Email Addresses - specify MapR user email addresses Permissions - give permissions to users Quota Defaults - settings for default quotas in the cluster Balancer Settings - settings to configure the disk space and role replication Configuring Balancer Settings on the cluster. SMTP - settings for sending email from MapR HTTP - settings for accessing the MapR Control System via a browser (this view is only available in versions 2.1 through 3.0.2) Manage Licenses - MapR license settings Metrics Database - Settings for the MapR Metrics MySQL database Email Addresses The Configure Email Addresses dialog lets you specify whether MapR gets user email addresses from an LDAP directory, or uses a company domain: Use Company Domain - specify a domain to append after each username to determine each user's email address Use LDAP - obtain each user's email address from an LDAP server Buttons: OK - save changes and exit Cancel - exit without saving changes Permissions The User Permissions dialog lets you grant specific cluster permissions to particular users and groups. User/Group field - the user or group to which permissions are to be granted (one user or group per row) Permissions field - the permissions to grant to the user or group (see the Permissions table below) Delete button ( ) - deletes the current row [ + Add Permission ] - adds a new row Cluster Permissions The following table lists the actions a user can perform on a cluster, and the corresponding codes used in the cluster ACL: Code Allowed Action login Log in to the MapR Control System, use the API and command-line interface, read access on cluster and volumes ss Start/stop services cv Create volumes a Administrative access (can edit and view ACLs, but cannot perform cluster operations) fc Full control over the cluster (this enables all cluster-related administrative options with the exception of changing the cluster ACLs) Buttons: OK - save changes and exit Cancel - exit without saving changes Quota Defaults The Configure Quota Defaults dialog lets you set the default quotas that apply to users and groups. The User Quota Defaults section contains the following fields: Default User Advisory Quota - if selected, sets the advisory quota that applies to all users without an explicit advisory quota. Default User Total Quota - if selected, sets the advisory quota that applies to all users without an explicit total quota. The Group Quota Defaults section contains the following fields: Default Group Advisory Quota - if selected, sets the advisory quota that applies to all groups without an explicit advisory quota. Default Group Total Quota - if selected, sets the advisory quota that applies to all groups without an explicit total quota. Buttons: OK - saves the settings Cancel - exits without saving the settings SMTP The Configure SMTP dialog lets you configure the email account from which the MapR cluster sends alerts and other notifications. The Configure Sending Email (SMTP) dialog contains the following fields: Provider - selects Gmail or another email provider; if you select Gmail, the other fields are partially populated to help you with the configuration SMTP Server specifies the SMTP server to use when sending email. The server requires an encrypted connection (SSL) - use SSL when connecting to the SMTP server SMTP Port - the port to use on the SMTP server Full Name - the name used in the From field when the cluster sends an alert email Email Address - the email address used in the From field when the cluster sends an alert email. Username - the username used to log onto the email account the cluster will use to send email. SMTP Password - the password to use when sending email. Buttons: OK - saves the settings Cancel - exits without saving the settings Balancer Settings The Balancer Configuration dialog enables you to configure the behaviors of the disk space and role replication Configuring Balancer Settings. The Balancer Configuration dialog has the following elements: Balancer Controls: Contains toggle settings for the Disk Balancer and the Role Balancer. Set a balancer's toggle to ON to enable that balancer. Disk Balancer Settings: Configures the behavior of the disk balancer. Disk Balancer Presets: These preconfigured settings enable quick setting of policies for Rapid, Moderate, and Relaxed disk balancing. The default setting is Moderate. Threshold: Move this slider to set a percentage usage of a storage pool that makes the storage pool eligible for rebalancing operations. The default value for this setting is 70%. % Concurrent Disk Rebalancers: Move this slider to set the maximum percentage of data that is actively being rebalanced at a given time. Rebalancing operations will not affect more data than the value of this slider. The default value for this setting is 10%. Role Balancer Settings: Configures the behavior of the role balancer. Role Balancer Presets: These preconfigured settings enable quick setting of policies for Rapid, Moderate, and Relaxed role balancing. The default setting is Moderate. % Concurrent Role Rebalancers: Move this slider to set the maximum percentage of data that is actively being rebalanced at a given time. Role rebalancing operations will not affect more data than the value of this slider. The default value for this setting is 10%. Delay For Active Data: Move this slider to set a time frame in seconds. Role rebalancing operations skip any data that was active within the specified time frame. The default value for this setting is 600 seconds. Buttons: OK - saves the settings Cancel - exits without saving the settings HTTP (versions 2.1 through 3.0.2 only) In versions 2.1 through 3.0.2, the MCS includes a Configure HTTP dialog that lets you configure access to the MapR Control System via HTTP and HTTPS. As an alternative, you can edit web.conf, which resides in the /opt/mapr/conf directory. In version 3.1 and later, the HTTP settings can only be modified by editing web.conf directly. The sections in the Configure HTTP dialog let you enable HTTP and HTTPS access, and set the session timeout, respectively: Enable HTTP Access - if selected, configure HTTP access with the following field: HTTP Port - the port on which to connect to the MapR Control System via HTTP Enable HTTPS Access - if selected, configure HTTPS access with the following fields: HTTPS Port - the port on which to connect to the MapR Control System via HTTPS HTTPS Keystore Path - a path to the HTTPS keystore HTTPS Keystore Password - a password to access the HTTPS keystore HTTPS Key Password - a password to access the HTTPS key Session Timeout - the number of seconds before an idle session times out. Buttons: OK - saves the settings Cancel - exits without saving the settings Manage Licenses The License Management dialog lets you add and activate licenses for the cluster, and displays the Cluster ID and the following information about existing licenses: Name - the name of each license Issued - the date each license was issued Expires - the expiration date of each license Nodes - the nodes to which each license applies Fields: Cluster ID - the unique identifier needed for licensing the cluster Buttons: Add Licenses via Web - navigates to the MapR licensing form online Add License via Upload - alternate licensing mechanism: upload via browser Add License via Copy/Paste - alternate licensing mechanism: paste license key Apply Licenses - validates the licenses and applies them to the cluster Cancel - closes the dialog. Metrics The Configure Metrics Database dialog enables you to specify the location and login credentials of the MySQL server that stores information for J ob Metrics. Fields: URL - the hostname and port of the machine running the MySQL server Username - the username for the MySQL metrics database Password - the password for the MySQL metrics database Buttons: OK - saves the MySQL information in the fields Cancel - closes the dialog Other Views In addition to the MapR Control System views, there are views that display detailed information about the system: CLDB View - information about the container location database HBase View - information about HBase on the cluster JobTracker View - information about the JobTracker Nagios View - information about the Nagios configuration script JobHistory Server View - information about MapReduce jobs after their ApplicationMaster terminates With the exception of the MapR Launchpad, the above views include the following buttons: - Refresh Button (refreshes the view) - Popout Button (opens the view in a new browser window) CLDB View The CLDB view provides information about the Container Location Database (CLDB). The CLDB is a management service that keeps track of container locations and the root of volumes. To display the CLDB view, open the MapR Control System and click CLDB in the navigation pane. The following table describes the fields on the CLDB view: Field Description CLDB Mode The CLDB node can be in the following modes: MASTER_READ_WRITE, SLAVE_READ_ONLY, or / CLDB BuildVersion Lists the build version. CLDB Status Can be RUNNING, or Cluster Capacity Lists the storage capacity for the cluster. Cluster Used Lists the amount of storage in use. Cluster Available Lists the amount of available storage. Active FileServers A list of FileServers, and the following information about each: ServerID (Hex) - The server's ID in hexadecimal notation. ServerID - The server's ID in decimal notation. HostPort - The IP address of the host HostName - The hostname assigned to that file server. Network Location - The network topology for that file server. Last Heartbeat (s) - The timestamp for the last received heartbeat. State - Can be ACTIVE or Capacity (MB) - Total storage capacity on this server. Used (MB) - Storage used on this server. Available (MB) - Storage available on this server. In Transit (MB) - Active NFS Servers A list of NFS servers, and the following information about each: ServerID (Hex) - The server's ID in hexadecimal notation. ServerID - The server's ID in decimal notation. HostPort - The IP address of the host HostName - The hostname assigned to that file server. Last Heartbeat (s) - The timestamp for the last received heartbeat. State - Can be Active or Volumes A list of volumes, and the following information about each: Volume Name Mount Point - The path of where the volume is mounted over NFS. Mounted - Can be Y or N. ReadOnly - Can be Y or N. Volume ID - The Volume ID Volume Topology - The path describing the topology to which the volume is assigned. Quota - The total size of the volume's quota. A quota of 0 means no quota is assigned. Advisory Quota - The usage level that triggers a disk usage warning. Used - Total size of data written to the volume LogicalUsed - Actual size of data written to the volume Root Container ID - The ID of the root container. Replication Guaranteed Replication - Accounting Entities A list of users and groups, and the following information about each: AE Name AE Type AE Quota AE Advisory Quota AE Used - Mirrors A list of mirrors, and the following information about each: Mirror Volume Name Mirror ID Mirror NextID Mirror Status Last Successful Mirror Time Mirror SrcVolume Mirror SrcRootContainerID Mirror SrcClusterName Mirror SrcSnapshot Mirror DataGenerator Volume - Snapshots A list of snapshots, and the following information about each: Snapshot ID RW Volume ID Snapshot Name Root Container ID Snapshot Size Snapshot InProgress - Containers A list of containers, and the following information about each: Container ID Volume ID Latest Epoch SizeMB Container Master Location Container Locations Inactive Locations Unused Locations Replication Type - Snapshot Containers A list of snapshot containers, and the following information about each: Snapshot Container ID - unique ID of the container Snapshot ID - ID of the snapshot corresponding to the container RW Container ID - corresponding source container ID Latest Epoch SizeMB - container size, in MB Container Master Location - location of the container's master replica Container Locations Inactive Locations - HBase View The HBase View provides information about HBase on the cluster. Field Description Local Logs A link to the HBase Local Logs View Thread Dump A link to the HBase Thread Dump View Log Level A link to the HBase Log Level View, a form for getting/setting the log level Master Attributes A list of attributes, and the following information about each: Attribute Name Value Description - Catalog Tables A list of tables, and the following information about each: Table Description - User Tables Region Servers A list of region servers in the cluster, and the following information about each: Address Start Code Load Total - HBase Local Logs View The HBase Local Logs view displays a list of the local HBase logs. Clicking a log name displays the contents of the log. Each log name can be copied and pasted into the HBase Log Level View to get or set the current log level. HBase Log Level View The HBase Log Level View is a form for getting and setting log levels that determine which information gets logged. The Log field accepts a log name (which can be copied from the HBase Local Logs View and pasted). The Level field takes any of the following valid log levels: ALL TRACE DEBUG INFO WARN ERROR OFF HBase Thread Dump View The HBase Thread Dump View displays a dump of the HBase thread. Example: Process Thread Dump: 40 active threads Thread 318 (1962516546@qtp-879081272-3): State: RUNNABLE Blocked count: 8 Waited count: 32 Stack: sun.management.ThreadImpl.getThreadInfo0(Native Method) sun.management.ThreadImpl.getThreadInfo(ThreadImpl.java:147) sun.management.ThreadImpl.getThreadInfo(ThreadImpl.java:123) org.apache.hadoop.util.ReflectionUtils.printThreadInfo(ReflectionUtils.java:149) org.apache.hadoop.http.HttpServer$StackServlet.doGet(HttpServer.java:695) javax.servlet.http.HttpServlet.service(HttpServlet.java:707) javax.servlet.http.HttpServlet.service(HttpServlet.java:820) org.mortbay.jetty.servlet.ServletHolder.handle(ServletHolder.java:511) org.mortbay.jetty.servlet.ServletHandler$CachedChain.doFilter(ServletHandler.java:1221 ) org.apache.hadoop.http.HttpServer$QuotingInputFilter.doFilter(HttpServer.java:826) org.mortbay.jetty.servlet.ServletHandler$CachedChain.doFilter(ServletHandler.java:1212 ) org.mortbay.jetty.servlet.ServletHandler.handle(ServletHandler.java:399) org.mortbay.jetty.security.SecurityHandler.handle(SecurityHandler.java:216) org.mortbay.jetty.servlet.SessionHandler.handle(SessionHandler.java:182) org.mortbay.jetty.handler.ContextHandler.handle(ContextHandler.java:766) org.mortbay.jetty.webapp.WebAppContext.handle(WebAppContext.java:450) org.mortbay.jetty.handler.ContextHandlerCollection.handle(ContextHandlerCollection.jav a:230) org.mortbay.jetty.handler.HandlerWrapper.handle(HandlerWrapper.java:152) org.mortbay.jetty.Server.handle(Server.java:326) org.mortbay.jetty.HttpConnection.handleRequest(HttpConnection.java:542) Thread 50 (perfnode51.perf.lab:60000-CatalogJanitor): State: TIMED_WAITING Blocked count: 1081 Waited count: 1350 Stack: java.lang.Object.wait(Native Method) org.apache.hadoop.hbase.util.Sleeper.sleep(Sleeper.java:91) org.apache.hadoop.hbase.Chore.run(Chore.java:74) Thread 49 (perfnode51.perf.lab:60000-BalancerChore): State: TIMED_WAITING Blocked count: 0 Waited count: 270 Stack: java.lang.Object.wait(Native Method) org.apache.hadoop.hbase.util.Sleeper.sleep(Sleeper.java:91) org.apache.hadoop.hbase.Chore.run(Chore.java:74) Thread 48 (MASTER_OPEN_REGION-perfnode51.perf.lab:60000-1): State: WAITING Blocked count: 2 Waited count: 3 Waiting on java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject@6d1cf4e5 Stack: sun.misc.Unsafe.park(Native Method) java.util.concurrent.locks.LockSupport.park(LockSupport.java:158) java.util.concurrent.locks.AbstractQueuedSynchronizer$ConditionObject.await(AbstractQu euedSynchronizer.java:1925) java.util.concurrent.LinkedBlockingQueue.take(LinkedBlockingQueue.java:399) java.util.concurrent.ThreadPoolExecutor.getTask(ThreadPoolExecutor.java:947) JobTracker View Field State Started Version Compiled Identifier Description Cluster Summary The heapsize, and the following information about the cluster: Running Map Tasks Running Reduce Tasks Total Submissions Nodes Occupied Map Slots Occupied Reduce Slots Reserved Map Slots Reserved Reduce Slots Map Task Capacity Reduce Task Capacity Avg. Tasks/Node Blacklisted Nodes Excluded Nodes MapTask Prefetch Capacity - Scheduling Information A list of queues, and the following information about each: Queue name State Scheduling Information - Filter A field for filtering results by Job ID, Priority, User, or Name Running Jobs A list of running MapReduce jobs, and the following information about each: JobId Priority User Name Start Time Map % Complete Current Map Slots Failed MapAttempts MapAttempt Time Avg/Max Cumulative Map CPU Current Map PMem Reduce % Complete Current Reduce Slots Failed ReduceAttempts ReduceAttempt Time Avg/Max Cumulative Reduce CPU Current Reduce PMem - Completed Jobs A list of current MapReduce jobs, and the following information about each: JobId Priority User Name Start Time Total Time Maps Launched Map Total Failed MapAttempts MapAttempt Time Avg/Max Cumulative Map CPU Reducers Launched Reduce Total Failed ReduceAttempts ReduceAttempt Time Avg/Max Cumulative Reduce CPU Cumulative Reduce PMem Vaidya Reports - Retired Jobs A list of retired MapReduce job, and the following information about each: JobId Priority User Name State Start Time Finish Time Map % Complete Reduce % Complete Job Scheduling Information Diagnostic Info - Local Logs A link to the local logs JobTracker Configuration A link to a page containing Hadoop JobTracker configuration values JobTracker Configuration View Field Default fs.automatic.close TRUE fs.checkpoint.dir ${hadoop.tmp.dir}/dfs/namesecondary fs.checkpoint.edits.dir ${fs.checkpoint.dir} fs.checkpoint.period 3600 fs.checkpoint.size 67108864 fs.default.name maprfs:/// fs.file.impl org.apache.hadoop.fs.LocalFileSystem fs.ftp.impl org.apache.hadoop.fs.ftp.FTPFileSystem fs.har.impl org.apache.hadoop.fs.HarFileSystem fs.har.impl.disable.cache TRUE fs.hdfs.impl org.apache.hadoop.hdfs.DistributedFileSystem fs.hftp.impl org.apache.hadoop.hdfs.HftpFileSystem fs.hsftp.impl org.apache.hadoop.hdfs.HsftpFileSystem fs.kfs.impl org.apache.hadoop.fs.kfs.KosmosFileSystem fs.maprfs.impl com.mapr.fs.MapRFileSystem fs.ramfs.impl org.apache.hadoop.fs.InMemoryFileSystem fs.s3.block.size 67108864 fs.s3.buffer.dir ${hadoop.tmp.dir}/s3 fs.s3.impl org.apache.hadoop.fs.s3.S3FileSystem fs.s3.maxRetries 4 fs.s3.sleepTimeSeconds 10 fs.s3n.block.size 67108864 fs.s3n.impl org.apache.hadoop.fs.s3native.NativeS3FileSystem fs.trash.interval 0 hadoop.job.history.location file:////opt/mapr/hadoop/hadoop-0.20.2/bin/../logs/history hadoop.logfile.count 10 hadoop.logfile.size 10000000 hadoop.native.lib TRUE hadoop.proxyuser.root.groups root hadoop.proxyuser.root.hosts (none) hadoop.rpc.socket.factory.class.default org.apache.hadoop.net.StandardSocketFactory hadoop.security.authentication simple hadoop.security.authorization FALSE hadoop.security.group.mapping org.apache.hadoop.security.ShellBasedUnixGroupsMapping hadoop.tmp.dir /tmp/hadoop-${user.name} hadoop.util.hash.type murmur io.bytes.per.checksum 512 io.compression.codecs org.apache.hadoop.io.compress.DefaultCodec,org.apache.hadoop.io.compress.GzipCodec,org. io.file.buffer.size 8192 io.map.index.skip 0 io.mapfile.bloom.error.rate 0.005 io.mapfile.bloom.size 1048576 io.seqfile.compress.blocksize 1000000 io.seqfile.lazydecompress TRUE io.seqfile.sorter.recordlimit 1000000 io.serializations org.apache.hadoop.io.serializer.WritableSerialization io.skip.checksum.errors FALSE io.sort.factor 256 io.sort.record.percent 0.17 io.sort.spill.percent 0.99 ipc.client.connect.max.retries 10 ipc.client.connection.maxidletime 10000 ipc.client.idlethreshold 4000 ipc.client.kill.max 10 ipc.client.tcpnodelay FALSE ipc.server.listen.queue.size 128 ipc.server.tcpnodelay FALSE job.end.retry.attempts 0 job.end.retry.interval 30000 jobclient.completion.poll.interval 5000 jobclient.output.filter FAILED jobclient.progress.monitor.poll.interval 1000 keep.failed.task.files FALSE local.cache.size 10737418240 map.sort.class org.apache.hadoop.util.QuickSort mapr.localoutput.dir output mapr.localspill.dir spill mapr.localvolumes.path /var/mapr/local mapred.acls.enabled FALSE mapred.child.oom_adj 10 mapred.child.renice 10 mapred.child.taskset TRUE mapred.child.tmp ./tmp mapred.cluster.ephemeral.tasks.memory.limit.mb 200 mapred.compress.map.output FALSE mapred.fairscheduler.allocation.file conf/pools.xml mapred.fairscheduler.assignmultiple TRUE mapred.fairscheduler.eventlog.enabled FALSE mapred.fairscheduler.smalljob.max.inputsize 10737418240 mapred.fairscheduler.smalljob.max.maps 10 mapred.fairscheduler.smalljob.max.reducer.inputsize 1073741824 mapred.fairscheduler.smalljob.max.reducers 10 mapred.fairscheduler.smalljob.schedule.enable TRUE mapred.healthChecker.interval 60000 mapred.healthChecker.script.timeout 600000 mapred.inmem.merge.threshold 1000 mapred.job.queue.name default mapred.job.reduce.input.buffer.percent 0 mapred.job.reuse.jvm.num.tasks -1 mapred.job.shuffle.input.buffer.percent 0.7 mapred.job.shuffle.merge.percent 0.66 mapred.job.tracker :9001 mapred.job.tracker.handler.count 10 mapred.job.tracker.history.completed.location /var/mapr/cluster/mapred/jobTracker/history/done mapred.job.tracker.http.address 0.0.0.0:50030 mapred.job.tracker.persist.jobstatus.active FALSE mapred.job.tracker.persist.jobstatus.dir /var/mapr/cluster/mapred/jobTracker/jobsInfo mapred.job.tracker.persist.jobstatus.hours 0 mapred.jobtracker.completeuserjobs.maximum 100 mapred.jobtracker.instrumentation org.apache.hadoop.mapred.JobTrackerMetricsInst mapred.jobtracker.job.history.block.size 3145728 mapred.jobtracker.jobhistory.lru.cache.size 5 mapred.jobtracker.maxtasks.per.job -1 mapred.jobtracker.port 9001 mapred.jobtracker.restart.recover TRUE mapred.jobtracker.retiredjobs.cache.size 1000 mapred.jobtracker.taskScheduler org.apache.hadoop.mapred.FairScheduler mapred.line.input.format.linespermap 1 mapred.local.dir ${hadoop.tmp.dir}/mapred/local mapred.local.dir.minspacekill 0 mapred.local.dir.minspacestart 0 mapred.map.child.java.opts -XX:ErrorFile=/opt/cores/mapreduce_java_error%p.log mapred.map.max.attempts 4 mapred.map.output.compression.codec org.apache.hadoop.io.compress.DefaultCodec mapred.map.tasks 2 mapred.map.tasks.speculative.execution TRUE mapred.max.maps.per.node -1 mapred.max.reduces.per.node -1 mapred.max.tracker.blacklists 4 mapred.max.tracker.failures 4 mapred.merge.recordsBeforeProgress 10000 mapred.min.split.size 0 mapred.output.compress FALSE mapred.output.compression.codec org.apache.hadoop.io.compress.DefaultCodec mapred.output.compression.type RECORD mapred.queue.names default mapred.reduce.child.java.opts -XX:ErrorFile=/opt/cores/mapreduce_java_error%p.log mapred.reduce.copy.backoff 300 mapred.reduce.max.attempts 4 mapred.reduce.parallel.copies 12 mapred.reduce.slowstart.completed.maps 0.95 mapred.reduce.tasks 1 mapred.reduce.tasks.speculative.execution FALSE mapred.running.map.limit -1 mapred.running.reduce.limit -1 mapred.skip.attempts.to.start.skipping 2 mapred.skip.map.auto.incr.proc.count TRUE mapred.skip.map.max.skip.records 0 mapred.skip.reduce.auto.incr.proc.count TRUE mapred.skip.reduce.max.skip.groups 0 mapred.submit.replication 10 mapred.system.dir /var/mapr/cluster/mapred/jobTracker/system mapred.task.cache.levels 2 mapred.task.profile FALSE mapred.task.profile.maps 0-2 mapred.task.profile.reduces 0-2 mapred.task.timeout 600000 mapred.task.tracker.http.address 0.0.0.0:50060 mapred.task.tracker.report.address 127.0.0.1:0 mapred.task.tracker.task-controller org.apache.hadoop.mapred.DefaultTaskController mapred.tasktracker.dns.interface default mapred.tasktracker.dns.nameserver default mapred.tasktracker.ephemeral.tasks.maximum 1 mapred.tasktracker.ephemeral.tasks.timeout 10000 mapred.tasktracker.ephemeral.tasks.ulimit 4294967296> mapred.tasktracker.expiry.interval 600000 mapred.tasktracker.indexcache.mb 10 mapred.tasktracker.instrumentation org.apache.hadoop.mapred.TaskTrackerMetricsInst mapred.tasktracker.map.tasks.maximum (CPUS > 2) ? (CPUS * 0.75) : 1 mapred.tasktracker.reduce.tasks.maximum (CPUS > 2) ? (CPUS * 0.50): 1 mapred.tasktracker.taskmemorymanager.monitoring-interval 5000 mapred.tasktracker.tasks.sleeptime-before-sigkill 5000 mapred.temp.dir ${hadoop.tmp.dir}/mapred/temp mapred.userlog.limit.kb 0 mapred.userlog.retain.hours 24 mapreduce.heartbeat.10 300 mapreduce.heartbeat.100 1000 mapreduce.heartbeat.1000 10000 mapreduce.heartbeat.10000 100000 mapreduce.job.acl-view-job mapreduce.job.complete.cancel.delegation.tokens TRUE mapreduce.job.split.metainfo.maxsize 10000000 mapreduce.jobtracker.recovery.dir /var/mapr/cluster/mapred/jobTracker/recovery mapreduce.jobtracker.recovery.maxtime 120 mapreduce.jobtracker.staging.root.dir /var/mapr/cluster/mapred/jobTracker/staging mapreduce.maprfs.use.compression TRUE mapreduce.reduce.input.limit -1 mapreduce.tasktracker.outofband.heartbeat FALSE mapreduce.tasktracker.prefetch.maptasks 1 mapreduce.use.fastreduce FALSE mapreduce.use.maprfs TRUE tasktracker.http.threads 2 topology.node.switch.mapping.impl org.apache.hadoop.net.ScriptBasedMapping topology.script.number.args 100 webinterface.private.actions FALSE Nagios View The Nagios view displays a dialog containing a Nagios configuration script. Example: ############# Commands ############# define command { command_name check_fileserver_proc command_line $USER1$/check_tcp -p 5660 } define command { command_name check_cldb_proc command_line $USER1$/check_tcp -p 7222 } define command { command_name check_jobtracker_proc command_line $USER1$/check_tcp -p 50030 } define command { command_name check_tasktracker_proc command_line $USER1$/check_tcp -p 50060 } define command { command_name check_nfs_proc command_line $USER1$/check_tcp -p 2049 } define command { command_name check_hbmaster_proc command_line $USER1$/check_tcp -p 60000 } define command { command_name check_hbregionserver_proc command_line $USER1$/check_tcp -p 60020 } define command { command_name check_webserver_proc command_line $USER1$/check_tcp -p 8443 } ################# HOST: perfnode51.perf.lab ############### define host { use linux-server host_name perfnode51.perf.lab address 10.10.30.51 check_command check-host-alive } ################# HOST: perfnode52.perf.lab ############### define host { use linux-server host_name perfnode52.perf.lab address 10.10.30.52 check_command check-host-alive } ################# HOST: perfnode53.perf.lab ############### define host { use linux-server host_name perfnode53.perf.lab address 10.10.30.53 check_command check-host-alive } ################# HOST: perfnode54.perf.lab ############### define host { use linux-server host_name perfnode54.perf.lab address 10.10.30.54 check_command check-host-alive } ################# HOST: perfnode55.perf.lab ############### define host { use linux-server host_name perfnode55.perf.lab address 10.10.30.55 check_command check-host-alive } ################# HOST: perfnode56.perf.lab ############### define host { use linux-server host_name perfnode56.perf.lab address 10.10.30.56 check_command check-host-alive } Node-Related Dialog Boxes This page describes the node-related dialog boxes, which are accessible in most views that list node details. This includes the following dialog boxes: Forget Node Manage Node Services Change Node Topology Forget Node The Forget Node dialog confirms that you wish to remove a node from active management in this cluster. Services on the node must be stopped before the node can be forgotten. Manage Node Services The Manage Node Services dialog lets you start and stop services on a node, or multiple nodes. The Service Changes section contains a dropdown menu for each service: No change - leave the service running if it is running, or stopped if it is stopped Start - start the service Stop - stop the service Restart - restart the service Buttons: OK - start and stop the selected services as specified by the dropdown menus Cancel - returns to the Node Properties View without starting or stopping any services You can also start and stop services in the Manage Node Services pane of the Node Properties view. Change Node Topology The Change Node Topology dialog lets you change the rack or switch path for one or more nodes. The Change Node Topology dialog consists of: Nodes affected (the node or nodes to be moved, as specified in the Nodes view) A field with a dropdown menu for the new node topology path The Change Node Topology dialog contains the following buttons: OK - changes the node topology Cancel - returns to the Nodes view without changing the node topology Hadoop Commands All Hadoop commands are invoked by the bin/hadoop script. When you run these commands, you can specify the MapReduce mode in two different ways: 1. Use the hadoop keyword and specify the mode explicitly, where classic mode refers to Hadoop 1.x and yarn mode refers to Hadoop 2.x. 2. Use the hadoop1 or hadoop2 keyword and do not specify the mode. For example, the following commands are equivalent: root@testnode100:/opt/mapr/conf/conf.d# hadoop2 conf | grep mapreduce.map.memory.mb mapreduce.map.memory.mb1024 root@testnode100:/opt/mapr/conf/conf.d# hadoop -yarn conf | grep mapreduce.map.memory.mb mapreduce.map.memory.mb1024 The following syntax summary applies to all commands. Syntax Summary hadoop [-yarn|-classic] [--config confdir] [COMMAND] [GENERIC_OPTIONS] [COMMAND_OPTIONS] hadoop1 [--config confdir] [COMMAND] [GENERIC_OPTIONS] [COMMAND_OPTIONS] hadoop2 [--config confdir] [COMMAND] [GENERIC_OPTIONS] [COMMAND_OPTIONS] Hadoop has an option parsing framework that employs parsing generic options as well as running classes. COMMAND_OPTION Description -mode Specifies the Hadoop version: yarn or classic Aternatively, you can use a hadoop1 or hadoop2 command without setting the mode. If you use a hadoop command (instead of hadoop1 or hadoop2) and do not set the mode, the command runs in the mode set by the MAPR_MAPREDUCE_MODE environment variable. If this variable is not set, the command runs in the mode set in the hadoop version file on the node (default_mode = yarn or classic). --config confdir Overwrites the default Configuration directory. Default is ${HADOOP_HOME}/conf. COMMAND Various commands with their options are described in the following sections. GENERIC_OPTIONS The common set of options supported by multiple commands. COMMAND_OPTIONS Various command options are described in the following sections. Useful Information Running the hadoop script without any arguments prints the help description for all commands. Supported Commands for Hadoop 1.x MapR supports the following hadoop commands for Hadoop 1.x: Command Description archive -archiveName NAME * The hadoop archive command creates a Hadoop archive, a file that contains other files. A Hadoop archive always has a *.har extension. classpath The hadoop classpath command prints the class path needed to access the Hadoop JAR and the required libraries. conf The hadoop conf command prints the configuration information for the current node. daemonlog The hadoop daemonlog command may be used to get or set the log level of Hadoop daemons. distcp The hadoop distcp command is a tool for large inter- and intra-cluster copying. It uses MapReduce to effect its distribution, error handling and recovery, and reporting. It expands a list of files and directories into input to map tasks, each of which will copy a partition of the files specified in the source list. fs The hadoop fs command runs a generic filesystem user client that interacts with the MapR filesystem (MapR-FS). jar The hadoop jar command runs a JAR file. Users can bundle their MapReduce code in a JAR file and execute it using this command. job Manipulates MapReduce jobs. jobtracker Runs the MapReduce Jobtracker node. mfs The hadoop mfs command performs operations on directories in the cluster. The main purposes of hadoop mfs are to display directory information and contents, to create symbolic links, and to set compression and chunk size on a directory. mradmin Runs a MapReduce admin client. pipes Runs a pipes job. queue Gets information about job queues. tasktracker The hadoop tasktracker command runs a MapReduce tasktracker node. version The hadoop version command prints the Hadoop software version. Supported Commands for Hadoop 2.x MapR supports the following hadoop commands for Hadoop 2.x: Command Description archive -archiveName NAME * Creates a Hadoop archive, a file that contains other files. A Hadoop archive always has a .har extension. CLASSNAME The hadoop script can be used to invoke any class. hadoop CLASSNAME runs the class named CLASSNAME. classpath Prints the class path needed to access the Hadoop JAR and the required libraries. conf The hadoop conf command prints the configuration information for the current node. daemonlog The hadoop daemonlog command may be used to get or set the log level of Hadoop daemons. distcp The hadoop distcp command is a tool for large inter- and intra-cluster copying. It uses MapReduce to effect its distribution, error handling and recovery, and reporting. It expands a list of files and directories into input to map tasks, each of which will copy a partition of the files specified in the source list. fs The hadoop fs command runs a generic filesystem user client that interacts with the MapR filesystem (MapR-FS). jar The hadoop jar command runs a JAR file. Users can bundle their MapReduce code in a JAR file and execute it using this command. jobtracker Runs the MapReduce Jobtracker node. mfs The hadoop mfs command performs operations on directories in the cluster. The main purposes of hadoop mfs are to display directory information and contents, to create symbolic links, and to set compression and chunk size on a directory. mradmin Runs a MapReduce admin client. tasktracker The hadoop tasktracker command runs a MapReduce tasktracker node. version The hadoop version command prints the Hadoop software version. For Hadoop2, some hadoop commands are deprecated and replaced by the mapred command. For example, if you run the hadoop job command, you see this message: # hadoop job DEPRECATED: Use of this script to execute mapred command is deprecated. Instead, use the mapred command for it. The syntax for the mapred command is: mapred [--config confdir] COMMAND Commands used with mapred include: Command Description historyserver Runs job history servers as a standalone daemon hsadmin The job history server admin interface job Manipulates MapReduce jobs pipes Runs a pipes job queue Gets information regarding JobQueues Unsupported Commands MapR does not support the following Hadoop commands: balancer datanode dfsadmin fsck fetchdt namenode secondarynamenode Generic Options Implement the Tool interface to make the following command-line options available for many of the Hadoop commands. Generic options are supported by the distcp, fs, job, mradmin, pipes, and queue Hadoop commands. Generic Option Description -conf Add the specified configuration files to the list of resources available in the configuration. -D Set a value for the specified Hadoop configuration property. -fs Set the URI of the default filesystem. -jt Specify a jobtracker for a given host and port. This command option is a shortcut for -D mapred.job.tracker=host:port -files Specify files to be copied to the map reduce cluster. -libjars Specify JAR files to be included in the classpath of the mapper and reducer tasks. -archives Specify archive files (JAR, tar, tar.gz, ZIP) to be copied and unarchived on the task node. hadoop archive The hadoop archive command creates a Hadoop archive, a file that contains other files. A Hadoop archive always has a *.har extension. Syntax hadoop [ Generic Options ] archive -archiveName [-p ] Parameters Parameter Description -archiveName Name of the archive to be created. -p The parent argument is to specify the relative path to which the files should be archived to. Filesystem pathnames which work as usual with regular expressions. Destination directory which would contain the archive. Examples Archive within a single directory hadoop archive -archiveName myArchive.har -p /foo/bar /outputdir The above command creates an archive of the directory /foo/bar in the directory /outputdir. Archive to another directory hadoop archive -archiveName myArchive.har -p /foo/bar a/b/c e/f/g The above command creates an archive of the directory /foo/bar/a/b/c in the directory /foo/bar/e/f/g. hadoop classpath The hadoop classpath command prints the class path needed to access the Hadoop jar and the required libraries. Syntax hadoop classpath Output $ hadoop classpath /opt/mapr/hadoop/hadoop-0.20.2/bin/../conf:/usr/lib/jvm/java-6-sun/lib/tools.jar:/opt/ mapr/hado op/hadoop-0.20.2/bin/..:/opt/mapr/hadoop/hadoop-0.20.2/bin/../hadoop*core*.jar:/opt/ma pr/hadoop /hadoop-0.20.2/bin/../lib/aspectjrt-1.6.5.jar:/opt/mapr/hadoop/hadoop-0.20.2/bin/../li b/aspectj tools-1.6.5.jar:/opt/mapr/hadoop/hadoop-0.20.2/bin/../lib/commons-cli-1.2.jar:/opt/map r/hadoop/ hadoop-0.20.2/bin/../lib/commons-codec-1.4.jar:/opt/mapr/hadoop/hadoop-0.20.2/bin/../l ib/common s-daemon-1.0.1.jar:/opt/mapr/hadoop/hadoop-0.20.2/bin/../lib/commons-el-1.0.jar:/opt/m apr/hadoo p/hadoop-0.20.2/bin/../lib/commons-httpclient-3.0.1.jar:/opt/mapr/hadoop/hadoop-0.20.2 /bin/../l ib/commons-logging-1.0.4.jar:/opt/mapr/hadoop/hadoop-0.20.2/bin/../lib/commons-logging -api-1.0. 4.jar:/opt/mapr/hadoop/hadoop-0.20.2/bin/../lib/commons-net-1.4.1.jar:/opt/mapr/hadoop /hadoop-0 .20.2/bin/../lib/core-3.1.1.jar:/opt/mapr/hadoop/hadoop-0.20.2/bin/../lib/eval-0.5.jar :/opt/map r/hadoop/hadoop-0.20.2/bin/../lib/hadoop-0.20.2-dev-capacity-scheduler.jar:/opt/mapr/h adoop/had oop-0.20.2/bin/../lib/hadoop-0.20.2-dev-core.jar:/opt/mapr/hadoop/hadoop-0.20.2/bin/.. /lib/hado op-0.20.2-dev-fairscheduler.jar:/opt/mapr/hadoop/hadoop-0.20.2/bin/../lib/hsqldb-1.8.0 .10.jar:/ opt/mapr/hadoop/hadoop-0.20.2/bin/../lib/jackson-core-asl-1.5.2.jar:/opt/mapr/hadoop/h adoop-0.2 0.2/bin/../lib/jackson-mapper-asl-1.5.2.jar:/opt/mapr/hadoop/hadoop-0.20.2/bin/../lib/ jasper-co mpiler-5.5.12.jar:/opt/mapr/hadoop/hadoop-0.20.2/bin/../lib/jasper-runtime-5.5.12.jar: /opt/mapr /hadoop/hadoop-0.20.2/bin/../lib/jets3t-0.6.1.jar:/opt/mapr/hadoop/hadoop-0.20.2/bin/. ./lib/jet ty-6.1.14.jar:/opt/mapr/hadoop/hadoop-0.20.2/bin/../lib/jetty-servlet-tester-6.1.14.ja r:/opt/ma pr/hadoop/hadoop-0.20.2/bin/../lib/jetty-util-6.1.14.jar:/opt/mapr/hadoop/hadoop-0.20. 2/bin/../ lib/junit-4.5.jar:/opt/mapr/hadoop/hadoop-0.20.2/bin/../lib/kfs-0.2.2.jar:/opt/mapr/ha doop/hado op-0.20.2/bin/../lib/log4j-1.2.15.jar:/opt/mapr/hadoop/hadoop-0.20.2/bin/../lib/loggin g-0.1.jar :/opt/mapr/hadoop/hadoop-0.20.2/bin/../lib/maprfs-0.1.jar:/opt/mapr/hadoop/hadoop-0.20 .2/bin/.. /lib/maprfs-test-0.1.jar:/opt/mapr/hadoop/hadoop-0.20.2/bin/../lib/mockito-all-1.8.2.j ar:/opt/m apr/hadoop/hadoop-0.20.2/bin/../lib/mysql-connector-java-5.0.8-bin.jar:/opt/mapr/hadoo p/hadoop0.20.2/bin/../lib/oro-2.0.8.jar:/opt/mapr/hadoop/hadoop-0.20.2/bin/../lib/servlet-api2.5-6.1.1 4.jar:/opt/mapr/hadoop/hadoop-0.20.2/bin/../lib/slf4j-api-1.4.3.jar:/opt/mapr/hadoop/h adoop-0.2 0.2/bin/../lib/slf4j-log4j12-1.4.3.jar:/opt/mapr/hadoop/hadoop-0.20.2/bin/../lib/xmlen c-0.52.ja r:/opt/mapr/hadoop/hadoop-0.20.2/bin/../lib/zookeeper-3.3.2.jar:/opt/mapr/hadoop/hadoo p-0.20.2/ bin/../lib/jsp-2.1/jsp-2.1.jar:/opt/mapr/hadoop/hadoop-0.20.2/bin/../lib/jsp-2.1/jsp-a pi-2.1.ja r hadoop daemonlog The hadoop daemonlog command gets and sets the log level for each daemon. Hadoop daemons all produce logfiles that you can use to learn about what is happening on the system. You can use the hadoop daemonlog co mmand to temporarily change the log level of a component when debugging the system. Syntax hadoop daemonlog -getlevel | -setlevel : [ ] Parameters The following command options are supported for hadoop daemonlog command: Parameter Description -getlevel Prints the log level of the daemon running at the specified host and port, by querying http://:/logLevel?log= : The host on which to get the log level. : The port by which to get the log level. : The daemon on which to get the log level. Usually the fully qualified classname of the daemon doing the logging. For example, org.apache.hadoop.mapred.JobTracker for the JobTracker daemon. -setlevel Sets the log level of the daemon running at the specified host and port, by querying http://:/logLevel?log= * : The host on which to set the log level. : The port by which to set the log level. : The daemon on which to set the log level. ] [-filelimit ] [-sizelimit ] [-delete ] Parameters Command Options The following command options are supported for the hadoop distcp command: Parameter Description Specify the source URL. Specify the destination URL. -p [rbugp] Preserve r: replication number b: block size u: user g: group p: permission -p alone is equivalent to -prbugp. Modification times are not preserved. When you specify -update, status updates are not synchronized unless the file sizes also differ. -i Ignore failures. As explained in the below, this option will keep more accurate statistics about the copy than the default case. It also preserves logs from failed copies, which can be valuable for debugging. Finally, a failing map will not cause the job to fail before all splits are attempted. -log Write logs to . The hadoop distcp command keeps logs of each file it attempts to copy as map output. If a map fails, the log output will not be retained if it is re-executed. -m Maximum number of simultaneous copies. Specify the number of maps to copy data. Note that more maps may not necessarily improve throughput. See Map Sizing. -overwrite Overwrite destination. If a map fails and -i is not specified, all the files in the split, not only those that failed, will be recopied. As discussed in the Overwriting Files Between Clusters, it also changes the semantics for generating destination paths, so users should use this carefully. -update Overwrite if size is different from size. As noted in the preceding, this is not a "sync" operation. The only criterion examined is the source and destination file sizes; if they differ, the source file replaces the destination file. See Updating Files Between Clusters -f Use list at as source list. This is equivalent to listing each source on the command line. The value of must be a fully qualified URI. -filelimit Limit the total number of files to be <= n. See Symbolic Representations. -sizelimit Limit the total size to be <= n bytes. See Symbolic Representations. -delete Delete the files existing in the but not in . The deletion is done by FS Shell. Generic Options The hadoop distcp command supports the following generic options: -conf , -D , -fs , -jt , -files , -libjars , and -archives . For more information on generic options, see Generic Options. Symbolic Representations The parameter in -filelimit and -sizelimit can be specified with symbolic representation. For example, 1230k = 1230 * 1024 = 1259520 891g = 891 * 1024^3 = 956703965184 Map Sizing The hadoop distcp command attempts to size each map comparably so that each copies roughly the same number of bytes. Note that files are the finest level of granularity, so increasing the number of simultaneous copiers (i.e. maps) may not always increase the number of simultaneous copies nor the overall throughput. If -m is not specified, distcp will attempt to schedule work for min (total_bytes / bytes.per.map, 20 * num_task_trackers) wher e bytes.per.map defaults to 256MB. Tuning the number of maps to the size of the source and destination clusters, the size of the copy, and the available bandwidth is recommended for long-running and regularly run jobs. Examples Basic inter-cluster copying The hadoop distcp commmand is most often used to copy files between clusters: hadoop distcp maprfs:///mapr/cluster1/foo \ maprfs:///mapr/cluster2/bar The command in the example expands the namespace under /foo/bar on cluster1 into a temporary file, partitions its contents among a set of map tasks, and starts a copy on each TaskTracker from cluster1 to cluster2. Note that the hadoop distcp command expects absolute paths. Only those files that do not already exist in the destination are copied over from the source directory. Updating files between clusters Use the hadoop distcp -update command to synchronize changes between clusters. $ hadoop distcp -update maprfs:///mapr/cluster1/foo maprfs:///mapr/cluster2/bar/foo Files in the /foo subtree are copied from cluster1 to cluster2 only if the size of the source file is different from that of the size of the destination file. Otherwise, the files are skipped over. Note that using the -update option changes distributed copy interprets the source and destination paths making it necessary to add the trailing / foo subdirectory in the second cluster. Overwriting files between clusters By default, distributed copy skips files that already exist in the destination directory, but you can overwrite those files using the -overwrite optio n. In this example, multiple source directories are specified: $ hadoop distcp -overwrite maprfs:///mapr/cluster1/foo/a \ maprfs:///mapr/cluster1/foo/b \ maprfs:///mapr/cluster2/bar As with using the -update option, using the -overwrite changes the way that the source and destination paths are interpreted by distributed copy: the contents of the source directories are compared to the contents of the destination directory. The distributed copy aborts in case of a conflict. Migrating Data from HDFS to MapR-FS The hadoop distcp command can be used to migrate data from an HDFS cluster to a MapR-FS where the HDFS cluster uses the same version of the RPC protocol as that used by MapR. For a discussion, see Copying Data from Apache Hadoop. $ hadoop distcp namenode1:50070/foo maprfs:///bar You must specify the IP address and HTTP port (usually 50070) for the namenode on the HDFS cluster. hadoop fs The hadoop fs command runs a generic filesystem user client that interacts with the MapR filesystem (MapR-FS). On the Windows client, make sure that the PATH contains the following directories: C:\Windows\system32 C:\Windows If they are not present, the hadoop fs command might fail silently. Syntax hadoop [ Generic Options ] fs [-cat ] [-chgrp [-R] GROUP PATH...] [-chmod [-R] PATH...] [-chown [-R] [OWNER][:[GROUP]] PATH...] [-copyFromLocal ... ] [-copyToLocal [-ignoreCrc] [-crc] ] [-count[-q] ] [-cp ] [-df ] [-du ] [-dus ] [-expunge] [-get [-ignoreCrc] [-crc] [-getmerge [addnl]] [-help [cmd]] [-ls ] [-lsr ] [-mkdir ] [-moveFromLocal ... ] [-moveToLocal ] [-mv ] [-put ... ] [-rm [-skipTrash] ] [-rmr [-skipTrash] ] [-stat [format] ] [-tail [-f] ] [-test -[ezd] ] [-text ] [-touchz ] Parameters Command Options The following command parameters are supported for hadoop fs: Parameter Description -cat Fetch all files that match the file pattern defined by the parameter and display their contents on stdout. -fs [local | ] Specify the file system to use. If not specified, the current configuration is used, taken from the following, in increasing precedence: core-default.xml inside the hadoop jar file core-site.xml in $HADOOP_CONF_DIR The local option means use the local file system as your DFS. specifies a particular file system to contact. This argument is optional but if used must appear appear first on the command line. Exactly one additional argument must be specified. -ls List the contents that match the specified file pattern. If path is not specified, the contents of /user/ will be listed. Directory entries are of the form dirName (full path) and file entries are of the form fileName(full path) . size where n is the number of replicas specified for the file and size is the size of the file, in bytes. -lsr Recursively list the contents that match the specified file pattern. Behaves very similarly to hadoop fs -ls, except that the data is shown for all the entries in the subtree. -df [] Shows the capacity, free and used space of the filesystem. If the filesystem has multiple partitions, and no path to a particular partition is specified, then the status of the root partitions will be shown. -du Show the amount of space, in bytes, used by the files that match the specified file pattern. Equivalent to the Unix command du -sb /* in case of a directory, and to du -b in case of a file. The output is in the form name(full path) size (in bytes). -dus Show the amount of space, in bytes, used by the files that match the specified file pattern. Equivalent to the Unix command du -sb. The output is in the form name(full path) size (in bytes). -mv Move files that match the specified file pattern to a destination . When moving multiple files, the destination must be a directory. -cp Copy files that match the file pattern to a destination. When copying multiple files, the destination must be a directory. -rm [-skipTrash] Delete all files that match the specified file pattern. Equivalent to the Unix command rm . The -skipTrash option bypasses trash, if enabled, and immediately deletes -rmr [-skipTrash] Remove all directories which match the specified file pattern. Equivalent to the Unix command rm -rf The -skipTrash option bypasses trash, if enabled, and immediately deletes -put ... Copy files from the local file system into fs. -copyFromLocal ... Identical to the -put command. -moveFromLocal ... Same as -put, except that the source is deleted after it's copied. -get [-ignoreCrc] [-crc] Copy files that match the file pattern to the local name. is kept. When copying multiple files, the destination must be a directory. -getmerge Get all the files in the directories that match the source file pattern and merge and sort them to only one file on local fs. is kept. -copyToLocal [-ignoreCrc] [-crc] Identical to the -get command. -moveToLocal Not implemented yet -mkdir Create a directory in specified location. -tail [-f] Show the last 1KB of the file. The -f option shows appended data as the file grows. -touchz Write a timestamp in yyyy-MM-dd HH:mm:ss format in a file at . An error is returned if the file exists with non-zero length. -test -[ezd] If file { exists, has zero length, is a directory then return 0, else return 1. -text Takes a source file and outputs the file in text format. The allowed formats are zip and TextRecordInputStream. -stat [format] Print statistics about the file/directory at in the specified format. Format accepts filesize in blocks (%b), filename (%n), block size (%o), replication (%r), modification date (%y, %Y) -chmod [-R] PATH... Changes permissions of a file. This works similar to shell's chmod with a few exceptions. -R modifies the files recursively. This is the only option currently supported. MODE Mode is same as mode used for chmod shell command. Only letters recognized are rwxXt. That is, +t,a+r,g-w,+rwx,o=r OCTALMODE Mode specifed in 3 or 4 digits. If 4 digits, the first may be 1 or 0 to turn the sticky bit on or off, respectively. Unlike shell command, it is not possible to specify only part of the mode E.g. 754 is same as u=rwx,g=rx,o=r If none of 'augo' is specified, 'a' is assumed and unlike shell command, no umask is applied. -chown [-R] [OWNER][:[GROUP]] PATH... Changes owner and group of a file. This is similar to shell's chown with a few exceptions. -R modifies the files recursively. This is the only option currently supported. If only owner or group is specified then only owner or group is modified.The owner and group names may only consists of digits, alphabet, and any of -.@/' i.e. [-.@/a-zA-Z0-9]. The names are case sensitive. Warning WARNING: Avoid using '.' to separate user name and group though Linux allows it. If user names have dots in them and you are using local file system, you might see surprising results since shell command chown is used for local files. -chgrp [-R] GROUP PATH... This is equivalent to -chown ... :GROUP ... -count[-q] Count the number of directories, files and bytes under the paths that match the specified file pattern. The output columns are: DIR_COUNT FILE_COUNT CONTENT_SIZE FILE_NAME or QUOTA REMAINING_QUATA SPACE_QUOTA REMAINING_SPACE_QUOTA DIR_COUNT FILE_COUNT CONTENT_SIZE FILE_NAME -help [cmd] Displays help for given command or all commands if none is specified. Generic Options The following generic options are supported for the hadoop fs command: -conf , -D , -fs , -jt , -files , -libjars , and -archives . For more information on generic options, see Generic Options. hadoop jar The hadoop jar command runs a program contained in a JAR file. Users can bundle their MapReduce code in a JAR file and execute it using this command. Syntax hadoop jar [] Parameters The following commands parameters are supported for hadoop jar: Parameter Description The JAR file. Arguments to the program specified in the JAR file. Examples Streaming Jobs Hadoop streaming jobs are run using the hadoop jar command. The Hadoop streaming utility enables you to create and run MapReduce jobs with any executable or script as the mapper and/or the reducer. $ hadoop jar $HADOOP_HOME/hadoop-streaming.jar \ -input myInputDirs \ -output myOutputDir \ -mapper org.apache.hadoop.mapred.lib.IdentityMapper \ -reducer /bin/wc The -input, -output, -mapper, and -reducer streaming command options are all required for streaming jobs. Either an executable or a Java class may be used for the mapper and the reducer. For more information about and examples of streaming jobs, see Streaming Options and Usage at the Apache project's page. Running from a JAR file The simple Word Count program is another example of a program that is run using the hadoop jar command. The wordcount functionality is built into the hadoop-0.20.2-dev-examples.jar program. You pass the file, along with the location, to Hadoop with the hadoop jar comm and and Hadoop reads the JAR file and executes the relevant instructions. The Word Count program reads files from an input directory, counts the words, and writes the results of the job to files in an output directory. $ hadoop jar /opt/mapr/hadoop/hadoop-0.20.2/hadoop-0.20.2-dev-examples.jar wordcount /myvolume/in /myvolume/out hadoop job The hadoop job command enables you to manage MapReduce jobs. Syntax hadoop job [Generic Options] [-submit ] [-status ] [-counter ] [-kill ] [-unblacklist ] [-unblacklist-tracker ] [-set-priority ] [-events <#-of-events>] [-history ] [-list [all]] [-list-active-trackers] [-list-blacklisted-trackers] [-list-attempt-ids ] [-kill-task ] [-fail-task ] [-blacklist-tasktracker ] [-showlabels] Parameters Command Options The following command options are supported for hadoop job: Parameter Description -submit Submits the job. -status Prints the map and reduce completion percentage and all job counters. -counter Prints the counter value. -kill Kills the job. -unblacklist Removes a tasktracker job from the jobtracker's blacklist. -unblacklist-tracker Admin only. Removes the TaskTracker at Changes the priority of the job. Valid priority values are VERY_HIGH, HIGH, NORMAL, LOW, and VERY_LOW. The job scheduler uses this property to determine the order in which jobs are run. -events <#-of-events> Prints the events' details received by jobtracker for the given range. -history Prints job details, failed and killed tip details. -list [all] The -list all option displays all jobs. The -list command without the all option displays only jobs which are yet to complete. -list-active-trackers Prints all active tasktrackers. -list-blackisted-trackers Prints blacklisted tasktrackers. -list-attempt-ids Lists the IDs of task attempts. -kill-task Kills the task. Killed tasks are not counted against failed attempts. -fail-task Fails the task. Failed tasks are counted against failed attempts. -blacklist-tasktracker Pauses all current tasktracker jobs and prevent additional jobs from being scheduled on the tasktracker. -showlabels Dumps label information of all active nodes. Generic Options The following generic options are supported for the hadoop job command: -conf , -D , -fs , -jt , -files , -libjars , and -archives . For more information on generic options, see Generic Options. Examples Submitting Jobs The hadoop job -submit command enables you to submit a job to the specified jobtracker. $ hadoop job -jt darwin:50020 -submit job.xml Stopping Jobs Gracefully Use the hadoop kill command to stop a running or queued job. $ hadoop job -kill Viewing Job History Logs Run the hadoop job -history command to view the history logs summary in specified directory. $ hadoop job -history output-dir This command will print job details, failed and killed tip details. Additional details about the job such as successful tasks and task attempts made for each task can be viewed by adding the -all option: $ hadoop job -history all output-dir Blacklisting Tasktrackers The hadoop job command when run as root or using sudo can be used to manually blacklist tasktrackers: hadoop job -blacklist-tasktracker Manually blacklisting a tasktracker pauses any running jobs and prevents additional jobs from being scheduled. For a detailed discussion see TaskTracker Blacklisting. hadoop jobtracker The hadoop jobtracker command runs the MapReduce jobtracker node. Syntax hadoop jobtracker [-dumpConfiguration] Parameters The hadoop jobtracker command supports the following command options: Parameter Description -dumpConfiguration Dumps the configuration used by the jobtracker along with queue configuration in JSON format into standard output used by the jobtracker and exits. hadoop mfs The hadoop mfs command performs operations on directories in the cluster. The main purposes of hadoop mfs are to display directory information and contents, to create symbolic links, and to set compression and chunk size on a directory. Syntax hadoop mfs [ -ln ] [ -ls ] [ -lsd ] [ -lsr ] [ -Lsr ] [ -lsrv ] [ -lss ] [ -setcompression on|off|lzf|lz4|zlib ] [ -setchunksize ] [ -setnetworkencryption on|off ] [ -help ] Parameters The normal command syntax is to specify a single option from the following table, along with its corresponding arguments. If compression and chunk size are not set explicitly for a given directory, the values are inherited from the parent directory. Parameter Description -ln Creates a symbolic link that points to the target path , similar to the standard Linux ln -s command. -ls Lists files in the directory specified by . The hadoop mfs -ls command corresponds to the standard h adoop fs -ls command, but provides the following additional information: Chunks used for each file Server where each chunk resides Whether compression is enabled for each file Whether encryption is enabled for each file -lsd Lists files in the directory specified by , and also provides information about the specified directory itself: Whether compression is enabled for the directory (indicated by z ) The configured chunk size (in bytes) for the directory. -lsr Lists files in the directory and subdirectories specified by , recursively, including dereferencing symbolic links. The hadoop mfs -lsr command corresponds to the standard hadoop fs -lsr command, but provides the following additional information: Chunks used for each file Server where each chunk resides -Lsr Equivalent to lsr, but additionally dereferences symbolic links -lsrv Lists all paths recursively without crossing volume links. -lss Lists files in the directory specified by , with an additional column that displays the number of disk blocks per file. Disk blocks are 8192 bytes. -setcompression on|off|lzf|lz4|zlib Turns compression on or off on the directory specified in , and sets the compression type: on — turns on compression using the default algorithm (LZ4) off — turns off compression lzf — turns on compression and sets the algorithm to LZF lz4 — turns on compression and sets the algorithm to LZ4 zlib — turns on compression and sets the algorithm to ZLIB -setchunksize Sets the chunk size in bytes for the directory specified in . The parameter must be a multiple of 65536. -setnetworkencryption on|off Sets network encryption on or off for the filesystem object defined in . The cluster encrypts network target to or from a file, directory, or MapR table with network security enabled. -help Displays help for the hadoop mfs command. Examples The hadoop mfs command is used to view file contents. You can use this command to check if compression is turned off in a directory or mounted volume. For example, # hadoop mfs -ls / Found 23 items vrwxr-xr-x Z - root root 13 2012-04-29 10:24 268435456 /.rw p mapr.cluster.root writeable 2049.35.16584 -> 2049.16.2 scale-50.scale.lab:5660 scale-51.scale.lab:5660 scale-52.scale.lab:5660 vrwxr-xr-x U - root root 7 2012-04-28 22:16 67108864 /hbase p mapr.hbase default 2049.32.16578 -> 2050.16.2 scale-50.scale.lab:5660 scale-51.scale.lab:5660 scale-52.scale.lab:5660 drwxr-xr-x Z - root root 0 2012-04-29 09:14 268435456 /tmp p 2049.41.16596 scale-50.scale.lab:5660 scale-51.scale.lab:5660 scale-52.scale.lab:5660 vrwxr-xr-x Z - root root 1 2012-04-27 22:59 268435456 /user p users default 2049.36.16586 -> 2055.16.2 scale-50.scale.lab:5660 scale-52.scale.lab:5660 scale-51.scale.lab:5660 drwxr-xr-x Z - root root 1 2012-04-27 22:37 268435456 /var p 2049.33.16580 scale-50.scale.lab:5660 scale-51.scale.lab:5660 scale-52.scale.lab:5660 In the above example, the letter Z indicates LZ4 compression on the directory; the letter U indicates that the directory is uncompressed. In the following example, the listed item is both uncompressed (first U) and unencrypted (second U). [root@node1-302 ~]# hadoop mfs -ls /hbase Found 10 items drwxr-xr-x U U - mapr mapr 3 2014-05-28 12:05 67108864 /hbase/-ROOTp 2050.34.3674200 node2-302:5660 node1-302:5660 node3-302:5660 ... Output When used with -ls, -lsd, -lsr, or -lss, hadoop mfs displays information about files and directories. For each file or directory hadoop mfs displays a line of basic information followed by lines listing the chunks that make up the file, in the following format: {mode} {compression} {encryption} {replication} {owner} {group} {size} {date} {chunk size} {name} {chunk} {fid} {host} [{host}...] {chunk} {fid} {host} [{host}...] ... Volume links are displayed as follows: {mode} {compression} {encryption} {replication} {owner} {group} {size} {date} {chunk size} {name} {chunk} {target volume name} {writability} {fid} -> {fid} [{host}...] For volume links, the first fid is the chunk that stores the volume link itself; the fid after the arrow (->) is the first chunk in the target volume. The following table describes the values: mode A text string indicating the read, write, and execute permissions for the owner, group, and other permissions. See also Mana ging Permissions. compression U: uncompressed L: LZf Z (Uppercase): LZ4 z (Lowercase): ZLIB encryption U: unencrypted; E: encrypted replication The replication factor of the file (directories display a dash instead) owner The owner of the file or directory group The group of the file of directory size The size of the file or directory date The date the file or directory was last modified chunk size The chunk size of the file or directory name The name of the file or directory chunk The chunk number. The first chunk is a primary chunk labeled "p", a 64K chunk containing the root of the file. Subsequent chunks are numbered in order. fid The chunk's file ID, which consists of three parts: The ID of the container where the file is stored The inode of the file within the container An internal version number host The host on which the chunk resides. When several hosts are listed, the first host is the first copy of the chunk and subsequent hosts are replicas. target volume name The name of the volume pointed to by a volume link. writability Displays whether the volume is writable. hadoop mradmin The hadoop mradmin command runs Map-Reduce administrative commands. Syntax hadoop [ Generic Options ] mradmin [-refreshServiceAcl] [-refreshQueues] [-refreshNodes] [-refreshUserToGroupsMappings] [-refreshSuperUserGroupsConfiguration] [-help [cmd]] Parameters The following command parameters are supported for hadoop mradmin: Parameter Description -refreshServiceAcl Reload the service-level authorization policy file Job tracker will reload the authorization policy file. -refreshQueues Reload the queue acls and state JobTracker will reload the mapred-queues.xml file. -refreshUserToGroupsMappings Refresh user-to-groups mappings. -refreshSuperUserGroupsConfiguration Refresh superuser proxy groups mappings. -refreshNodes Refresh the hosts information at the job tracker. -help [cmd] Displays help for the given command or all commands if none is specified. The following generic options are supported for hadoop mradmin: Generic Option Description -conf Specify an application configuration file. -D Use value for given property. -fs Specify a file system. -jt Specify a job tracker. -files Specify comma separated files to be copied to the map reduce cluster. -libjars Specify comma separated jar files to include in the classpath. -archives Specify comma separated archives to be unarchived on the computer machines. hadoop pipes The hadoop pipes command runs a pipes job. Hadoop Pipes is the C++ interface to Hadoop Reduce. Hadoop Pipes uses sockets to enable tasktrackers to communicate processes running the C++ map or reduce functions. See also Compiling Pipes Programs. Syntax hadoop [GENERIC OPTIONS ] pipes [-output ] [-jar ] [-inputformat ] [-map ] [-partitioner ] [-reduce ] [-writer ] [-program ] [-reduces ] Parameters Command Options The following command parameters are supported for hadoop pipes: Parameter Description -output Specify the output directory. -jar Specify the jar filename. -inputformat InputFormat class. -map Specify the Java Map class. -partitioner Specify the Java Partitioner. -reduce Specify the Java Reduce class. -writer Specify the Java RecordWriter. -program Specify the URI of the executable. -reduces Specify the number of reduces. Generic Options The following generic options are supported for the hadoop pipes command: -conf , -D , -f s , -jt , -files , -libjars , and -archives . For more information on generic options, see Generic Options. hadoop queue The hadoop queue command displays job queue information. Syntax hadoop [ Generic Options ] queue [-list] | [-info [-showJobs]] | [-showacls] Parameters Command Options The hadoop queue command supports the following command options: Parameter Description -list Gets list of job queues configured in the system. Along with scheduling information associated with the job queues. -info [-showJobs] Displays the job queue information and associated scheduling information of particular job queue. If -showJobs o ption is present, a list of jobs submitted to the particular job queue is displayed. -showacls Displays the queue name and associated queue operations allowed for the current user. The list consists of only those queues to which the user has access. Generic Options The following generic options are supported for the hadoop queue command: -conf , -D , -f s , -jt , -files , -libjars , and -archives . For more information on generic options, see Generic Options. hadoop tasktracker The hadoop tasktracker command runs a MapReduce tasktracker node. When you change the name of a node, you must restart the tasktracker. Syntax hadoop tasktracker Output mapr@mapr-desktop:~$ hadoop tasktracker 12/03/21 21:19:56 INFO mapred.TaskTracker: STARTUP_MSG: /************************************************************ STARTUP_MSG: Starting TaskTracker STARTUP_MSG: host = mapr-desktop/127.0.1.1 STARTUP_MSG: args = [] STARTUP_MSG: version = 0.20.2-dev STARTUP_MSG: build = -r ; compiled by 'root' on Thu Dec 8 22:43:13 PST 2011 ************************************************************/ 12/03/21 21:19:56 INFO mapred.TaskTracker: /*-------------- TaskTracker System Properties ---------------- java.runtime.name: Java(TM) SE Runtime Environment sun.boot.library.path: /usr/lib/jvm/java-6-sun-1.6.0.26/jre/lib/amd64 java.vm.version: 20.1-b02 hadoop.root.logger: INFO,console java.vm.vendor: Sun Microsystems Inc. java.vendor.url: http://java.sun.com/ path.separator: : java.vm.name: Java HotSpot(TM) 64-Bit Server VM file.encoding.pkg: sun.io sun.java.launcher: SUN_STANDARD user.country: US sun.os.patch.level: unknown java.vm.specification.name: Java Virtual Machine Specification user.dir: /home/mapr java.runtime.version: 1.6.0_26-b03 java.awt.graphicsenv: sun.awt.X11GraphicsEnvironment java.endorsed.dirs: /usr/lib/jvm/java-6-sun-1.6.0.26/jre/lib/endorsed os.arch: amd64 java.io.tmpdir: /tmp line.separator: hadoop.log.file: hadoop.log java.vm.specification.vendor: Sun Microsystems Inc. os.name: Linux hadoop.id.str: sun.jnu.encoding: UTF-8 java.library.path: /opt/mapr/hadoop/hadoop-0.20.2/bin/../lib/native/Linux-amd64-64: hadoop.home.dir: /opt/mapr/hadoop/hadoop-0.20.2/bin/.. java.specification.name: Java Platform API Specification java.class.version: 50.0 sun.management.compiler: HotSpot 64-Bit Tiered Compilers hadoop.pid.dir: /opt/mapr/hadoop/hadoop-0.20.2/bin/../pids os.version: 2.6.32-33-generic user.home: /home/mapr user.timezone: America/Los_Angeles java.awt.printerjob: sun.print.PSPrinterJob file.encoding: UTF-8 java.specification.version: 1.6 java.class.path: /opt/mapr/hadoop/hadoop-0.20.2/bin/../conf:/usr/lib/jvm/java-6-sun-1.6.0.26/lib/tools. jar:/opt/mapr/hadoop/hadoop-0.20.2/bin/..:/opt/mapr/hadoop/hadoop-0.20.2/bin/../hadoop *core*.jar:/opt/mapr/hadoop/hadoop-0.20.2/bin/../lib/aspectjrt-1.6.5.jar:/opt/mapr/had oop/hadoop-0.20.2/bin/../lib/aspectjtools-1.6.5.jar:/opt/mapr/hadoop/hadoop-0.20.2/bin /../lib/commons-cli-1.2.jar:/opt/mapr/hadoop/hadoop-0.20.2/bin/../lib/commons-codec-1. 4.jar:/opt/mapr/hadoop/hadoop-0.20.2/bin/../lib/commons-daemon-1.0.1.jar:/opt/mapr/had oop/hadoop-0.20.2/bin/../lib/commons-el-1.0.jar:/opt/mapr/hadoop/hadoop-0.20.2/bin/../ lib/commons-httpclient-3.0.1.jar:/opt/mapr/hadoop/hadoop-0.20.2/bin/../lib/commons-log ging-1.0.4.jar:/opt/mapr/hadoop/hadoop-0.20.2/bin/../lib/commons-logging-api-1.0.4.jar :/opt/mapr/hadoop/hadoop-0.20.2/bin/../lib/commons-net-1.4.1.jar:/opt/mapr/hadoop/hado op-0.20.2/bin/../lib/core-3.1.1.jar:/opt/mapr/hadoop/hadoop-0.20.2/bin/../lib/eval-0.5 .jar:/opt/mapr/hadoop/hadoop-0.20.2/bin/../lib/hadoop-0.20.2-dev-capacity-scheduler.ja r:/opt/mapr/hadoop/hadoop-0.20.2/bin/../lib/hadoop-0.20.2-dev-core.jar:/opt/mapr/hadoo p/hadoop-0.20.2/bin/../lib/hadoop-0.20.2-dev-fairscheduler.jar:/opt/mapr/hadoop/hadoop -0.20.2/bin/../lib/hsqldb-1.8.0.10.jar:/opt/mapr/hadoop/hadoop-0.20.2/bin/../lib/jacks on-core-asl-1.5.2.jar:/opt/mapr/hadoop/hadoop-0.20.2/bin/../lib/jackson-mapper-asl-1.5 .2.jar:/opt/mapr/hadoop/hadoop-0.20.2/bin/../lib/jasper-compiler-5.5.12.jar:/opt/mapr/ hadoop/hadoop-0.20.2/bin/../lib/jasper-runtime-5.5.12.jar:/opt/mapr/hadoop/hadoop-0.20 .2/bin/../lib/jets3t-0.6.1.jar:/opt/mapr/hadoop/hadoop-0.20.2/bin/../lib/jetty-6.1.14. jar:/opt/mapr/hadoop/hadoop-0.20.2/bin/../lib/jetty-servlet-tester-6.1.14.jar:/opt/map r/hadoop/hadoop-0.20.2/bin/../lib/jetty-util-6.1.14.jar:/opt/mapr/hadoop/hadoop-0.20.2 /bin/../lib/junit-4.5.jar:/opt/mapr/hadoop/hadoop-0.20.2/bin/../lib/kfs-0.2.2.jar:/opt /mapr/hadoop/hadoop-0.20.2/bin/../lib/log4j-1.2.15.jar:/opt/mapr/hadoop/hadoop-0.20.2/ bin/../lib/logging-0.1.jar:/opt/mapr/hadoop/hadoop-0.20.2/bin/../lib/maprfs-0.1.jar:/o pt/mapr/hadoop/hadoop-0.20.2/bin/../lib/maprfs-test-0.1.jar:/opt/mapr/hadoop/hadoop-0. 20.2/bin/../lib/mockito-all-1.8.2.jar:/opt/mapr/hadoop/hadoop-0.20.2/bin/../lib/mysqlconnector-java-5.0.8-bin.jar:/opt/mapr/hadoop/hadoop-0.20.2/bin/../lib/oro-2.0.8.jar:/ opt/mapr/hadoop/hadoop-0.20.2/bin/../lib/servlet-api-2.5-6.1.14.jar:/opt/mapr/hadoop/h adoop-0.20.2/bin/../lib/slf4j-api-1.4.3.jar:/opt/mapr/hadoop/hadoop-0.20.2/bin/../lib/ slf4j-log4j12-1.4.3.jar:/opt/mapr/hadoop/hadoop-0.20.2/bin/../lib/xmlenc-0.52.jar:/opt /mapr/hadoop/hadoop-0.20.2/bin/../lib/zookeeper-3.3.2.jar:/opt/mapr/hadoop/hadoop-0.20 .2/bin/../lib/jsp-2.1/jsp-2.1.jar:/opt/mapr/hadoop/hadoop-0.20.2/bin/../lib/jsp-2.1/js p-api-2.1.jar user.name: mapr java.vm.specification.version: 1.0 sun.java.command: org.apache.hadoop.mapred.TaskTracker java.home: /usr/lib/jvm/java-6-sun-1.6.0.26/jre sun.arch.data.model: 64 user.language: en java.specification.vendor: Sun Microsystems Inc. hadoop.log.dir: /opt/mapr/hadoop/hadoop-0.20.2/bin/../logs java.vm.info: mixed mode java.version: 1.6.0_26 java.ext.dirs: /usr/lib/jvm/java-6-sun-1.6.0.26/jre/lib/ext:/usr/java/packages/lib/ext sun.boot.class.path: /usr/lib/jvm/java-6-sun-1.6.0.26/jre/lib/resources.jar:/usr/lib/jvm/java-6-sun-1.6.0.2 6/jre/lib/rt.jar:/usr/lib/jvm/java-6-sun-1.6.0.26/jre/lib/sunrsasign.jar:/usr/lib/jvm/ java-6-sun-1.6.0.26/jre/lib/jsse.jar:/usr/lib/jvm/java-6-sun-1.6.0.26/jre/lib/jce.jar: /usr/lib/jvm/java-6-sun-1.6.0.26/jre/lib/charsets.jar:/usr/lib/jvm/java-6-sun-1.6.0.26 /jre/lib/modules/jdk.boot.jar:/usr/lib/jvm/java-6-sun-1.6.0.26/jre/classes java.vendor: Sun Microsystems Inc. file.separator: / java.vendor.url.bug: http://java.sun.com/cgi-bin/bugreport.cgi sun.io.unicode.encoding: UnicodeLittle sun.cpu.endian: little hadoop.policy.file: hadoop-policy.xml sun.desktop: gnome sun.cpu.isalist: ------------------------------------------------------------*/ 12/03/21 21:19:57 INFO mapred.TaskTracker: /tmp is not tmpfs or ramfs. Java Hotspot Instrumentation will be disabled by default 12/03/21 21:19:57 INFO mapred.TaskTracker: Cleaning up config files from the job history folder 12/03/21 21:19:57 INFO mapred.TaskTracker: TT local config is /opt/mapr/hadoop/hadoop-0.20.2/conf/mapred-site.xml 12/03/21 21:19:57 INFO mapred.TaskTracker: Loading resource properties file : /opt/mapr//logs/cpu_mem_disk 12/03/21 21:19:57 INFO mapred.TaskTracker: Physical memory reserved for mapreduce tasks = 2105540608 bytes 12/03/21 21:19:57 INFO mapred.TaskTracker: CPUS: 1 12/03/21 21:19:57 INFO mapred.TaskTracker: Total MEM: 1.9610939GB 12/03/21 21:19:57 INFO mapred.TaskTracker: Reserved MEM: 2008MB 12/03/21 21:19:57 INFO mapred.TaskTracker: Reserved MEM for Ephemeral slots 0 12/03/21 21:19:57 INFO mapred.TaskTracker: DISKS: 2 12/03/21 21:19:57 INFO mapred.TaskTracker: Map slots 1, Default heapsize for map task 873 mb 12/03/21 21:19:57 INFO mapred.TaskTracker: Reduce slots 1, Default heapsize for reduce task 1135 mb 12/03/21 21:19:57 INFO mapred.TaskTracker: Ephemeral slots 0, memory given for each ephemeral slot 200 mb 12/03/21 21:19:57 INFO mapred.TaskTracker: Prefetch map slots 1 12/03/21 21:20:07 INFO mortbay.log: Logging to org.slf4j.impl.Log4jLoggerAdapter(org.mortbay.log) via org.mortbay.log.Slf4jLog 12/03/21 21:20:08 INFO http.HttpServer: Added global filtersafety (class=org.apache.hadoop.http.HttpServer$QuotingInputFilter) 12/03/21 21:20:08 WARN mapred.TaskTracker: Error while writing to TaskController config filejava.io.FileNotFoundException: /opt/mapr/hadoop/hadoop-0.20.2/bin/../conf/taskcontroller.cfg (Permission denied) 12/03/21 21:20:08 ERROR mapred.TaskTracker: Can not start TaskTracker because java.io.IOException: Cannot run program "/opt/mapr/hadoop/hadoop-0.20.2/bin/../bin/Linux-amd64-64/bin/task-controller": java.io.IOException: error=13, Permission denied at java.lang.ProcessBuilder.start(ProcessBuilder.java:460) at org.apache.hadoop.util.Shell.runCommand(Shell.java:267) at org.apache.hadoop.util.Shell.run(Shell.java:249) at org.apache.hadoop.util.Shell$ShellCommandExecutor.execute(Shell.java:442) at org.apache.hadoop.mapred.LinuxTaskController.setup(LinuxTaskController.java:142) at org.apache.hadoop.mapred.TaskTracker.(TaskTracker.java:2149) at org.apache.hadoop.mapred.TaskTracker.main(TaskTracker.java:5216) Caused by: java.io.IOException: java.io.IOException: error=13, Permission denied at java.lang.UNIXProcess.(UNIXProcess.java:148) at java.lang.ProcessImpl.start(ProcessImpl.java:65) at java.lang.ProcessBuilder.start(ProcessBuilder.java:453) ... 6 more 12/03/21 21:20:08 INFO mapred.TaskTracker: SHUTDOWN_MSG: /************************************************************ SHUTDOWN_MSG: Shutting down TaskTracker at mapr-desktop/127.0.1.1 ************************************************************/ hadoop version The hadoop version command prints the hadoop software version. Syntax hadoop version Output mapr@mapr-desktop:~$ hadoop version Hadoop 0.20.2-dev Subversion -r Compiled by root on Thu Dec 8 22:43:13 PST 2011 From source with checksum 19fa44df0cb831c45ef984f21feb7110 hadoop conf The hadoop conf command outputs the configuration information for this node to standard output. Syntax hadoop [ generic options ] conf [ -dump ] [ -key ] Parameters Parameter Description -dump Dumps the entire configuration set to standard output. -key Displays the configured value for the specified parameter. Examples Dumping a node's entire configuration to a text file hadoop conf -dump > nodeconfiguration.txt The above command creates a text file named nodeconfiguration.txt that contains the node's configuration information. Using the tail util ity to examine the last few lines of the file displays the following information: [user@hostame:01] tail nodeconfiguration.txt mapred.merge.recordsBeforeProgress=10000 io.mapfile.bloom.error.rate=0.005 io.bytes.per.checksum=512 mapred.cluster.ephemeral.tasks.memory.limit.mb=200 mapred.fairscheduler.smalljob.max.inputsize=10737418240 ipc.client.tcpnodelay=true mapreduce.tasktracker.reserved.physicalmemory.mb.low=0.80 fs.s3.sleepTimeSeconds=10 mapred.task.tracker.report.address=127.0.0.1:0 *** MapR Configuration Dump: END *** [user@hostname:02] Displaying the configured value of a specific parameter [user@hostame:01] hadoop conf -key io.bytes.per.checksum 512 [user@hostname:02] The above command returns 512 as the configured value of the io.bytes.per.checksum parameter. Hadoop 2 Commands MapR supports the following commands in Hadoop Version 2.x. Command Description CLASSNAME The hadoop script can be used to invoke any class. hadoop CLASSNAME runs the class named CLASSNAME. conf The hadoop conf command prints the configuration information for the current node. daemonlog The hadoop daemonlog command may be used to get or set the log level of Hadoop daemons. fs The hadoop fs command runs a generic filesystem user client that interacts with the MapR filesystem (MapR-FS). jar The hadoop jar command runs a JAR file. Users can bundle their MapReduce code in a JAR file and execute it using this command. jobtracker Runs the MapReduce Jobtracker node. mfs The hadoop mfs command performs operations on directories in the cluster. The main purposes of hadoop mfs are to display directory information and contents, to create symbolic links, and to set compression and chunk size on a directory. mradmin Runs a MapReduce admin client. tasktracker The hadoop tasktracker command runs a MapReduce tasktracker node. version The hadoop version command prints the Hadoop software version. The following mapred commands are deprecated for Hadoop 2.x, and can no longer be executed by using the hadoop job command. Use the mapred command instead. Command Description archive -archiveName NAME * The hadoop archive command creates a Hadoop archive, a file that contains other files. A Hadoop archive always has a .har extension. classpath The hadoop classpath command prints the class path needed to access the Hadoop JAR and the required libraries. distcp The hadoop distcp command is a tool for large inter- and intra-cluster copying. It uses MapReduce to effect its distribution, error handling and recovery, and reporting. It expands a list of files and directories into input to map tasks, each of which will copy a partition of the files specified in the source list. historyserver Runs job history server as a standalone daemon. hsadmin The job history server admin interface. job Manipulates MapReduce jobs. pipes Runs a pipes job. queue Gets information about job queues. If you try to execute these commands by using the hadoop job command, you will see this message: # hadoop job DEPRECATED: Use of this script to execute mapred command is deprecated. Instead use the mapred command for it. Most Hadoop commands print help when invoked without parameters. Hadoop 1 Commands The following commands are supported in Hadoop Version 1.x. Command Description archive -archiveName NAME * The hadoop archive command creates a Hadoop archive, a file that contains other files. A Hadoop archive always has a .har extension. CLASSNAME The hadoop script can be used to invoke any class. hadoop CLASSNAME runs the class named CLASSNAME. classpath The hadoop classpath command prints the class path needed to access the Hadoop JAR and the required libraries. conf The hadoop conf command prints the configuration information for the current node. daemonlog The hadoop daemonlog command may be used to get or set the log level of Hadoop daemons. distcp The hadoop distcp command is a tool for large inter- and intra-cluster copying. It uses MapReduce to effect its distribution, error handling and recovery, and reporting. It expands a list of files and directories into input to map tasks, each of which will copy a partition of the files specified in the source list. fs The hadoop fs command runs a generic filesystem user client that interacts with the MapR filesystem (MapR-FS). jar The hadoop jar command runs a JAR file. Users can bundle their MapReduce code in a JAR file and execute it using this command. job Manipulates MapReduce jobs. jobtracker Runs the MapReduce Jobtracker node. mfs The hadoop mfs command performs operations on directories in the cluster. The main purposes of hadoop mfs are to display directory information and contents, to create symbolic links, and to set compression and chunk size on a directory. mradmin Runs a MapReduce admin client. pipes Runs a pipes job. queue Gets information about job queues. tasktracker The hadoop tasktracker command runs a MapReduce tasktracker node. version The hadoop version command prints the Hadoop software version. Useful Information Most Hadoop commands print help when invoked without parameters. YARN commands All YARN commands are invoked by the /usr/bin/yarn script. Usage: yarn [--config confdir] [COMMAND] [COMMAND_OPTIONS] COMMAND_OPTION Description --config confdir Overwrites the default Configuration directory. Default is ${HADOOP_HOME}/conf. COMMAND Commands COMMAND_OPTIONS Command options Commands The following yarn commands may be run on MapR distribution of Apache Hadoop: Command Description jar Runs jar file application Lists applications, or prints the status or kills the specified application. node Prints node report(s) logs Dumps container logs classpath Prints the class path needed to get the Hadoop jar and the required libraries version Print the version rmadmin Performs administrative tasks for Resource Manager daemonlog Gets and sets the log level for each daemon The following yarn commands are not supported on the MapR distribution of Apache Hadoop: yarn resourcemanager yarn nodemanager yarn proxyserver Instead, use the maprcli node services command or the MCS to start the services. For more information, see Starting, Stopping, and Restarting Services. yarn application The yarn application lists applications, or prints the status or kills the specified application. Syntax yarn application [-list [<-appStates States>] [<-appTypes Types>] ] [-status ApplicationId] [-kill ApplicationId] Parameters The following commands parameters are supported for yarn application: Parameter Description -list [<-appStates States>] [<-appTypes Types>] Lists applications. Optionally, you can filter the applications based on type or state. Use -appTypes to filer applications based on a comma-separated list of application types. Use-appStates to filter applications based on a comma-separated list of the following valid application states:ALL, NEW, NEW_SAVING, SUBMITTED, ACCEPTED, RUNNING, FINISHED, FAILED, KILLED -status ApplicationId Prints the status of the application. -kill ApplicationId Kills the application. yarn classpath The yarn classpath command prints the class path needed to access the Hadoop jar and the required libraries. Syntax yarn classpath Output $ yarn classpath /opt/mapr/hadoop/hadoop-2.4.1/etc/hadoop:/opt/mapr/hadoop/hadoop-2.4.1/etc/hadoop:/opt /mapr/hadoop/hadoop-2.4.1/etc/hadoop:/opt/mapr/hadoop/hadoop-2.4.1/share/hadoop/common /lib/*: /opt/mapr/hadoop/hadoop-2.4.1/share/hadoop/common/*: /opt/mapr/hadoop/hadoop-2.4.1/share/hadoop/hdfs: /opt/mapr/hadoop/hadoop-2.4.1/share/hadoop/hdfs/lib/*: /opt/mapr/hadoop/hadoop-2.4.1/share/hadoop/hdfs/*: /opt/mapr/hadoop/hadoop-2.4.1/share/hadoop/yarn/lib/*: /opt/mapr/hadoop/hadoop-2.4.1/share/hadoop/yarn/*: /opt/mapr/hadoop/hadoop-2.4.1/share/hadoop/mapreduce/lib/*: /opt/mapr/hadoop/hadoop-2.4.1/share/hadoop/mapreduce/*: /contrib/capacity-scheduler/*.jar: /opt/mapr/hadoop/hadoop-2.4.1/share/hadoop/yarn/*: /opt/mapr/hadoop/hadoop-2.4.1/share/hadoop/yarn/lib/* yarn daemonlog Gets or sets the log level for each daemon. Syntax yarn daemonlog [-getlevel ] Parameters | [-setlevel ] Parameter Description -getlevel Prints the log level of the daemon running at . This command internally connects to http:///logLevel?log=. -setlevel Sets the log level of the daemon running at . This command internally connects to http:///logLevel?log=. yarn jar Runs a jar file that contains YARN code. Syntax yarn jar [] [] Parameters The following commands parameters are supported for yarn jar: Parameter Description The JAR file. Sets the applications entry point. Arguments to the program specified in the JAR file. yarn logs Dumps the YARN contains logs. Syntax yarn logs -applicationID [-appowner ] [-containerId ] [-nodeAddress ] Parameters Parameter Description -applicationID Specifies an application Id. -appowner Specifies the application owner. This defaults to the current user if this option is not specified. -containerID Specifies the container ID. This is required when -nodeAddress is specified. -nodeAddress Specifies the node address in the following format: nodename:port This is required when -containerID is specified. yarn node Prints node report(s) Syntax yarn node [-lists [-states ] | [-all]] [-status NodeId] Parameters Parameter Description -list [-states ] | [-all]] Lists all running nodes. Optionally, filter nodes based on state or choose to list all the nodes. Use -states to filter nodes based on a comma-separate list of node states. Use -all to list all nodes. -status NodeId Prints the status report of the node. yarn rmadmin Runs the ResourceManager admin client. Syntax yarn rmadmin [-refreshQueues] [-refreshNodes] [-refreshUserToGroupsMapping] [-refreshSuperUserGroupsConfiguration] [-refreshAdminAcls] [-refreshServiceAcl] [-getGroups ] [-help ] [-transitionToActive ] [-transitionToStandby ] [-getServiceState ] [-checkHealth ] Parameters Parameter Description -refreshQueues Reloads the queues' acls, states, and scheduler specific properties. The ResourceManager reloads the mapred-queues configuration file. -refreshNodes Refreshes the host information at the ResourceManager. -refreshUserToGroupsMappings Refreshes user-to-groups mappings. -refreshSuperUserGroupsConfiguration Refreshes superuser proxy groups mappings. -refreshAdminAcls Refreshes acls for administration of ResourceManager. -refreshServiceAcl Reloads the service-level authorization policy file. The ResourceManager reloads the authorization policy file. -getGroups Gets the groups that the user belongs to. -help Displays help for the given parameter or all parameters if no parameter is specified. -transitionToActive Transitions the service into the Active state. -transitionToStandby Transitions the service into Standby state. -getServiceState Returns the service state. -checkHealth Requests a health check for the service. If the health check fails, the RMAdmin tool exits with a non-zero exit code. yarn version Prints the YARN version. Syntax yarn version Output $ yarn version Hadoop 2.4.1-mapr-1408-SNAPSHOT Subversion [email protected]:mapr/private-hadoop-common -r 6067f2d92ecb3202328799bc42a28ec7f6ffbff0 Compiled by root on 2014-08-27T00:00Z Compiled with protoc 2.5.0 From source with checksum 932a326cc7e285d0a94adbc10e84e41 This command was run using /opt/mapr/hadoop/hadoop-2.4.1/share/hadoop/common/hadoop-common-2.4.1-mapr-1408-SNAPSH OT.jar API Reference Overview This guide provides information about the MapR command API. Most commands can be run on the command-line interface (CLI), or by making REST requests programmatically or in a browser. To run CLI commands, use a Client machine or an ssh connection to any node in the cluster. To use the REST interface, make HTTP requests to a node that is running the WebServer service. Each command reference page includes the command syntax, a table that describes the parameters, and examples of command usage. In each parameter table, required parameters are in bold text. For output commands, the reference pages include tables that describe the output fields. Values that do not apply to particular combinations are marked NA. REST API Syntax MapR REST calls use the following format: https://:/rest/[/...]? Construct the list from the required and optional parameters, in the format = separated by the ampersand (&) character. Example: https://r1n1.qa.sj.ca.us:8443/rest/volume/mount?name=test-volume&path=/test Values in REST API calls must be URL-encoded. For readability, the values in this document use the actual characters, rather than the URL-encoded versions. Authentication To make REST calls using curl or wget, provide the username and password. To configure PAM for REST API, see PAM Configuration. Curl Syntax curl -k -u https://:/rest/... To keep your password secure, do not provide it on the command line. Curl will prompt you for your password, and you can enter it securely. Wget Syntax wget --no-check-certificate --user --ask-password https://:/rest/... To keep your password secure, do not provide it on the command line. Use the --ask-password option instead; then wget will prompt you for your password and you can enter it securely. Command-Line Interface (CLI) Syntax The MapR CLI commands are documented using the following conventions: [Square brackets] indicate an optional parameter indicate a value to enter The following syntax example shows that the volume mount command requires the -name parameter, for which you must enter a list of volumes, and all other parameters are optional: maprcli volume mount [ -cluster ] -name [ -path ] For clarity, the syntax examples show each parameter on a separate line; in practical usage, the command and all parameters and options are typed on a single line. Example: maprcli volume mount -name test-volume -path /test Common Parameters The following parameters are available for many commands in both the REST and command-line contexts. Parameter Description cluster The cluster on which to run the command. If this parameter is omitted, the command is run on the same cluster where it is issued. In multi-cluster contexts, you can use this parameter to specify a different cluster on which to run the command. zkconnect A ZooKeeper connect string, which specifies a list of the hosts running ZooKeeper, and the port to use on each, in the format: '< host>[:][,[:]...]' Default: 'localhost:5181' In most cases the ZooKeeper connect string can be omitted, but it is useful in certain cases when the CLDB is not running. Common Options The following options are available for most commands in the command-line context. Option Description -noheader When displaying tabular output from a command, omits the header row. -long Shows the entire value. This is useful when the command response contains complex information. When -long is omitted, complex information is displayed as an ellipsis (...). -json Displays command output in JSON format. When -json is omitted, the command output is displayed in tabular format. -cli.loglevel Specifies a log level for API output. Legal values for this option are: DEBUG INFO ERROR WARN TRACE FATAL Filters Some MapR CLI commands use filters, which let you specify large numbers of nodes or volumes by matching specified values in specified fields rather than by typing each name explicitly. Filters use the following format: [""][""] ... field Field on which to filter. The field depends on the command with which the filter is used. operator An operator for that field: == - Exact match != - Does not match > - Greater than < - Less than >= - Greater than or equal to <= - Less than or equal to value Value on which to filter. Wildcards (using *) are allowed for operators == and !=. There is a special value all that matches all values. You can use the wildcard (*) for partial matches. For example, you can display all volumes whose owner is root and whose name begins with t est as follows: maprcli volume list -filter [n=="test*"]and[on=="root"] Response The commands return responses in JSON or in a tabular format. When you run commands from the command line, the response is returned in tabular format unless you specify JSON using the -json option; when you run commands through the REST interface, the response is returned in JSON. Success On a successful call, each command returns the error code zero (OK) and any data requested. When JSON output is specified, the data is returned as an array of records along with the status code and the total number of records. In the tabular format, the data is returned as a sequence of rows, each of which contains the fields in the record separated by tabs. JSON { "status":"OK", "total":, "data":[ { } ... ] } Tabular status 0 Or ... ... ... Error When an error occurs, the command returns the error code and descriptive message. JSON { "status":"ERROR", "errors":[ { "id":, "desc":": " } ] } Tabular ERROR () - : acl The acl commands let you work with access control lists (ACLs): acl edit - modifies a specific user's access to a cluster or volume acl set - modifies the ACL for a cluster or volume acl show - displays the ACL associated with a cluster or volume In order to use the acl edit command, you must have full control (fc) permission on the cluster or volume for which you are running the command. Specifying Permissions Specify permissions for a user or group with a string that lists the permissions for that user or group. To specify permissions for multiple users or groups, use a string for each, separated by spaces. The format is as follows: Users - :[,...][ :[,:[,...][ :[, ] [ -group ] [ -name ] -type cluster|volume [ -user ] REST http[s]:///rest/acl/edit? Parameters Parameter Description cluster The cluster on which to run the command. group Groups and allowed actions for each group. See acl. Format: :[,...][ :[,:[,...][ :[, ] [ -group ] [ -name ] -type cluster|volume [ -user ] REST http[s]:///rest/acl/edit? Parameters Parameter Description cluster The cluster on which to run the command. group Groups and allowed actions for each group. See acl. Format: :[,...][ :[,:[,...][ :[, ] [ -group ] [ -name ] [ -output long|short|terse ] [ -perm ] -type cluster|volume [ -user ] REST http[s]:///rest/acl/show? Parameters Parameter Description cluster The name of the cluster on which to run the command group The group for which to display permissions name The cluster or volume name output The output format: long short terse perm When this option is specified, acl show displays the permissions available for the object type specified in the type parameter. type Cluster or volume. user The user for which to display permissions Output The actions that each user or group is allowed to perform on the cluster or the specified volume. For information about each allowed action, see a cl. Principal User root Group root All users Examples Allowed actions [r, ss, cv, a, fc] [r, ss, cv, a, fc] [r] Show the ACL for "test-volume": CLI maprcli acl show -type volume -name test-volume Show the permissions that can be set on a cluster: CLI maprcli acl show -type cluster -perm alarm The alarm commands perform functions related to system alarms: alarm clear - clears one or more alarms alarm clearall - clears all alarms alarm config load - displays the email addresses to which alarm notifications are to be sent alarm config save - saves changes to the email addresses to which alarm notifications are to be sent alarm list - displays alarms on the cluster alarm names - displays all alarm names alarm raise - raises a specified alarm Alarm Notification Fields The following fields specify the configuration of alarm notifications. Field Description alarm The named alarm. individual Specifies whether individual alarm notifications are sent to the default email address for the alarm type. 0 - do not send notifications to the default email address for the alarm type 1 - send notifications to the default email address for the alarm type email A custom email address for notifications about this alarm type. If specified, alarm notifications are sent to this email address, regardless of whether they are sent to the default email address Alarm Types See Alarms Reference. Alarm History To see a history of alarms that have been raised, look at the file /opt/mapr/logs/cldb.log on the master CLDB node. Example: grep ALARM /opt/mapr/logs/cldb.log alarm clear Clears one or more alarms. Permissions required: fc or a Syntax CLI maprcli alarm clear -alarm [ -cluster ] [ -entity ] REST http[s]://:/rest/alarm/clear? Parameters Parameter Description alarm The named alarm to clear. See Alarm Types. cluster The cluster on which to run the command. entity The entity on which to clear the alarm. Examples Clear a specific alarm: CLI REST maprcli alarm clear -alarm NODE_ALARM_DEBUG_LOGGING https://r1n1.sj.us:8443/rest/alarm/clear?alarm=NODE_ALARM_DEBUG_LOGGING alarm clearall Clears all alarms. Permissions required: fc or a Syntax CLI REST maprcli alarm clearall [ -cluster ] http[s]://:/rest/alarm/clearall? Parameters Parameter Description cluster The cluster on which to run the command. Examples Clear all alarms: CLI REST maprcli alarm clearall https://r1n1.sj.us:8443/rest/alarm/clearall alarm config load Displays the configuration of alarm notifications. Permissions required: login, fc or a Syntax CLI maprcli alarm config load [ -cluster ] [ -output terse|verbose ] REST http[s]://:/rest/alarm/config/load Parameters Parameter Description cluster The cluster on which to run the command. output Whether the output should be terse or verbose. Output A list of configuration values for alarm notifications. Output Fields See Alarm Notification Fields. Sample output alarm CLUSTER_ALARM_BLACKLIST_TTS CLUSTER_ALARM_UPGRADE_IN_PROGRESS CLUSTER_ALARM_UNASSIGNED_VIRTUAL_IPS VOLUME_ALARM_SNAPSHOT_FAILURE VOLUME_ALARM_MIRROR_FAILURE VOLUME_ALARM_DATA_UNDER_REPLICATED VOLUME_ALARM_DATA_UNAVAILABLE VOLUME_ALARM_ADVISORY_QUOTA_EXCEEDED VOLUME_ALARM_QUOTA_EXCEEDED NODE_ALARM_CORE_PRESENT NODE_ALARM_DEBUG_LOGGING NODE_ALARM_DISK_FAILURE NODE_ALARM_OPT_MAPR_FULL NODE_ALARM_VERSION_MISMATCH NODE_ALARM_TIME_SKEW NODE_ALARM_SERVICE_CLDB_DOWN NODE_ALARM_SERVICE_FILESERVER_DOWN NODE_ALARM_SERVICE_JT_DOWN NODE_ALARM_SERVICE_TT_DOWN NODE_ALARM_SERVICE_HBMASTER_DOWN NODE_ALARM_SERVICE_HBREGION_DOWN NODE_ALARM_SERVICE_NFS_DOWN NODE_ALARM_SERVICE_WEBSERVER_DOWN NODE_ALARM_SERVICE_HOSTSTATS_DOWN NODE_ALARM_ROOT_PARTITION_FULL AE_ALARM_AEADVISORY_QUOTA_EXCEEDED AE_ALARM_AEQUOTA_EXCEEDED individual 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 1 email Examples Display the alarm notification configuration: CLI maprcli alarm config load REST https://r1n1.sj.us:8443/rest/alarm/config/load alarm config save Sets notification preferences for alarms. Permissions required: fc or a Alarm notifications can be sent to the default email address and a specific email address for each named alarm. If individual is set to 1 for a specific alarm, then notifications for that alarm are sent to the default email address for the alarm type. If a custom email address is provided, notifications are sent there regardless of whether they are also sent to the default email address. Syntax CLI maprcli alarm config save [ -cluster ] -values REST http[s]://:/rest/alarm/config/save? Parameters Parameter Description cluster The cluster on which to run the command. values A comma-separated list of configuration values for one or more alarms, in the following format: ,, See Alarm Notification Fields. Examples Send alert emails for the AE_ALARM_AEQUOTA_EXCEEDED alarm to the default email address and a custom email address: CLI maprcli alarm config save -values "AE_ALARM_AEQUOTA_EXCEEDED,1,[email protected]" REST https://r1n1.sj.us:8443/rest/alarm/config/save?values=AE_ALARM_AEQUOTA_EXCEEDED,1,[email protected] alarm list Lists alarms in the system. Permissions required: login, fc or a You can list all alarms, alarms by type (Cluster, Node or Volume), or alarms on a particular node or volume. To retrieve a count of all alarm types, pass 1 in the summary parameter. You can specify the alarms to return by filtering on type and entity. Use start and limit to retrieve only a specified window of data. Syntax CLI maprcli alarm list [ -alarm ] [ -cluster ] [ -entity ] [ -limit ] [ -output (terse|verbose) ] [ -start ] [ -summary (0|1) ] [ -type ] REST http[s]://:/rest/alarm/list? Parameters Parameter Description alarm The alarm type to return. See Alarm Types. cluster The cluster on which to list alarms. entity The name of the cluster, node, volume, user, or group to check for alarms. limit The number of records to retrieve. Default: 2147483647 output Whether the output should be terse or verbose. start The list offset at which to start. summary Specifies the type of data to return: 1 = count by alarm type 0 = List of alarms type The entity type: cluster node volume ae Output Information about one or more named alarms on the cluster, or for a specified node, volume, user, or group. Output Fields Field Description alarm state State of the alarm: 0 = Clear 1 = Raised description A description of the condition that raised the alarm entity The name of the volume, node, user, or group. alarm name The name of the alarm. alarm statechange time The date and time the alarm was most recently raised. Sample Output alarm state description entity alarm name statechange time 1 Volume desired replication is 1, current replication is 0 mapr.qa-node173.qa.prv.local.logs VOLUME_ALARM_DATA_UNDER_REPLICATED 1 Volume data unavailable mapr.qa-node173.qa.prv.local.logs VOLUME_ALARM_DATA_UNAVAILABLE 1 Volume desired replication is 1, current replication is 0 mapr.qa-node235.qa.prv.local.mapred VOLUME_ALARM_DATA_UNDER_REPLICATED 1 Volume data unavailable mapr.qa-node235.qa.prv.local.mapred VOLUME_ALARM_DATA_UNAVAILABLE 1 Volume desired replication is 1, current replication is 0 mapr.qa-node175.qa.prv.local.logs VOLUME_ALARM_DATA_UNDER_REPLICATED Examples alarm 1296707707872 1296707707871 1296708283355 1296708283099 1296706343256 List a summary of all alarms CLI maprcli alarm list -summary 1 REST https://r1n1.sj.us:8443/rest/alarm/list?summary=1 List cluster alarms CLI maprcli alarm list -type 0 REST https://r1n1.sj.us:8443/rest/alarm/list?type=cluster alarm names Displays a list of alarm names. No permission is necessary (any user can run this command). Syntax CLI maprcli alarm names REST http[s]://:/rest/alarm/names Examples Display all alarm names: CLI maprcli alarm names REST https://r1n1.sj.us:8443/rest/alarm/names alarm raise Raises a specified alarm or alarms. Permissions required fc or a. Syntax CLI REST Parameters maprcli alarm raise -alarm [ -cluster ] [ -description ] [ -entity ] http[s]://:/rest/alarm/raise? Parameter Description alarm The alarm type to raise. See Alarm Types. cluster The cluster on which to run the command. description A brief description. entity The entity on which to raise alarms. Examples Raise a specific alarm: CLI maprcli alarm raise -alarm NODE_ALARM_DEBUG_LOGGING REST https://r1n1.sj.us:8443/rest/alarm/raise?alarm=NODE_ALARM_DEBUG_LOGGING cluster The cluster commands let you view and manage cluster-wide settings. cluster mapreduce get displays the cluster-wide MapReduce mode. cluster mapreduce set sets the cluster-wide MapReduce mode. cluster mapreduce get Displays the cluster-wide default for the MapReduce mode. Syntax CLI maprcli cluster mapreduce get REST http[s]://:/rest/cluster/mapreduce/get Output Fields Field Description default_mode Displays either yarn or classic. mapreduce_version Displays the hadoop version associated with the default_mode. Examples CLI maprcli cluster mapreduce get REST https://r1n1.sj.us:8443/rest/cluster/mapreduce/get cluster mapreduce set Sets the cluster-wide MapReduce mode. Syntax CLI maprcli cluster mapreduce set [-mode ] REST http[s]://:/rest/cluster/mapreduce/set? Parameters Parameter Description Mode The MapReduce mode of the cluster. Enter one of the following: classic: Use JobTracker and TaskTracker to run MapReduce Jobs. yarn: Use Resource Manager and Node Manager to run MapReduce jobs or applications. Examples Sets the MapReduce mode for the cluster to classic. CLI maprcli cluster mapreduce set -mode classic REST https://r1n1.sj.us:8443/rest/cluster/mapreduce/set?mode=classic config The config commands let you work with configuration values for the MapR cluster: config load displays the values config save makes changes to the stored values Configuration Fields Field Default Value cldb.balancer.disk.max.switches.in.nodes.percentage 10 cldb.balancer.disk.paused 1 cldb.balancer.disk.sleep.interval.sec 2 * 60 cldb.balancer.disk.threshold.percentage 70 cldb.balancer.logging 0 cldb.balancer.role.max.switches.in.nodes.percentage 10 cldb.balancer.role.paused 1 cldb.balancer.role.sleep.interval.sec 15 * 60 cldb.balancer.startup.interval.sec 30 * 60 cldb.cluster.almost.full.percentage 90 cldb.container.alloc.selector.algo 0 cldb.container.assign.buffer.sizemb 1 * 1024 Description The percentage at which the CLUSTER_ALARM_CLUSTER_ALMOST_FULL alarm is triggered. cldb.container.create.diskfull.threshold 80 cldb.container.sizemb 16 * 1024 cldb.default.chunk.sizemb 256 cldb.default.volume.topology The default topology for new volumes. cldb.dialhome.metrics.rotation.period 365 cldb.fileserver.activityreport.interval.hb.multiplier 3 cldb.fileserver.containerreport.interval.hb.multiplier 1800 cldb.fileserver.heartbeat.interval.sec 1 cldb.force.master.for.container.minutes 1 cldb.fs.mark.inactive.sec 5 * 60 cldb.fs.mark.rereplicate.sec 60 * 60 cldb.fs.workallocator.num.volume.workunits 20 cldb.fs.workallocator.num.workunits 80 cldb.ganglia.cldb.metrics 0 cldb.ganglia.fileserver.metrics 0 cldb.heartbeat.monitor.sleep.interval.sec 60 cldb.log.fileserver.timeskew.interval.mins 60 cldb.max.parallel.resyncs.star 2 cldb.mfs.heartbeat.timeout.multiple 10 cldb.min.containerid 1 cldb.min.fileservers 1 cldb.min.snap.containerid 1 cldb.min.snapid 1 cldb.replication.manager.start.mins 15 cldb.replication.process.num.containers 60 cldb.replication.sleep.interval.sec 15 cldb.replication.tablescan.interval.sec 2 * 60 cldb.restart.wait.time.sec 180 cldb.snapshots.inprogress.cleanup.minutes 30 cldb.topology.almost.full.percentage 90 cldb.volume.default.replication The number of seconds a node can fail to heartbeat before it is considered dead. Once a node is considered dead, the CLDB re-replicates any data contained on the node. Specifies a heartbeat timeout multiple. For small clusters, the heartbeat interval is 1 second and the multiple is 10 by default, which makes the heartbeat timeout 10 seconds. The minimum CLDB fileservers. The delay between CLDB startup and replication manager startup, to allow all nodes to register and heartbeat The default replication for the CLDB volumes. cldb.volume.epoch cldb.volumes.default.min.replication 2 cldb.volumes.default.replication 3 mapr.domainname The domain name MapR uses to get operating system users and groups (in domain mode). mapr.entityquerysource Sets MapR to get user information from LDAP (LDAP mode) or from the operating system of a domain (domain mode): ldap domain mapr.eula.user mapr.eula.time mapr.fs.nocompression "bz2,gz,tgz,tbz2, zip,z,Z,mp3,jpg, jpeg,mpg,mpeg,avi, gif,png" The file types that should not be compressed. See Extensions Not Compressed. mapr.fs.permissions.supergroup The super group of the MapR-FS layer. mapr.fs.permissions.superuser The super user of the MapR-FS layer. mapr.ldap.attribute.group The LDAP server group attribute. mapr.ldap.attribute.groupmembers The LDAP server groupmembers attribute. mapr.ldap.attribute.mail The LDAP server mail attribute. mapr.ldap.attribute.uid The LDAP server uid attribute. mapr.ldap.basedn The LDAP server Base DN. mapr.ldap.binddn The LDAP server Bind DN. mapr.ldap.port The port MapR is to use on the LDAP server. mapr.ldap.server The LDAP server MapR uses to get users and groups (in LDAP mode). mapr.ldap.sslrequired Specifies whether the LDAP server requires SSL: 0 == no 1 == yes mapr.license.exipry.notificationdays 30 mapr.quota.group.advisorydefault The default group advisory quota; see Managing Quotas. mapr.quota.group.default The default group quota; see Managing Quotas. mapr.quota.user.advisorydefault The default user advisory quota; see Managing Quotas. mapr.quota.user.default The default user quota; see Managing Quotas. mapr.smtp.port The port MapR uses on the SMTP server (1-65535). mapr.smtp.sender.email The reply-to email address MapR uses when sending notifications. mapr.smtp.sender.fullname The full name MapR uses in the Sender field when sending notifications. mapr.smtp.sender.password The password MapR uses to log in to the SMTP server when sending notifications. mapr.smtp.sender.username The username MapR uses to log in to the SMTP server when sending notifications. mapr.smtp.server The SMTP server that MapR uses to send notifications. mapr.smtp.sslrequired Specifies whether SSL is required when sending email: 0 == no 1 == yes mapr.targetversion mapr.webui.http.port The port MapR uses for the MapR Control System over HTTP (0-65535); if 0 is specified, disables HTTP access. mapr.webui.https.certpath The HTTPS certificate path. mapr.webui.https.keypath The HTTPS key path. mapr.webui.https.port The port MapR uses for the MapR Control System over HTTPS (0-65535); if 0 is specified, disables HTTPS access. mapr.webui.timeout The number of seconds the MapR Control System allows to elapse before timing out. mapreduce.cluster.permissions.supergroup The super group of the MapReduce layer. mapreduce.cluster.permissions.superuser The super user of the MapReduce layer. mfs.feature.devicefile.support 1 Defines whether named pipes can be used over NFS. config load Displays information about the cluster configuration. You can use the keys parameter to specify which information to display. Syntax CLI REST maprcli config load [ -cluster ] -keys http[s]://:/rest/config/load? Parameters Parameter Description cluster The cluster for which to display values. keys The fields for which to display values; see the Configuration Fields table Output Information about the cluster configuration. See the Configuration Fields table. Sample Output { "status":"OK", "total":1, "data":[ { "mapr.webui.http.port":"8080", "mapr.fs.permissions.superuser":"root", "mapr.smtp.port":"25", "mapr.fs.permissions.supergroup":"supergroup" } ] } Examples Display several keys: CLI REST maprcli config load -keys mapr.webui.http.port,mapr.webui.https.port,mapr.webui.https.keystorepath,mapr.webui.h https://r1n1.sj.us:8443/rest/config/load?keys=mapr.webui.http.port,mapr.webui.https.port,mapr.webui.https.keyst config save Saves configuration information, specified as key/value pairs. Permissions required: fc or a. See the Configuration Fields table. Syntax CLI REST maprcli config save [ -cluster ] -values http[s]://:/rest/config/save? Parameters Parameter Description cluster The cluster on which to run the command. values A JSON object containing configuration fields; see the Configuration Fields table. Examples Configure MapR SMTP settings: CLI REST maprcli config save -values '{"mapr.smtp.provider":"gmail","mapr.smtp.server":"smtp.gmail.com","mapr.smtp.sslre Cd","mapr.smtp.sender.email":"[email protected]","mapr.smtp.sender.username":"[email protected]","mapr.smtp.sender.pass https://r1n1.sj.us:8443/rest/config/save?values={"mapr.smtp.provider":"gmail","mapr.smtp.server":"smtp.gmail.co Cd","mapr.smtp.sender.email":"[email protected]","mapr.smtp.sender.username":"[email protected]","mapr.smtp.sender.pass dashboard The dashboard info command displays a summary of information about the cluster. dashboard info Displays a summary of information about the cluster. For best results, use the -json option when running dashboard info from the command line. Syntax CLI REST maprcli dashboard info [ -cluster ] [ -multi_cluster_info true|false. default: false ] [ -version true|false. default: false ] [ -zkconnect ] http[s]://:/rest/dashboard/info? Parameters Parameter Description cluster The cluster on which to run the command. multi_cluster_info Specifies whether to display cluster information from multiple clusters. version Specifies whether to display the version. zkconnect ZooKeeper Connect String Output A summary of information about the services, volumes, mapreduce jobs, health, and utilization of the cluster. Output Fields Field Description Timestamp The time at which the dashboard info data was retrieved, expressed as a Unix epoch time. Status The success status of the dashboard info command. Total The number of clusters for which data was queried in the dashboard info command. Version The MapR software version running on the cluster. Cluster The following information about the cluster: name — the cluster name ip — the IP address of the active CLDB id — the cluster ID services The number of active, stopped, failed, and total installed services on the cluster: CLDB File server Job tracker Task tracker HB master HB region server volumes The number and size (in GB) of volumes that are: Mounted Unmounted mapreduce The following mapreduce information: Queue time Running jobs Queued jobs Running tasks Blacklisted jobs maintenance The following information about system health: Failed disk nodes Cluster alarms Node alarms Versions utilization The following utilization information: CPU: Memory Disk space compression Sample Output # maprcli dashboard info -json { "timestamp":1336760972531, "status":"OK", "total":1, "data":[ { "version":"2.0.0", "cluster":{ "name":"mega-cluster", "ip":"192.168.50.50", "id":"7140172612740778586" }, "volumes":{ "mounted":{ "total":76, "size":88885376 }, "unmounted":{ "total":1, "size":6 } }, "utilization":{ "cpu":{ "util":14, "total":528, "active":75 }, "memory":{ "total":2128177, "active":896194 }, "disk_space":{ "total":707537, "active":226848 }, "compression":{ "compressed":86802, "uncompressed":116655 } }, "services":{ "fileserver":{ "active":22, "stopped":0, "failed":0, "total":22 }, "nfs":{ "active":1, "stopped":0, "failed":0, "total":1 }, "webserver":{ "active":1, "stopped":0, "failed":0, "total":1 }, "cldb":{ "active":1, "stopped":0, "failed":0, "total":1 }, "tasktracker":{ "active":21, "stopped":0, "failed":0, "total":21 }, "jobtracker":{ "active":1, "standby":0, "stopped":0, "failed":0, "total":1 }, "hoststats":{ "active":22, "stopped":0, "failed":0, "total":22 } }, "mapreduce":{ "running_jobs":1, "queued_jobs":0, "running_tasks":537, "blacklisted":0 } } ] } Examples Display dashboard information: CLI maprcli dashboard info -json REST https://r1n1.sj.us:8443/rest/dashboard/info dialhome The dialhome commands let you change the Dial Home status of your cluster: dialhome dialhome dialhome dialhome dialhome ackdial - acknowledges a successful Dial Home transmission. enable - enables or disables Dial Home. lastdialed - displays the last Dial Home transmission. metrics - displays the metrics collected by Dial Home. status - displays the current Dial Home status. dialhome ackdial Acknowledges the most recent Dial Home on the cluster. Permissions required: fc or a Syntax CLI REST maprcli dialhome ackdial [ -forDay ] http[s]://:/rest/dialhome/ackdial[?parameters] Parameters Parameter Description forDay Date for which the recorded metrics were successfully dialed home. Accepted values: UTC timestamp or a UTC date in MM/DD/YY format. Default: yesterday Examples Acknowledge Dial Home: CLI REST maprcli dialhome ackdial https://r1n1.sj.us:8443/rest/dialhome/ackdial dialhome enable Enables Dial Home on the cluster. Permissions required: fc or a Syntax CLI maprcli dialhome enable -enable 0|1 REST http[s]://:/rest/dialhome/enable Parameters Parameter Description enable Specifies whether to enable or disable Dial Home: 0 - Disable 1 - Enable Output A success or failure message. Sample output pconrad@s1-r1-sanjose-ca-us:~$ maprcli dialhome enable -enable 1 Successfully enabled dialhome pconrad@s1-r1-sanjose-ca-us:~$ maprcli dialhome status Dial home status is: enabled Examples Enable Dial Home: CLI REST maprcli dialhome enable -enable 1 https://r1n1.sj.us:8443/rest/dialhome/enable?enable=1 dialhome lastdialed Displays the date of the last successful Dial Home call. Permissions required: fc or a Syntax CLI maprcli dialhome lastdialed REST http[s]://:/rest/dialhome/lastdialed Output The date of the last successful Dial Home call. Sample output $ maprcli dialhome lastdialed date 1322438400000 Examples Show the date of the most recent Dial Home: CLI REST maprcli dialhome lastdialed https://r1n1.sj.us:8443/rest/dialhome/lastdialed dialhome metrics Returns a compressed metrics object. Permissions required: fc or a Syntax CLI REST maprcli dialhome metrics [ -forDay ] http[s]://:/rest/dialhome/metrics Parameters Parameter Description forDay Date for which the recorded metrics were successfully dialed home. Accepted values: UTC timestamp or a UTC date in MM/DD/YY format. Default: yesterday Output Sample output $ maprcli dialhome metrics metrics [B@48067064 Examples Show the Dial Home metrics: CLI maprcli dialhome metrics REST https://r1n1.sj.us:8443/rest/dialhome/metrics dialhome status Displays the Dial Home status. Permissions required: fc or a Syntax CLI maprcli dialhome status REST http[s]://:/rest/dialhome/status Output The current Dial Home status. Sample output $ maprcli dialhome status enabled 1 Examples Display the Dial Home status: CLI maprcli dialhome status REST https://r1n1.sj.us:8443/rest/dialhome/status disk The disk commands lets you work with disks: disk add adds a disk to a node disk list lists disks disk listall lists all disks disk remove removes a disk from a node Disk Fields The following table shows the fields displayed in the output of the disk list and disk listall commands. You can choose which fields (columns) to display and sort in ascending or descending order by any single field. Verbose Name Terse Name Description hostname hn Hostname of node which owns this disk/partition. diskname n Name of the disk or partition. status st Disk status: 0 = Good 1 = Bad disk powerstatus pst Disk power status: 0 = Active/idle (normal operation) 1 = Standby (low power mode) 2 = Sleeping (lowest power mode, drive is completely shut down) mount mt Disk mount status 0 = unmounted 1 = mounted fstype fs File system type modelnum mn Model number serialnum sn Serial number firmwareversion fw Firmware version vendor ven Vendor name totalspace dst Total disk space, in MB usedspace dsu Disk space used, in MB availablespace dsa Disk space available, in MB err Disk error message, in English. Note that this will not be translated. Only sent if st == 1. ft Disk failure time, MapR disks only. Only sent if st == 1. disk add Adds one or more disks to the specified node. Permissions required: fc or a If you are running MapR 1.2.2 or earlier, do not use the disk add command or the MapR Control System to add disks to MapR-FS. You must either upgrade to MapR 1.2.3 before adding or replacing a disk, or use the following procedure (which avoids the disk add comma nd): 1. Use the MapR Control System to remove the failed disk. All other disks in the same storage pool are removed at the same time. Make a note of which disks have been removed. 2. 2. Create a text file /tmp/disks.txt containing a list of the disks you just removed. See Setting Up Disks for MapR. 3. Add the disks to MapR-FS by typing the following command (as root or with sudo): /opt/mapr/server/disksetup -F /tmp/disks.txt Syntax CLI maprcli disk add [ -cluster ] -disks -host REST http[s]://:/rest/disk/add? Parameters Parameter Description cluster The cluster on which to add disks. If not specified, the default is the current cluster. disks A comma-separated list of disk names. Examples: /dev/sdc /dev/sdd,/dev/sde,/dev/sdf The hostname or IP address of the machine on which to add the disk. host Output Output Fields Field Description ip The IP address of the machine that owns the disk(s). disk The name of a disk or partition. Example sca or sca/sca1 all The string all, meaning all unmounted disks for this node. Examples Add a disk: CLI REST maprcli disk add -disks /dev/sda1 -host 10.250.1.79 https://r1n1.sj.us:8443/rest/disk/add?disks=["/dev/sda1"] disk list The maprcli disk list command lists the disks on a node. Syntax CLI maprcli disk list -host [ -output terse|verbose ] [ -system 1|0 ] REST http[s]://:/rest/disk/list? Parameters Parameter Description host The node on which to list the disks. output Whether the output should be terse or verbose. Default is verbose. system Show only operating system disks: 0 - shows only MapR-FS disks 1 - shows only operating system disks Not specified - shows both MapR-FS and operating system disks Output Information about the specified disks. See the Disk Fields table. Examples List disks on a host: CLI REST maprcli disk list -host 10.10.100.22 https://r1n1.sj.us:8443/rest/disk/list?host=10.10.100.22 disk listall Lists all disks Syntax CLI maprcli disk listall [ -cluster ] [ -limit ] [ -output terse|verbose ] [ -start ] REST http[s]://:/rest/disk/listall? Parameters Parameter Description cluster The cluster on which to run the command. limit The number of rows to return, beginning at start. Default: 0 output Always the string terse. start The offset from the starting row according to sort. Default: 0 Output Information about all disks. See the Disk Fields table. Examples List all disks: CLI maprcli disk listall REST https://r1n1.sj.us:8443/rest/disk/listall disk remove Removes a disk from MapR-FS. Permissions required: fc or a The disk remove command does not remove a disk containing unreplicated data unless forced. To force disk removal, specify -force with the value 1. Only use the -force 1 option if you are sure that you do not need the data on the disk. This option removes the disk without regard to replication factor or other data protection mechanisms, and may result in permanent data loss. Removing a disk in the storage pool that contains Container ID 1 will stop your cluster. Container ID 1 contains CLDB data for the master CLDB. Run disk remove without the -force 1 option first and examine the warning messages to make sure you aren't removing the disk with Container ID 1. To safely remove such a disk, perform a CLDB Failover to make one of the other CLDB nodes the primary CLDB, then remove the disk as normal. Syntax CLI maprcli disk remove [ -cluster ] -disks [ -force 0|1 ] -host REST http[s]://:/rest/disk/remove? Parameters Parameter Description cluster The cluster on which to run the command. disks A list of disks in the form: ["disk"]or["disk","disk","disk"...]or[] force Whether to force 0 (default) - do not remove the disk or disks if there is unreplicated data on the disk 1 - remove the disk or disks regardless of data loss or other consequences The hostname or ip address of the node from which to remove the disk. host Output Output Fields Field Description disk The name of a disk or partition. Example: sca or sca/sca1 all The string all, meaning all unmounted disks attached to the node. disks A comma-separated list of disks which have non-replicated volumes. "sca" or "sca/sca1,scb" Examples Remove a disk: CLI maprcli disk remove -disks ["sda1"] REST https://r1n1.sj.us:8443/rest/disk/remove?disks=["sda1"] dump The maprcli dump commands can be used to view key information about volumes, containers, storage pools, and MapR cluster services for debugging and troubleshooting. dump balancerinfo returns detailed information about the storage pools on a cluster. If there are any active container moves, the command returns information about the source and destination storage pools. dump balancermetrics returns a cumulative count of container moves and MB of data moved between storage pools. dump cldbnodes returns the IP address and port number of the CLDB nodes on the cluster. dump containerinfo returns detailed information about one or more specified containers. dump replicationmanagerinfo returns information about volumes and the containers on those volumes including the nodes on which the containers have been replicated and the space allocated to each container. dump replicationmanagerqueueinfo returns information that enables you to identify containers that are under-replicated or over-replicated. dump rereplicationinfo returns information about the ongoing re-replication of replica containers including the destination IP address and port number, the ID number of the destination file server, and the ID number of the destination storage pool. dump rolebalancerinfo returns information about active replication role switches. dump rolebalancermetrics returns the cumulative number of times that the replication role balancer has switched the replication role of name containers and data containers on the cluster. dump volumeinfo returns information about volumes and the associated containers. dump volumenodes returns the IP address and port number of volume nodes. dump zkinfo returns the ZooKeeper znodes. This command is used by the mapr-support-collect.sh script to gather cluster diagnostics for troubleshooting. dump balancerinfo The maprcli dump balancerinfo command enables you to see how much space is used in storage pools and to track active container moves. The disk space balancer is a tool that balances disk space usage on a cluster by moving containers between storage pools. Whenever a storage pool is over 70% full (or a threshold defined by the cldb.balancer.disk.threshold.percentage parameter), the disk space balancer distributes containers to other storage pools that have lower utilization than the average for that cluster. The disk space balancer aims to ensure that the percentage of space used on all the disks in the node is similar. For more information, see Disk Space Balancer. Syntax maprcli dump balancerinfo [-cluster ] Parameters Parameter Description -cluster The cluster on which to run the command. If this parameter is omitted, the command is run on the same cluster where it is issued. In multi-cluster contexts, you can use this parameter to specify a different cluster on which to run the command. Output The maprcli dump balancerinfo command returns detailed information about the storage pools on a cluster. If there are any active container moves, the command returns information about the source and destination storage pools. # maprcli dump balancerinfo -cluster my.cluster.com -json { "timestamp":1337036566035, "status":"OK", "total":187, "data":[ { "spid":"4bc329ce06752062004fa1a537abcdef", "fsid":5410063549464613987, "ip:port":"10.50.60.72:5660-", "capacityMB":1585096, "usedMB":1118099, "percentage":70, "fullnessLevel":"AboveAverage", "inTransitMB":0, "outTransitMB":31874 }, { "spid":"761fec1fabf32104004fad9630ghijkl", "fsid":3770844641152008527, "ip:port":"10.50.60.73:5660-", "capacityMB":1830364, "usedMB":793679, "percentage":47, "fullnessLevel":"BelowAverage", "inTransitMB":79096, "outTransitMB":0 }, .... { "containerid":4034, "sizeMB":16046, "From fsid":5410063549464613987, "From IP:Port":"10.50.60.72:5660-", "From SP":"4bc329ce06752062004fa1a537abcefg", "To fsid":3770844641152008527, "To IP:Port":"10.50.60.73:5660-", "To SP":"761fec1fabf32104004fad9630ghijkl" }, Output fields Field Description spid The unique ID number of the storage pool. fsid The unique ID number of the file server. The FSID identifies an MapR-FS instance or a node that has MapR-FS running in the cluster. Typically, each node has a group of storage pools, so the same FSID will correspond to multiple SPIDs. ip:port The host IP address and MapR-FS port. capacityMB The total capacity of the storage pool (in MB). usedMB The amount of space used on the storage pool (in MB). percentage The percentage of the storage pool currently utilized. A ratio of the space used (usedMB) to the total capacity (capacityMB) of the storage pool. fullnessLevel The fullness of the storage pool relative to the fullness of the rest of the cluster. Possible values are OverUsed, AboveAverage, Average, BelowAverage, and UnderUsed. For more information, see Monitoring storage pool space usage below. inTransitMB The amount of data (in MB) that the disk space balancer is currently moving into a storage pool. outTransitMB The amount of data (in MB) that the disk space balancer is currently moving out of a storage pool. The following fields are returned only if the disk space balancer is actively moving one or more containers at the time the command is run. Field Description containerid The unique ID number of the container. sizeMB The amount of data (in MB) being moved. From fsid The FSID (file server ID number) of the source file server. From IP:Port The IP address and port number of the source node. From SP The SPID (storage pool ID) of the source storage pool. To fsid The FSID (file server ID number) of the destination file server. To IP:Port The IP address and port number of the destination node. To SP The SPID (storage pool ID number) of the destination storage pool. Examples Monitoring storage pool space usage You can use the maprcli dump balancerinfo command to monitor space usage on storage pools. # maprcli dump balancerinfo -json .... { "spid":"4bc329ce06752062004fa1a537abcefg", "fsid":5410063549464613987, "ip:port":"10.50.60.72:5660-", "capacityMB":1585096, "usedMB":1118099, "percentage":70, "fullnessLevel":"AboveAverage", "inTransitMB":0, "outTransitMB":31874 }, Tracking active container moves Using the maprcli dump balancerinfo command you can monitor the activity of the disk space balancer. Whenever there are active container moves, the command returns information about the source and destination storage pools. # maprcli dump balancerinfo -json .... { "containerid":7840, "sizeMB":15634, "From fsid":8081858704500413174, "From IP:Port":"10.50.60.64:5660-", "From SP":"9e649bf0ac6fb9f7004fa19d20rstuvw", "To fsid":3770844641152008527, "To IP:Port":"10.50.60.73:5660-", "To SP":"fefcc342475f0286004fad963flmnopq" } The example shows that a container (7840) is being moved from a storage pool on node 10.50.60.64 to a storage pool on node 10.50.60.73. Tip You can use the storage pool IDs (SPIDs) to search the CLDB and MFS logs for activity (balancer moves, container moves, creates, deletes, etc.) related to specific storage pools. dump balancermetrics The maprcli dump balancermetrics command returns a cumulative count of container moves and MB of data moved between storage pools. You can run this command periodically to determine how much data has been moved by the disk space balancer between two intervals. The disk space balancer is a tool that balances disk space usage on a cluster by moving containers between storage pools. Whenever a storage pool is over 70% full (or it reaches a threshold defined by the cldb.balancer.disk.threshold.percentage parameter), the disk space balancer distributes containers to other storage pools that have lower utilization than the average for that cluster. The disk space balancer aims to ensure that the percentage of space used on all the disks in the node is similar. For more information, see Disk Space Balancer. Syntax maprcli dump balancermetrics [-cluster ] Parameters Parameter Description -cluster The cluster on which to run the command. If this parameter is omitted, the command is run on the same cluster where it is issued. In multi-cluster contexts, you can use this parameter to specify a different cluster on which to run the command. Output The maprcli dump balancermetrics command returns a cumulative count of container moves and MB of data moved between storage pools since the current CLDB became the master CLDB. # maprcli dump balancermetrics -json { "timestamp":1337770325979, "status":"OK", "total":1, "data":[ { "numContainersMoved":10090, "numMBMoved":3147147, "timeOfLastMove": "Wed May 23 03:51:44 PDT 2012" } ] } Output fields Field Description numContainersMoved The number of containers moved between storage pools by the disk space balancer. numMBMoved The total MB of data moved between storage pools on the cluster. timeOfLastMove The date and time of most recent container move. dump cldbnodes The maprcli dump cldbnodes command lists the nodes that contain container location database (CLDB) data. The CLDB is a service running on one or more MapR nodes that maintains the location of cluster containers, services, and other information. The CLDB automatically replicates its data to other nodes in the cluster, preserving at least two (and generally three) copies of the CLDB data. If the CLDB process dies, it is automatically restarted on the node. Syntax maprcli dump cldbnodes [-cluster ] -zkconnect Parameters Parameter Description -cluster The cluster on which to run the command. If this parameter is omitted, the command is run on the same cluster where it is issued. In multi-cluster contexts, you can use this parameter to specify a different cluster on which to run the command. -zkconnect [:][,[:]...]' Output The maprcli dump cldbnodes command returns the IP address and port number of the CLDB nodes on the cluster. $ maprcli dump cldbnodes -zkconnect localhost:5181 -json { { "timestamp":1309882069107, "status":"OK", "total":1, "data":[ { "valid":[ "10.10.30.39:5660-10.50.60.39:5660-", "10.10.30.38:5660-10.50.60.38:5660-", "10.10.30.35:5660-10.50.60.35:5660-" ] } ] } Examples Disaster Recovery In the event that all CLDB nodes fail, you can restore the CLDB from a backup. It is a good idea to set up an automatic backup of the CLDB volume at regular intervals. You can use the maprcli dump cldbnodes command to set up cron jobs to back up CLDB volumes locally or to external media such as a USB drive. For more information, see Disaster Recovery. To back up a CLDB volume from a remote cluster: 1. Set up a cron job to save the container information on the remote cluster using the following command: # maprcli dump cldbnodes -zkconnect > 2. Set up a cron job to copy the container information file to a volume on the local cluster. 3. Create a mirror volume on the local cluster, choosing the volume mapr.cldb.internal from the remote cluster as the source volume. Set the mirror sync schedule so that it will run at the same time as the cron job. To back up a CLDB volume locally: 1. Set up a cron job to save the container information to a file on external media by running the following command: # maprcli dump cldbnodes -zkconnect > 2. Set up a cron job to create a dump file of the local volume mapr.cldb.internal on external media. Example: # maprcli volume dump create -name mapr.cldb.internal -dumpfile For information about restoring from a backup of the CLDB, contact MapR Support. dump containerinfo The maprcli dump containerinfo command enables you to view detailed information about one or more specified containers. A container is a unit of sharded storage in a MapR cluster. Every container in a MapR volume is either a name container or a data container. The name container is the first container in a volume and holds that volume's namespace and file chunk locations. Depending on its replication role, a name container may be either a master container (part of the original copy of the volume) or a replica container (one of the replicas in the replication chain). Every data container is either a master container, an intermediate container, or a tail container. Syntax maprcli dump containerinfo [-clustername ] -ids Parameters Parameter [-clustername ] -ids Description The cluster on which to run the command. If this parameter is omitted, the command is run on the same cluster where it is issued. In multi-cluster contexts, you can use this parameter to specify a different cluster on which to run the command. Specifies one or more container IDs. Container IDs are comma separated. Output The maprcli dump containerinfo command returns information about one or more containers. # maprcli dump containerinfo -ids 2049 -json { "timestamp":1335831624586, "status":"OK", "total":1, "data":[ { "ContainerId":2049, "Epoch":11, "Master":"10.250.1.15:5660-172.16.122.1:5660-192.168.115.1:5660--11-VALID", "ActiveServers":{ "IP:Port":"10.250.1.15:5660-172.16.122.1:5660-192.168.115.1:5660--11-VALID" }, "InactiveServers":{ }, "UnusedServers":{ }, "OwnedSizeMB":"0 MB", "SharedSizeMB":"0 MB", "LogicalSizeMB":"0 MB", "Mtime":"Thu Mar 22 15:44:22 PDT 2012", "NameContainer":"true", "VolumeName":"mapr.cluster.root", "VolumeId":93501816, "VolumeReplication":3, "VolumeMounted":true } ] } Output fields Field Description ContainerID The unique ID number for the container. Epoch A sequence number that indicates the most recent copy of the container. The CLDB uses the epoch to ensure that an out-of-date copy cannot become the master for the container. Master The physical IP address and port number of the master copy. The master copy is part of the original copy of the volume. ActiveServers The physical IP address and port number of each active node on which the container resides. InactiveServers The physical IP address and port number of each inactive node on which the container resides. UnusedServers The physical IP address and port number of servers from which no "heartbeat" has been received for quite some time. OwnedSizeMB The size on disk (in MB) dedicated to the container. SharedSizeMB The size on disk (in MB) shared by the container. LogicalSizeMB The logical size on disk (in MB) of the container. TotalSizeMB The total size on disk (in MB) allocated to the container. Combines the Owned Size and Shared Size. Mtime The time of the last modification to the contents of the container. NameContainer Indicates if the container is the name container for the volume. If true, the container holds the volume's namespace information and file chunk locations. VolumeName The name of the volume. VolumeId The unique ID number of the volume. VolumeReplication The replication factor, the number of copies of a volume excluding the original. VolumeMounted Indicates whether the volume is mounted. If true, the volume is currently mounted. If false, the volume is not mounted. dump replicationmanagerinfo The maprcli dump replicationmanagerinfo enables you to see which containers are under or over replicated in a specified volume. For each container, the command displays the current state of that container. Syntax maprcli dump replicationmanagerinfo [-cluster ] -volumename Parameters Parameter Description -cluster The cluster on which to run the command. If this parameter is omitted, the command is run on the same cluster where it is issued. In multi-cluster contexts, you can use this parameter to specify a different cluster on which to run the command. -volumename Specifies the name of the volume. Output The maprcli dump replicationmanagerinfo returns information about volumes and the containers on those volumes including the nodes on which the containers have been replicated and the space allocated to each container. # maprcli dump replicationmanagerinfo -cluster my.cluster.com -volumename mapr.metrics -json { "timestamp":1335830006872, "status":"OK", "total":2, "data":[ { "VolumeName":"mapr.metrics", "VolumeId":182964332, "VolumeTopology":"/", "VolumeUsedSizeMB":1, "VolumeReplication":3, "VolumeMinReplication":2 }, { "ContainerId":2053, "Epoch":9, "Master":"10.250.1.15:5660-172.16.122.1:5660-192.168.115.1:5660--9-VALID", "ActiveServers":{ "IP:Port":"10.250.1.15:5660-172.16.122.1:5660-192.168.115.1:5660--9-VALID" }, "InactiveServers":{ }, "UnusedServers":{ }, "OwnedSizeMB":"1 MB", "SharedSizeMB":"0 MB", "LogicalSizeMB":"1 MB", "Mtime":"Mon Apr 30 16:40:41 PDT 2012", "NameContainer":"true" } ] } Output fields Field Description VolumeName Indicates the name of the volume. VolumeId Indicates the ID number of the volume. VolumeTopology The volume topology corresponds to the node topology of the rack or nodes where the volume resides. By default, new volumes are created with a topology of / (root directory). For more information, see Volume Topology VolumeUsedSizeMB The size on disk (in MB) of the volume. VolumeReplication The desired replication factor, the number of copies of a volume excluding the original. The default value is 3. VolumeMinReplication The minimum replication factor, the number of copies of a volume (excluding the original) that should be maintained by the MapR cluster for normal operation. When the replication factor falls below this minimum, writes to the volume are disabled. The default value is 2. ContainerId The unique ID number for the container. Epoch A sequence number that indicates the most recent copy of the container. The CLDB uses the epoch to ensure that an out-of-date copy cannot become the master for the container. Master The physical IP address and port number of the master copy. The master copy is part of the original copy of the volume. ActiveServers The physical IP address and port number of each active node on which the container resides. InactiveServers The physical IP address and port number of each inactive node on which the container resides. UnusedServers The physical IP address and port number of each on which the container does not reside. OwnedSizeMB The size on disk (in MB) dedicated to the container. SharedSizeMB The size on disk (in MB) shared by the container. LogicalSizeMB The logical size on disk (in MB) of the container. Mtime Indicates the time of the last modification to the container's contents. NameContainer Indicates if the container is the name container for the volume. If true, the container is the volume's first container and replication occurs simultaneously from the master to the intermediate and tail containers. dump replicationmanagerqueueinfo The maprcli dump replicationmanagerqueueinfo command enables you to determine the status of under-replicated containers and over-replicated containers. Syntax maprcli dump replicationmanagerqueueinfo [-cluster ] -queue Parameters Parameter Description cluster The cluster on which to run the command. If this parameter is omitted, the command is run on the same cluster where it is issued. In multi-cluster contexts, you can use this parameter to specify a different cluster on which to run the command. queue The name of the queue. Valid values are 0, 1, or 2. Queue 0 includes containers that have copies below the minimum replication factor for the volume. Queue 1 includes containers that have copies below the replication for the volume, but above the minimum replication factor. Queue 2 includes containers that are over-replicated. Output The maprcli dump replicationmanagerqueueinfo command returns information about one of three queues: 0, 1, or 2. Depending on the queue value entered, the command displays information about containers that are under-replicated or over-replicated. You can use this information to decide if you need to change the replication factor for that volume. # maprcli dump replicationmanagerqueueinfo -queue 0 Mtime LogicalSizeMB UnusedServers ActiveServers TotalSizeMB NameContainer InactiveServers ContainerId Master Epoch SharedSizeMB OwnedSizeMB Thu May 17 10:32:59 PDT 2012 0 MB ... 0 MB false 2065 10.250.1.103:5660--3-VALID 3 0 MB 0 MB Thu May 17 10:32:59 PDT 2012 0 MB ... 0 MB false 0 MB true 0 MB Thu May false 0 MB Thu May false 0 MB Thu May false 0 MB Thu May false 0 MB Thu May false 0 MB Thu May false 0 MB Thu May false 0 MB Thu May false 0 MB Thu May false 0 MB Thu May false 0 MB Thu May false 0 MB Thu May true 0 MB Thu May true 0 MB Thu May true 0 MB Thu May true 0 MB Fri May true 0 MB Thu May true 0 MB Thu May true 0 MB 2064 17 10:32:59 PDT 2012 17 10:32:59 PDT 2012 17 10:32:59 PDT 2012 17 10:32:59 PDT 2012 17 10:32:59 PDT 2012 17 10:32:59 PDT 2012 17 10:32:59 PDT 2012 17 10:32:59 PDT 2012 17 10:32:59 PDT 2012 17 10:32:59 PDT 2012 17 10:32:59 PDT 2012 17 10:36:30 PDT 2012 17 10:36:36 PDT 2012 17 10:32:59 PDT 2012 17 10:37:06 PDT 2012 18 14:33:44 PDT 2012 17 10:32:59 PDT 2012 17 10:32:59 PDT 2012 10.250.1.103:5660--3-VALID 3 0 MB 1 ... 10.250.1.103:5660--8-VALID 8 0 MB 2066 ... 10.250.1.103:5660--3-VALID 3 1 MB 2069 ... 10.250.1.103:5660--5-VALID 5 1 MB 2068 ... 10.250.1.103:5660--5-VALID 5 0 MB 2071 ... 10.250.1.103:5660--3-VALID 3 0 MB 2070 ... 10.250.1.103:5660--3-VALID 3 0 MB 2073 ... 10.250.1.103:5660--3-VALID 3 0 MB 2072 ... 10.250.1.103:5660--3-VALID 3 0 MB 2075 ... 10.250.1.103:5660--3-VALID 3 0 MB 2074 ... 10.250.1.103:5660--3-VALID 3 0 MB 2077 ... 10.250.1.103:5660--3-VALID 3 0 MB 2076 ... 10.250.1.103:5660--3-VALID 3 0 MB 2049 ... 10.250.1.103:5660--7-VALID 7 0 MB 2050 ... 10.250.1.103:5660--7-VALID 7 0 MB 2051 ... 10.250.1.103:5660--6-VALID 6 0 MB 2053 ... 10.250.1.103:5660--6-VALID 6 0 MB 2054 ... 10.250.1.103:5660--5-VALID 5 0 MB 2055 ... 10.250.1.103:5660--3-VALID 3 0 MB 2056 ... 10.250.1.103:5660--3-VALID 3 0 MB 0 MB 0 MB 0 MB 0 MB 0 MB 0 MB 0 MB 0 MB 0 MB 0 MB 0 MB 0 MB 0 MB 0 MB 0 MB 0 MB 0 MB 0 MB 0 MB 0 MB 0 MB 0 MB 0 MB 0 MB 0 MB 0 MB 0 MB 0 MB 0 MB 0 MB 0 MB 0 MB 0 MB 0 MB 0 MB 0 MB 0 MB 0 MB Thu May false 0 MB Thu May false 0 MB Thu May false 0 MB Thu May false 0 MB Thu May false 0 MB Thu May false 0 MB Thu May 17 10:32:59 PDT 2012 17 10:32:59 PDT 2012 17 10:32:59 PDT 2012 17 10:32:59 PDT 2012 17 10:32:59 PDT 2012 17 10:32:59 PDT 2012 17 10:32:59 PDT 2012 0 MB 2057 ... 10.250.1.103:5660--5-VALID 5 0 MB 2058 ... 10.250.1.103:5660--3-VALID 3 0 MB 2059 ... 10.250.1.103:5660--3-VALID 3 0 MB 2060 ... 10.250.1.103:5660--3-VALID 3 0 MB 2061 ... 10.250.1.103:5660--3-VALID 3 0 MB 2062 ... 10.250.1.103:5660--3-VALID 3 0 MB ... 0 MB 0 MB 0 MB 0 MB 0 MB 0 MB 0 MB 0 MB 0 MB 0 MB 0 MB 0 MB 0 MB false 0 MB 2063 10.250.1.103:5660--3-VALID 3 0 MB Output fields Field Description ContainerID The unique ID number of the container. Epoch A sequence number that indicates the most recent copy of the container. The CLDB uses the epoch to ensure that an out-of-date copy cannot become the master for the container. Master The physical IP address and port number of the master copy. The master copy is part of the original copy of the volume. ActiveServers The physical IP address and port number of each active node on which the container resides. InactiveServers The physical IP address and port number of each inactive node on which the container resides. UnusedServers The physical IP address and port number of servers from which no "heartbeat" has been received for quite some time. OwnedSizeMB The size on disk (in MB) dedicated to the container. SharedSizeMB The size on disk (in MB) shared by the container. LogicalSizeMB The logical size on disk (in MB) of the container. TotalSizeMB The total size on disk (in MB) allocated to the container. Combines the Owned Size and Shared Size. Mtime The time of the last modification to the contents of the container. NameContainer Indicates if the container is the name container for the volume. If true, the container holds the volume's namespace information and file chunk locations. dump rereplicationinfo The maprcli dump rereplicationinfo command enables you to view information about the re-replication of containers. Re-replication occurs whenever the number of available replica containers drops below the number prescribed by that volume's replication factor. Re-replication may occur for a variety of reasons including replica container corruption, node unavailability, hard disk failure, or an increase in replication factor. Syntax maprcli dump rereplicationinfo [-cluster ] Parameters Parameter Description -cluster The cluster on which to run the command. If this parameter is omitted, the command is run on the same cluster where it is issued. In multi-cluster contexts, you can use this parameter to specify a different cluster on which to run the command. Output The maprcli dump rereplicationinfo command returns information about the ongoing re-replication of replica containers including the destination IP address and port number, the ID number of the destination file server, and the ID number of the destination storage pool. # maprcli dump rereplicationinfo -json { "timestamp":1338222709331, "status":"OK", "total":7, "data":[ { "containerid":2158, "replica":{ "sizeMB":15467, "To fsid":9057314602141502940, "To IP:Port":"192.0.2.28:5660-", "To SP":"03b5970f41abbe48004f828abaabcdef" } }, { "containerid":3367, "replica":{ "sizeMB":658, "To fsid":3684488804112157043, "To IP:Port":"192.0.2.33:5660-", "To SP":"3b86b4ce5bfd6bbf004f87e9b6ghijkl" } }, { "containerid":3376, "replica":{ "sizeMB":630, "To fsid":3684488804112157043, "To IP:Port":"192.0.2.33:5660-", "To SP":"3b86b4ce5bfd6bbf004f87e9b6ghijkl" } }, { "containerid":3437, "replica":{ "sizeMB":239, "To fsid":6776586767180745590, "To IP:Port":"192.0.2.32:5660-", "To SP":"6cd440fad0426db7004f828b2amnopqr" } }, { "containerid":8833, "replica":{ "sizeMB":7327, "To fsid":9057314602141502940, "To IP:Port":"192.0.2.28:5660-", "To SP":"33885e3c5be9a04d004f828abcstuvwx" } } ] } Output fields Field Description sizeMB The amount of data (in MB) being moved. To fsid The ID number (FSID) of the destination file server. To IP:Port The IP address and port number of the destination node. To SP The ID number (SPID) of the destination storage pool. dump rolebalancerinfo The maprcli dump rolebalancerinfo command enables you to monitor the replication role balancer and view information about active replication role switches. The replication role balancer is a tool that switches the replication roles of containers to ensure that every node has an equal share of master and replica containers (for name containers) and an equal share of master, intermediate, and tail containers (for data containers). The replication role balancer changes the replication role of the containers in a cluster so that network bandwidth is spread evenly across all nodes during the replication process. A container's replication role determines how it is replicated to the other nodes in the cluster. For name containers (the volume's first container), replication occurs simultaneously from the master to all replica containers. For data containers, replication proceeds from the master to the intermediate container(s) until it reaches the tail containers. For more information, see Replication Role Balancer. Syntax maprcli dump rolebalancerinfo [-cluster ] Parameters Parameter Description -cluster The cluster on which to run the command. If this parameter is omitted, the command is run on the same cluster where it is issued. In multi-cluster contexts, you can use this parameter to specify a different cluster on which to run the command. Output The maprcli dump rolebalancerinfo command returns information about active replication role switches. # maprcli dump rolebalancerinfo -json { "timestamp":1335835436698, "status":"OK", "total":1, "data":[ { "containerid": 36659, "Tail IP:Port":"10.50.60.123:5660-", "Updates blocked Since":"Wed May 23 05:48:15 PDT 2012" } ] } Output fields Field Description containerid The unique ID number of the container. Tail IP:Port The IP address and port number of the tail container node. Updates blocked Since During a replication role switch, updates to that container are blocked. dump rolebalancermetrics The maprcli dump rolebalancermetrics command enables you to view the number of times that the replication role balancer has switched the replication role of the name containers and data containers to ensure that containers are balanced across the nodes in the cluster. The replication role balancer is a tool that switches the replication roles of containers to ensure that every node has an equal share of master and replica containers (for name containers) and an equal share of master, intermediate, and tail containers (for data containers). The replication role balancer changes the replication role of the containers in a cluster so that network bandwidth is spread evenly across all nodes during the replication process. A container's replication role determines how it is replicated to the other nodes in the cluster. For name containers (the volume's first container), replication occurs simultaneously from the master to all replica containers. For data containers, replication proceeds from the master to the intermediate container(s) until it reaches the tail containers. For more information, see Replication Role Balancer. Syntax maprcli dump rolebalancermetrics [-cluster ] Parameters Parameter Description -cluster The cluster on which to run the command. If this parameter is omitted, the command is run on the same cluster where it is issued. In multi-cluster contexts, you can use this parameter to specify a different cluster on which to run the command. Output The maprcli dump rolebalancerinfo command returns the cumulative number of times that the replication role balancer has switched the replication role of name containers and data containers on the cluster. # maprcli dump rolebalancermetrics -json { "timestamp":1337777286527, "status":"OK", "total":1, "data":[ { "numNameContainerSwitches":60, "numDataContainerSwitches":28, "timeOfLastMove":"Wed May 23 05:48:00 PDT 2012" } ] } Output fields Field Description numNameContainerSwitches The number of times that the replication role balancer has switched the replication role of name containers. numDataContainerSwitches The number of times that the replication role balancer has switched the replication role of data containers. timeOfLastMove The date and time of the last replication role change. dump volumeinfo The maprcli dump volumeinfo command enables you to view information about a volume and the containers within that volume. A volume is a logical unit that allows you to apply policies to a set of files, directories, and sub-volumes. Using volumes, you can enforce disk usage limits, set replication levels, establish ownership and accountability, and measure the cost generated by different projects or departments. For more information, see Managing Data with Volumes. Syntax maprcli dump volumeinfo [-cluster ] -volumename Parameters Parameter Description cluster The cluster on which to run the command. If this parameter is omitted, the command is run on the same cluster where it is issued. In multi-cluster contexts, you can use this parameter to specify a different cluster on which to run the command. volumename The name of the volume. Output The maprcli volume info returns information about the volume and the containers associated with that volume. Volume information includes the ID, volume name, and replication factor. For each container on the specified volume, the command returns information about nodes and storage. # maprcli dump volumeinfo -volumename mapr.cluster.root -json { "timestamp":1335830155441, "status":"OK", "total":2, "data":[ { "VolumeName":"mapr.cluster.root", "VolumeId":93501816, "VolumeTopology":"/", "VolumeUsedSizeMB":0, "VolumeReplication":3, "VolumeMinReplication":2 }, { "ContainerId":2049, "Epoch":11, "Master":"10.250.1.15:5660-172.16.122.1:5660-192.168.115.1:5660--11-VALID", "ActiveServers":{ "IP:Port":"10.250.1.15:5660-172.16.122.1:5660-192.168.115.1:5660--11-VALID" }, "InactiveServers":{ }, "UnusedServers":{ }, "OwnedSizeMB":"0 MB", "SharedSizeMB":"0 MB", "LogicalSizeMB":"0 MB", "Mtime":"Thu Mar 22 15:44:22 PDT 2012", "NameContainer":"true" } ] } Output fields Field Description VolumeName The name of the volume. VolumeId The unique ID number of the volume. VolumeTopology The volume topology corresponds to the node topology of the rack or nodes where the volume resides. By default, new volumes are created with a topology of / (root directory). For more information, see Volume Topology. VolumeUsedSizeMB The size on disk (in MB) of the volume. VolumeReplication The desired replication factor, the number of copies of a volume. The default value is 3. The maximum value is 6. VolumeMinReplication The minimum replication factor, the number of copies of a volume (excluding the original) that should be maintained by the MapR cluster for normal operation. When the replication factor falls below this minimum, writes to the volume are disabled. The default value is 2. ContainerId The unique ID number of the container. Epoch A sequence number that indicates the most recent copy of the container. The CLDB uses the epoch to ensure that an out-of-date copy cannot become the master for the container. Master The physical IP address and port number of the master copy. The master copy is part of the original copy of the volume. ActiveServers The physical IP address and port number of each active node on which the container resides. InactiveServers The physical IP address and port number of each inactive node on which the container resides. UnusedServers The physical IP address and port number of servers from which no "heartbeat" has been received for quite some time. OwnedSizeMB The size on disk (in MB) dedicated to the container. SharedSizeMB The size on disk (in MB) shared by the container. LogicalSizeMB The logical size on disk (in MB) of the container. TotalSizeMB The total size on disk (in MB) allocated to the container. Combines the Owned Size and Shared Size. Mtime Indicates the time of the last modification to the contents of the container. NameContainer Indicates if the container is the name container for the volume. If true, the container is the volume's first container and replication occurs simultaneously from the master to the intermediate and tail containers. dump volumenodes The maprcli dump volumenodes command enables you to view information about the nodes on a volume. Syntax maprcli dump volumenodes [-cluster ] -volumename Parameters Parameter Description cluster The cluster on which to run the command. If this parameter is omitted, the command is run on the same cluster where it is issued. In multi-cluster contexts, you can use this parameter to specify a different cluster on which to run the command. volumename The name of the volume. Output The maprcli dump volumenodes command returns the IP address and port number of volume nodes. # maprcli dump volumenodes -volumename mapr.hbase -json { "timestamp":1337280188850, "status":"OK", "total":1, "data":[ { "Servers":{ "IP:Port":"10.250.1.103:5660--7-VALID" } } ] } Output fields Field Description IP:Port The IP address and MapR-FS port. dump zkinfo The maprcli dump zkinfo command enables you to view a snapshot of the data stored in Zookeeper as a result of cluster operations. ZooKeeper prevents service coordination conflicts by enforcing a rigid set of rules and conditions, provides cluster-wide information about running services and their configuration, and provides a mechanism for almost instantaneous service failover. The warden will not start any services unless ZooKeeper is reachable and more than half of the configured ZooKeeper nodes are live. The mapr-support-collect.sh script calls the maprcli dump supportdump command to gather cluster diagnostics for troubleshooting. For more information, see mapr-support-collect.sh. Syntax maprcli dump zkinfo [-cluster ] [-zkconnect ] Parameters Parameter Description -cluster The cluster on which to run the command. If this parameter is omitted, the command is run on the same cluster where it is issued. In multi-cluster contexts, you can use this parameter to specify a different cluster on which to run the command. -zkconnect A ZooKeeper connect string, which specifies a list of the hosts running ZooKeeper, and the port to use on each, in the format: '[:][,[:]...]' Output The maprcli dump zkinfo command is run as part of support dump tools to view the current state of the Zookeeper service. The command should always be run using the -json flag. Output in the tabular format is not useful. Command output displays the data stored in the ZooKeepr hierarchical tree of znodes. # maprcli dump zkinfo -json { "timestamp":1335825202157, "status":"OK", "total":1, "data":[ { "/_Stats":"\ncZxid = 0,ctime = Wed Dec 31 16:00:00 PST 1969,mZxid = 0,mtime = Wed Dec 31 16:00:00 PST 1969,pZxid = 516,cversion = 12,dataVersion = 0,aclVersion = 0,ephemeralOwner = 0,dataLength = 0,numChildren = 13", "/":[ { .... } ] } Output fields You can use the maprcli dump zkinfo command as you would use a database snapshot. The /services, /services_config, /server s, and /*_locks znodes are used by the wardens to store and exchange information. Field Description services The /services directory is used by the wardens to store and exchange information about services. datacenter The /datacenter directory contains CLDB "vital signs" that you can to identify the CLDB master, the most recent epoch, and other key data. For more information, see Moving CLDB Data below. services_config The /services_config directory is used by the wardens to store and exchange information. zookeeper The /zookeeper directory stores information about the ZooKeeper service. servers The /servers directory is used by the wardens to store and exchange information. nodes The /nodes directory (znode) stores key information about the nodes. Examples Moving CLDB Data In an M3-licensed cluster, CLDB data must be recovered from a failed CLDB node and installed on another node. The cluster can continue normally as soon as the CLDB is started on another node. For more information, see Recovering from a Failed CLDB Node on an M3 Cluster. Use the maprcli dump zkinfo command to identify the latest epoch of the CLDB, identify the nodes where replicates of the CLDB are stored, and select one of those nodes to serve the new CLDB node. Perform the following steps on any cluster node: 1. Log in as root or use sudo for the following commands. 2. Issue the maprcli dump zkinfo command using the -json flag. # maprcli dump zkinfo -json The output displays the ZooKeeper znodes. 3. In the /datacenter/controlnodes/cldb/epoch/1 directory, locate the CLDB with the latest epoch. 3. { "/datacenter/controlnodes/cldb/epoch/1/KvStoreContainerInfo":" Container ID:1 VolumeId:1 Master:10.250.1.15:5660-172.16.122.1:5660-192.168.115.1:5660--13-VALID Servers: 10.250.1.15:5660-172.16.122.1:5660-192.168.115.1:5660--13-VALID Inactive Servers: Unused Servers: Latest epoch:13" } The Latest Epoch field identifies the current epoch of the CLDB data. In this example, the latest epoch is 13. 4. Select a CLDB from among the copies at the latest epoch. For example, 10.250.2.41:5660--13-VALID indicates that the node has a copy at epoch 13 (the latest epoch). entity The entity commands let you work with entities (users and groups): entity info shows information about a specified user or group entity list lists users and groups in the cluster entity modify edits information about a specified user or group entity info Displays information about an entity. Syntax CLI REST maprcli entity info [ -cluster ] -name [ -output terse|verbose ] -type http[s]://:/rest/entity/info? Parameters Parameter Description cluster The cluster on which to run the command. name The entity name. output Whether to display terse or verbose output. type The entity type Output DiskUsage EntityId 864415 EntityQuota EntityType EntityName VolumeCount EntityAdvisoryquota 0 0 root 208 0 Output Fields Field Description DiskUsage Disk space used by the user or group EntityQuota The user or group quota EntityType The entity type EntityName The entity name VolumeCount The number of volumes associated with the user or group EntityAdvisoryquota The user or group advisory quota EntityId The ID of the user or group Examples Display information for the user 'root': CLI REST maprcli entity info -type 0 -name root https://r1n1.sj.us:8443/rest/entity/info?type=0&name=root entity list Syntax CLI REST Parameters maprcli entity list [ -alarmedentities true|false ] [ -cluster ] [ -columns ] [ -filter ] [ -limit ] [ -output terse|verbose ] [ -start ] http[s]://:/rest/entity/list? 0 Parameter Description alarmedentities Specifies whether to list only entities that have exceeded a quota or advisory quota. cluster The cluster on which to run the command. columns A comma-separated list of fields to return in the query. See the Fields table below. filter A filter specifying entities to display. See Filters for more information. limit The number of rows to return, beginning at start. Default: 0 output Specifies whether output should be terse or verbose. start The offset from the starting row according to sort. Default: 0 Output Information about the users and groups. Fields Field Description EntityType Entity type 0 = User 1 = Group EntityName User or Group name EntityId User or Group id EntityQuota Quota, in MB. 0 = no quota. EntityAdvisoryquota Advisory quota, in MB. 0 = no advisory quota. VolumeCount The number of volumes this entity owns. DiskUsage Disk space used for all entity's volumes, in MB. Sample Output DiskUsage EntityId 5859220 EntityQuota EntityType EntityName VolumeCount EntityAdvisoryquota 0 0 root 209 0 Examples List all entities: CLI maprcli entity list REST https://r1n1.sj.us:8443/rest/entity/list entity modify 0 Modifies a user or group quota or email address. Permissions required: fc or a Syntax CLI maprcli entity modify [ -advisoryquota [ -cluster ] [ -email ] [ -entities ] -name [ -quota ] -type REST http[s]://:/rest/entity/modify? Parameters Parameter Description advisoryquota The advisory quota. cluster The cluster on which to run the command. email Email address. entities A comma-separated list of entities, in the format :. Example: 0:,0:,1:,1:... name The entity name. quota The quota for the entity. type The entity type: 0=user 1-group Examples Modify the email address for the user 'root': CLI REST maprcli entity modify -name root -type 0 -email [email protected] https://r1n1.sj.us:8443/rest/entity/[email protected] job The job commands enable you to manipulate information about the Hadoop jobs that are running on your cluster: job changepriority - Changes the priority of a specific job. job kill - Kills a specific job. job linklogs - Uses Centralized Logging to create symbolic links to all the logs relating to the activity of a specific job. job table - Retrieves detailed information about the jobs running on the cluster. job changepriority Changes the priority of the specified job. Syntax CLI maprcli job changepriority [ -cluster cluster name ] -jobid job ID -priority NORMAL|LOW|VERY_LOW|HIGH|VERY_HIGH REST http[s]://:/rest/job/changepriority? Parameters Parameter Description cluster Cluster name jobid Job ID priority New job priority Examples Changing a Job's Priority: CLI REST maprcli job changepriority -jobid job_201120603544_8282 -priority LOW https://r1n1.sj.us:8443/rest/job/changepriority?jobid=job_201120603544_8282&priority=LOW job kill The job kill command kills the specified job. Syntax CLI REST maprcli job kill [ -cluster cluster name ] -jobid job ID http[s]://:/rest/job/kill?[cluster=cluster_name&]jobid=job_ID Parameters Parameter Description cluster Cluster name jobid Job ID Examples Killing a Job CLI maprcli job kill -jobid job_201120603544_8282 REST https://r1n1.sj.us:8443/rest/job/kill?jobid=job_201120603544_8282 job linklogs The maprcli job linklogs command performs Centralized Logging, which provides a job-centric view of all log files generated by tracker nodes during job execution. The output of job linklogs is a directory populated with symbolic links to all log files related to tasks, map attempts, and reduce attempts pertaining to the specified job(s). The command can be performed during or after a job. Syntax CLI maprcli job linklogs -jobid -todir REST http[s]://:/rest/job/linklogs?jobid=&todir= Parameters Parameter Description jobid A regular expression specifying the target jobs. todir The target location to dump the Centralized Logging output directories. Output The following directory structure will be created in the location specified by todir for all jobids matching the jobid parameter. /hosts// contains symbolic links to log directories of tasks executed for on /mappers/ contains symbolic links to log directories of all map task attempts for across the whole cluster /reducers/ contains symbolic links to log directories of all reduce task attempts for across the whole cluster Examples Link logs for all jobs named "wordcount1" and dump output to /myvolume/joblogviewdir: CLI REST maprcli job linklogs -jobid job_*_wordcount1 -todir /myvolume/joblogviewdir https://r1n1.sj.us:8443/rest/job/linklogs?jobid=job_*_wordcount1&todir=/myvolume/joblogviewdir job table Retrieves histograms and line charts for job metrics. Use the job table API to retrieve job metrics for your job. The metrics data can be formatted for histogram display or line chart display. Syntax REST http[s]://:/rest/job/table?output=terse&filter=string&chart=chart_type&columns=list_of_columns&sca Parameters Parameter Description filter Filters results to match the value of a specified string. chart Chart type to use: line for a line chart, bar for a histogram. columns Comma-separated list of column names to return. bincount Number of histogram bins. scale Scale to use for the histogram. Specify linear for a linear scale and log for a logarithmic scale. Column Names Parameter Description Notes jmadavg Job Average Map Attempt Duration jradavg Job Average Reduce Attempt Duration jtadavg Job Average Task Duration Filter Only jcmtct Job Complete Map Task Count Filter Only jcrtct Job Complete Reduce Task Count Filter Only jctct Job Complete Task Count Filter Only jccpu Job Cumulative CPU jcmem Job Cumulative Physical Memory jcpu Job Current CPU Filter Only jmem Job Current Memory Filter Only jfmtact Job Failed Map Task Attempt Count jfmtct Job Failed Map Task Count jfrtact Job Failed Reduce Task Attempt Count jfrtct Job Failed Reduce Task Count jftact Job Failed Task Attempt Count Filter Only jftct Job Failed Task Count Filter Only jmibps Job Map Input Bytes Rate Per-second throughput rate jmirps Job Map Input Records Rate Per-second throughput rate jmobps Job Map Output Bytes Rate Per-second throughput rate jmorps Job Map Output Records Rate Per-second throughput rate jmtact Job Mask Task Attempt Count Per-second throughput rate jmtct Job Map Task Count jmadmax Job Maximum Map Attempt Duration jradmax Job Maximum Reduce Attempt Duration jtadmax Job Maximum Task Duration jribps Job Reduce Input Bytes Rate jrirps Job Reduce Input Records Rate jrobps Job Reduce Output Bytes Rate jrorps Job Reduce Output Records Rate jrsbps Job Reduce Shuffle Bytes Rate jrtact Job Reduce Task Attempt Count jrtct Job Reduce Task Count jrumtct Job Running Map Task Count Filter Only jrurtct Job Running Reduce Task Count Filter Only jrutct Job Running Task Count Filter Only jtact Job Task Attempt Count Filter Only jtct Job Total Task Count Filter Only jd Job Duration Histogram Only jid Job ID Filter Only jn Job Name Filter Only ju Job User Filter Only js Job Status Filter Only jmcmem Job Map Cumulative Memory Bytes Histogram Only jrcmem Job Reduce Cumulative Memory Bytes Histogram Only jpri Job Priority Filter Only jmpro Job Map Progress Filter Only jrpro Job Reduce Progress Filter Only jmtst Map Tasks Start Time jmtft Map Tasks Finish Time jrtst Reduce Tasks Start Time jrtft Reduce Tasks Finish Time Filter Only jsbt Job Submit Time Filter Only jst Job Start Time Filter Only jft Job Finish Time Filter Only jrrsw Job Reduce Reserve Slot Wait jmrsw Job Map Reserve Slot Wait jdlmt Job Data-local Map Tasks jolmt Job Non-local Map Tasks jrlmt Job Rack-local Map Tasks jrtd Job Reduce Tasks Duration jmtd Job Map Tasks Duration jmmbr Job MapR-FS Map Bytes Read jrmbr Job MapR-FS Reduce Bytes Read jtmbr Job MapR-FS Total Bytes Read jmmbw Job MapR-FS Map Bytes Written jrmbw Job MapR-FS Reduce Bytes Written jtmbw Job MapR-FS Total Bytes Written jmfbw Job Map File Bytes Written jrfbw Job Reduce File Bytes Written jtfbw Job Total File Bytes Written jmir Cumulative Job Map Input Records jrir Cumulative Job Reduce Input Records jcir Cumulative Job Combine Input Records jmor Cumulative Job Map Output Records jror Cumulative Job Reduce Output Records jcor Job Combine Output Records jrsb Cumulative Job Reduce Shuffle Bytes jmsr Job Map Spilled Records jrsr Job Reduce Spilled Records jtsr Job Total Spilled Records jmob Cumulative Job Map Output Bytes jmib Cumulative Job Map Input Bytes jmcpu Job Map CPU jrcpu Job Reduce CPU jmsrb Job Map Split Raw Bytes jrsrb Job Reduce Split Raw Bytes jsrb Job Split Raw Bytes jrig Job Reduce Input Groups jvmb Virtual Memory Bytes jmvmb Job Map Virtual Memory Bytes jrvmb Job Reduce Virtual Memory Bytes jgct Job Total GC Time jmgct Job Map GC Time jrgct Job Reduce GC Time Examples Retrieve a Histogram: REST https://centos-n1.user.prv:8443/rest/job/table?columns=jn,jid,jd&filter=[jn==*select*]and[jst>1404246318000] CURL curl -d @json https://r1n1.sj.us:8443/rest/job/table REST Example Output {"total":4,"status":"OK","data":[["select disks from loadtest(Stage-1)","job_201406121453_0017","4088"],["select id from usertable(Stage-1)","job_201406121453_0016","1"],["select disks from loadtest(Stage-1)","job_201406121453_0015","3865"],["select disks from loadtest(Stage-1)","job_201406121453_0014","1481"]],"columns":["jn","jid","jd"]} CURL Example Request and Response In the curl example above, the json file contains a URL-encoded version of the information in the Request section below. Request GENERAL_PARAMS: { [chart: "bar"|"line"], columns: , [filter: "[{operator}]and[...]",] [output: terse,] [start: int,] [limit: int] } REQUEST_PARAMS_HISTOGRAM: { chart:bar columns:jd filter: } REQUEST_PARAMS_LINE: { chart:line, columns:jmem, filter: NOT PARSED, UNUSED IN BACKEND } REQUEST_PARAMS_GRID: { columns:jid,jn,js,jd filter: output:terse, start:0, limit:50 } Response RESPONSE_SUCCESS_HISTOGRAM: { "status" : "OK", "total" : 15, "columns" : ["jd"], "binlabels" : ["0-5s","5-10s","10-30s","30-60s","60-90s","90s-2m","2m-5m","5m-10m","10m-30m","30m-1h ","1h-2h","2h-6h","6h-12h","12h-24h",">24h"], "binranges" : [ [0,5000], [5000,10000], [10000,30000], [30000,60000], [60000,90000], [90000,120000], [120000,300000], [300000,600000], [600000,1800000], [1800000,3600000], [3600000,7200000], [7200000,21600000], [21600000,43200000], [43200000,86400000], [86400000] ], "data" : [33,919,1,133,9820,972,39,2,44,80,11,93,31,0,0] } RESPONSE_SUCCESS_GRID: { "status": "OK", "total" : 50, "columns" : ["jid","jn","ju","jg","js","jcpu","jccpu","jmem","jcmem","jpri","jmpro","jrpro","jsbt" ,"jst","jft","jd","jmtct","jfmtact","jrtct","jmtact","jrtact","jtact","jfrtact","jftac t","jfmtct","jfrtct","jftct","jtct","jrumtct","jrurtct","jrutct","jctct","jcmtct","jcr tct","jmirps","jmorps","jmibps","jmobps","jrirps","jrorps","jribps","jrobps","jtadavg" ,"jmadavg","jmadmax","jtadmax","jradavg","jradmax"], "data" : [ ["job_201210216041_7311","Billboard Top 10","heman","jobberwockies","PREP",69,9106628,857124,181087410,"LOW",30,48,13099922755 80,1316685654403,1324183149687,7497495284,72489,25227,6171223,95464,6171184,6266648,-3 8,25189,13115,-4,13111,6243712,5329,4,6243712,6225268,54045,6171223,403,128,570,137,17 2,957,490,179,246335,367645,1758151,1758151,125024,514028], ["job_201129309372_8897","Super Big","srivas","jobberwockies","KILLED",59,3125830,2895159,230270693,"LOW",91,1,1313111 819653,1323504739893,1326859602015,3354862122,8705980,3774739,7691269,12515000,1627363 1,28788631,8196156,11970895,2706470,4365698,7072168,16397249,215570,35,16397249,910947 6,5783940,3325536,707,509,345,463,429,93,88,752,406336,553455,3392429,3392429,259216,5 11285], ["job_201165490737_7144","Trending Human Interaction","mickey","jobberwockies","PREP",100,1304791,504092,728635524,"VERY_LOW",5 7,90,1301684627596,1331548331890,1331592957521,44625631,7389503,3770494,5433308,110114 95,15048822,26060317,9769362,13539856,2544349,4315172,6859521,12822811,21932,327,12822 811,5941031,4823222,1117809,739,654,561,426,925,23,420,597,292024,470314,1854688,18546 88,113733,566672], ["job_201152533959_6159","Star Search","darth","fbi","FAILED",82,7151113,2839682,490527441,"NORMAL",51,61,13053670422 24,1325920952001,1327496965896,1576013895,8939964,2041524,4965024,10795895,8786681,195 82576,3924842,5966366,1130482,3422544,4553026,13904988,833761,2,13904988,8518199,69757 21,1542478,665,916,10,34,393,901,608,916,186814,331708,2500504,2500504,41920,251453] ] } RESPONSE_SUCCESS_LINE: { "status" : "OK", "total" : 22, "columns" : ["jcmem"], "data" : [ [1329891055016,0], [1329891060016,8], [1329891065016,16], [1329891070016,1024], [1329891075016,2310], [1329891080016,3243], [1329891085016,4345], [1329891090016,7345], [1329891095016,7657], [1329891100016,8758], [1329891105016,9466], [1329891110016,10345], [1329891115016,235030], [1329891120016,235897], [1329891125016,287290], [1329891130016,298390], [1329891135016,301355], [1329891140016,302984], [1329891145016,303985], [1329891150016,304403], [1329891155016,503030], [1329891160016,983038] ] } license The license commands let you work with MapR licenses: license add - adds a license license addcrl - adds a certificate revocation list (CRL) license apps - displays the features included in the current license license list - lists licenses on the cluster license listcrl - lists CRLs license remove - removes a license license showid - displays the cluster ID license add Adds a license. Permissions required: fc or a The license can be specified either by passing the license string itself to license add, or by specifying a file containing the license string. Syntax CLI maprcli license add [ -cluster ] [ -is_file true|false ] -license REST http[s]://:/rest/license/add? Parameters Parameter Description cluster The cluster on which to run the command. is_file Specifies whether the license specifies a file. If false, the license parameter contains a long license string. license The license to add to the cluster. If -is_file is true, license specifies the filename of a license file. Otherwise, license cont ains the license string itself. Examples Adding a License from a File Assuming a file /tmp/license.txt containing a license string, the following command adds the license to the cluster. CLI maprcli license add -is_file true -license /tmp/license.txt license addcrl Adds a certificate revocation list (CRL). Permissions required: fc or a Syntax CLI REST maprcli license addcrl [ -cluster ] -crl [ -is_file true|false ] http[s]://:/rest/license/addcrl? Parameters Parameter Description cluster The cluster on which to run the command. crl The CRL to add to the cluster. If file is set, crl specifies the filename of a CRL file. Otherwise, crl contains the CRL string itself. is_file Specifies whether the license is contained in a file. license apps Displays the features authorized for the current license. Permissions required: login Syntax CLI REST maprcli license apps [ -cluster ] http[s]://:/rest/license/apps? Parameters Parameter Description cluster The cluster on which to run the command. license list Lists licenses on the cluster. Permissions required: login Syntax CLI maprcli license list [ -cluster ] REST http[s]://:/rest/license/list? Parameters Parameter Description cluster The cluster on which to run the command. license listcrl Lists certificate revocation lists (CRLs) on the cluster. Permissions required: login Syntax CLI REST maprcli license listcrl [ -cluster ] http[s]://:/rest/license/listcrl? Parameters Parameter Description cluster The cluster on which to run the command. license remove Removes a license. Permissions required: fc or a Syntax CLI REST maprcli license remove [ -cluster ] -license_id http[s]://:/rest/license/remove? Parameters Parameter Description cluster The cluster on which to run the command. license_id The license to remove. license showid Displays the cluster ID for use when creating a new license. Permissions required: login Syntax CLI maprcli license showid [ -cluster ] REST http[s]://:/rest/license/showid? Parameters Parameter Description cluster The cluster on which to run the command. Metrics API A Hadoop job sets the rules that the JobTracker service uses to break an input data set into discrete tasks and assign those tasks to individual nodes. The MapR Metrics service provides two API calls that enable you to retrieve grids of job data or task attempt data depending on the parameters you send: /api/job/table retrieves information about the jobs running on your cluster. You can use this API to retrieve information about the number of task attempts for jobs on the cluster, job duration, job computing resource use (CPU and memory), and job data throughput (both records and bytes per second). /api/task/table retrieves information about the tasks that make up a specific job, as well as the specific task attempts. You can use this API to retrieve information about a task attempt's data throughput, measured in number of records per second as well as in bytes per second. Both of these APIs provide robust filtering capabilities to display data with a high degree of specificity. nagios The nagios generate command generates a topology script for Nagios nagios generate Generates a Nagios Object Definition file that describes the cluster nodes and the services running on each. Syntax CLI REST Parameters maprcli nagios generate [ -cluster ] http[s]://:/rest/nagios/generate? Parameter Description cluster The cluster on which to run the command. Output Sample Output ############# Commands ############# define command { command_name command_line } check_fileserver_proc $USER1$/check_tcp -p 5660 define command { command_name command_line } check_cldb_proc $USER1$/check_tcp -p 7222 define command { command_name command_line } check_jobtracker_proc $USER1$/check_tcp -p 50030 define command { command_name command_line } check_tasktracker_proc $USER1$/check_tcp -p 50060 define command { command_name command_line } check_nfs_proc $USER1$/check_tcp -p 2049 define command { command_name command_line } check_hbmaster_proc $USER1$/check_tcp -p 60000 define command { command_name command_line } check_hbregionserver_proc $USER1$/check_tcp -p 60020 define command { command_name command_line } check_webserver_proc $USER1$/check_tcp -p 8443 ################# HOST: host1 ############### define host { use linux-server host_name host1 address 192.168.1.1 check_command check-host-alive } ################# HOST: host2 ############### define host { use linux-server host_name host2 address 192.168.1.2 check_command check-host-alive } Examples Generate a nagios configuration, specifying cluster name and ZooKeeper nodes: CLI maprcli nagios generate -cluster cluster-1 REST https://host1:8443/rest/nagios/generate?cluster=cluster-1 Generate a nagios configuration and save to the file nagios.conf: CLI maprcli nagios generate > nagios.conf nfsmgmt The nfsmgmt refreshexports command refreshes the NFS exports on the specified host and/or port. nfsmgmt refreshexports Refreshes the list of clusters and mount points available to mount with NFS. Permissions required: fc or a. Syntax CLI maprcli nfsmgmt refreshexports [ -nfshost ] [ -nfsport ] REST http[s]://<:port>/rest/nfsmgmt/refreshexports? Parameters Parameter Description nfshost The hostname of the node that is running the MapR NFS server. nfsport The port to use. node The node commands let you work with nodes in the cluster: node allow-into-cluster node cldbmaster node heatmap node list node listcldbs node listcldbzks node listzookeepers node maintenance node metrics node move node remove node services node topo Fields The following table lists the data fields that provide information about each node. Each field has two names: Field name - displayed in the output of the node list command Short name - used to specify the columns displayed using the columns parameter The short name is also used when specifying rows with a filter, for example when specifying nodes on which to perform an action with the node services command. Field Name Short Name Description blockMovesIn bmi Block moves in. blockMovesOut bmo Block moves out. bytesReceived br Bytes received by the node since the last CLDB heartbeat. bytesSent bs Bytes sent by the node since the last CLDB heartbeat. configuredservice csvc Services that are configured as roles on the node. CorePresentAlarm ncp Cores Present Alarm (NODE_ALARM_CORE_PRESENT): 0 = Clear 1 = Raised cpus cpc The total number of CPUs on the node. davail dsa Disk space available on the node. DiskFailureAlarm fda Failed Disks alarm (DISK_FAILURE): 0 = Clear 1 = Raised disks dsc Total number of disks on the node. dreadK drk Disk Kbytes read since the last heartbeat. dreads dro Disk read operations since the last heartbeat. dtotal dst Total disk space on the node. dused dsu Disk space used on the node. dwriteK dwk Disk Kbytes written since the last heartbeat. dwrites dwo Disk write ops since the last heartbeat. faileddisks nfd Number of failed MapR-FS disks on the node. 0 = Clear 1 = Raised fs-heartbeat fhb Time since the last heartbeat to the CLDB, in seconds. gw gw The gateway address. health h Overall node health, calculated from various alarm states: 0 = Healthy 1 = Needs attention 2 = Degraded 3 = Maintenance 4 = Critical healthDesc hd The health description. HighMfsMemoryAlarm nhmm The High FileServer Memory Alarm (NODE_ALARM_HIGH_MFS_MEMORY): 0 = Clear 1 = Raised HomeMapRFullAlarm hmf Installation Directory Full Alarm (NODE_ALARM_OPT_MAPR_FULL): 0 = Clear 1 = Raised hostname hn The host name. id id The node ID. ip ip A list of IP addresses associated with the node. jt-heartbeat jhb Time since the last heartbeat to the JobTracker, in seconds. LogLevelAlarm lla Excessive Logging Alarm (NODE_ALARM_DEBUG_LOGGING): 0 = Clear 1 = Raised mac mac The MAC address. MapRfs disks nmd Number of disks for use by MapR-FS maxContainersThreshold mct Max containers threshold. mtotal mt Total memory, in GB. mused mu Memory used, in GB. nm nm The netmask NodeDuplicateHostIdAlarm ndh The Duplicate Host ID alarm (NODE_ALARM_DUPLICATE_HOSTID): 0 = Clear 1 = Raised NodeNoHeartbeatAlarm nha Node no heartbeat alarm (): 0 = Clear 1 = Raised NodeM7ConfigMismatchAlarm nmc Node M7 Configuration Mismatch alarm (NODE_ALARM_M7_CONFIG_MISMATCH): 0 = Clear 1 = Raised NodeMaprUserMismatch nma Node MapR User Mismatch alarm (NODE_ALARM_MAPRUSER_MISMATCH): 0 = Clear 1 = Raised NodeMetricsWriteProblemAlarm nmw Node Metrics Write Problem Alarm (NODE_ALARM_METRICS_WRITE_PROBLEM): 0 = Clear 1 = Raised numResyncSlots nrs The number of resync slots. NodeTooManyContainersAlarm nmc Node Too Many Containers alarm (): 0 = Clear 1 = Raised PamMisconfiguredAlarm pma PAM misconfigured alarm (NODE_ALARM_PAM_MISCONFIGURED): 0 = Clear 1 = Raised RootPartitionFullAlarm rpf Root Partition Full Alarm (NODE_ALARM_ROOT_PARTITION_FULL): 0 = Clear 1 = Raised rpcin rpi RPC bytes received since the last heartbeat. rpcout rpo RPC bytes sent since the last heartbeat. rpcs rpc Number of RPCs since the last heartbeat. service svc A comma-separated list of services running on the node: cldb - CLDB fileserver - MapR-FS jobtracker - JobTracker tasktracker - TaskTracker hbmaster - HBase Master hbregionserver - HBase RegionServer nfs - NFS Gateway Example: "cldb,fileserver,nfs" ServiceCLDBDownNotRunningAlarm sca CLDB Service Down Alarm (NODE_ALARM_SERVICE_CLDB_DOWN) 0 = Clear 1 = Raised ServiceFileserverDownNotRunningAlarm sfsa Fileserver Service Down Alarm (NODE_ALARM_SERVICE_FILESERVER_DOWN) 0 = Clear 1 = Raised ServiceHBMasterDownNotRunningAlarm shma HBase Master Service Down Alarm (NODE_ALARM_SERVICE_HBMASTER_DOWN) 0 = Clear 1 = Raised ServiceHBRegionDownNotRunningAlarm shra HBase Regionserver Service Down Alarm" (NODE_ALARM_SERVICE_HBREGION_DOWN) 0 = Clear 1 = Raised ServiceHoststatsDownNotRunningAlarm sha Hoststats Service Down Alarm (NODE_ALARM_SERVICE_HOSTSTATS_DOWN) 0 = Clear 1 = Raised ServiceJTDownNotRunningAlarm sja Jobtracker Service Down Alarm (NODE_ALARM_SERVICE_JT_DOWN) 0 = Clear 1 = Raised ServiceMiscDownNotRunningAlarm (): 0 = Clear 1 = Raised ServiceNFSDownNotRunningAlarm sna NFS Service Down Alarm (NODE_ALARM_SERVICE_NFS_DOWN): 0 = Clear 1 = Raised ServiceTTDownNotRunningAlarm sta Tasktracker Service Down Alarm (NODE_ALARM_SERVICE_TT_DOWN): 0 = Clear 1 = Raised ServicesWebserverDownNotRunningAlarm swa Webserver Service Down Alarm (NODE_ALARM_SERVICE_WEBSERVER_DOWN) 0 = Clear 1 = Raised switchtopo sp The switch path. TimeSkewAlarm tsa Time Skew alarm (NODE_ALARM_TIME_SKEW): 0 = Clear 1 = Raised TTLocaldirFullAlarm tla TaskTracker local directory full alarm (NODE_ALARM_TT_LOCALDIR_FULL): 0 = Clear 1 = Raised racktopo rp The rack path. ttmapSlots tms TaskTracker map slots. ttmapUsed tmu TaskTracker map slots used. ttReduceSlots trs TaskTracker reduce slots. ttReduceUsed tru TaskTracker reduce slots used. uptime cpt Date when the node came up. utilization cpu CPU use percentage since the last heartbeat. VersionMismatchAlarm vma Software Version Mismatch Alarm (NODE_ALARM_VERSION_MISMATCH): 0 = Clear 1 = Raised vip vip The virtuap IP address. vipe vipe The end of the virtual IP address range add-to-cluster Allows host IDs to join the cluster after duplicates have been resolved. When the CLDB detects duplicate nodes with the same host ID, all nodes with that host ID are removed from the cluster and prevented from joining it again. After making sure that all nodes have unique host IDs, you can use the node allow-into-cluster command to un-ban the host ID that was previously duplicated among several nodes. Syntax CLI REST maprcli node allow-into-cluster [ -hostids ] http[s]://:/rest/node/allow-into-cluster? Parameters Parameter Description hostids A comma-separated list of host IDs. Examples Allow former duplicate host IDs node1 and node2 to join the cluster: CLI REST maprcli node allow-into-cluster -hostids node1,node2 https://r1n1.sj.us:8443/rest/node/allow-into-cluster?hostids=node1,node2 node allow-into-cluster Allows host IDs to join the cluster after duplicates have been resolved. When the CLDB detects duplicate nodes with the same host ID, all nodes with that host ID are removed from the cluster and prevented from joining it again. After making sure that all nodes have unique host IDs, you can use the node allow-into-cluster command to un-ban the host ID that was previously duplicated among several nodes. Syntax CLI maprcli node allow-into-cluster [ -hostids ] REST http[s]://:/rest/node/allow-into-cluster? Parameters Parameter Description hostids A comma-separated list of host IDs. Examples Allow former duplicate host IDs node1 and node2 to join the cluster: CLI REST maprcli node allow-into-cluster -hostids node1,node2 https://r1n1.sj.us:8443/rest/node/allow-into-cluster?hostids=node1,node2 node cldbmaster Returns the address of the master CLDB node. The node cldbmaster API returns the server ID and hostname of the node serving as the CLDB master node. Syntax CLI maprcli node cldbmaster [ -cluster ] REST http[s]://:/rest/node/cldbmaster? Parameters Parameter Description cluster name The name of the cluster for which to return the CLDB master node information. Examples Return the CLDB master node information for the cluster my.cluster.com: CLI REST maprcli node cldbmaster -cluster my.cluster.com https://r1n1.sj.us:8443/rest/node/cldbmaster?cluster=my.cluster.com node heatmap Displays a heatmap for the specified nodes. Syntax CLI REST maprcli node heatmap [ -cluster ] [ -filter ] [ -view ] http[s]://:/rest/node/heatmap? Parameters Parameter Description cluster The cluster on which to run the command. filter A filter specifying snapshots to preserve. See Filters for more information. view Name of the heatmap view to show: status = Node status: 0 = Healthy 1 = Needs attention 2 = Degraded 3 = Maintenance 4 = Critical cpu = CPU utilization, as a percent from 0-100. memory = Memory utilization, as a percent from 0-100. diskspace = MapR-FS disk space utilization, as a percent from 0-100. DISK_FAILURE = Status of the DISK_FAILURE alarm. 0 if clear, 1 if raised. SERVICE_NOT_RUNNING = Status of the SERVICE_NOT_RUNNING alarm. 0 if clear, 1 if raised. CONFIG_NOT_SYNCED = Status of the CONFIG_NOT_SYNCED alarm. 0 if clear, 1 if raised. Output Description of the output. { status:"OK", data: [{ "{{rackTopology}}" "{{nodeName}}" : "{{nodeName}}" : "{{nodeName}}" : ... }, "{{rackTopology}}" "{{nodeName}}" : "{{nodeName}}" : "{{nodeName}}" : ... }, ... }] : { {{heatmapValue}}, {{heatmapValue}}, {{heatmapValue}}, : { {{heatmapValue}}, {{heatmapValue}}, {{heatmapValue}}, } Output Fields Field Description rackTopology The topology for a particular rack. nodeName The name of the node in question. heatmapValue The value of the metric specified in the view parameter for this node, as an integer. Examples Display a heatmap for the default rack: CLI maprcli node heatmap REST https://r1n1.sj.us:8443/rest/node/heatmap Display memory usage for the default rack: CLI maprcli node heatmap -view memory REST https://r1n1.sj.us:8443/rest/node/heatmap?view=memory node list Lists nodes in the cluster. You can specify a set of nodes for which to retrieve information in several ways: To list only nodes with raised alarms, set alarmednodes to 1. To list only NFS nodes, set nfsnodes to 1. To window the list, use the start and limit options to select only a portion of the results. To list nodes that match certain criteria, pass a filter to the filter parameter. See the Fields table on the node page for the fields available to filter. See the API Reference page for information on filters. Syntax CLI maprcli node list [ -alarmednodes 1 ] [ -cluster ] [ -filter ] [ -limit ] [ -nfsnodes 1 ] [ -output terse|verbose ] [ -start ] [ -zkconnect ] REST http[s]://:/rest/node/list? Parameters Parameter Description alarmednodes If set to 1, displays only nodes with raised alarms. Cannot be used if nfsnodes is set. cluster The cluster on which to run the command. columns A comma-separated list of fields to return in the query, specified by the short names. See the Fields table below. filter A filter specifying nodes on which to start or stop services. See the Fields table on the node page for the fields available to filter. See the API Reference page for information on filters. limit The number of rows to return, beginning at start. Default: 0 nfsnodes If set to 1, displays only nodes running NFS. Cannot be used if alarmednodes is set. output Specifies whether the output should be terse or verbose. start The offset from the starting row according to sort. Default: 0 zkconnect ZooKeeper Connect String Output Information about the nodes.See the Fields table below. Sample Output bytesSent dreads davail TimeSkewAlarm servicesHoststatsDownAlarm ServiceHBMasterDownNotRunningAlarm ServiceNFSDownNotRunningAlarm ttmapUsed DiskFailureAlarm mused id mtotal cpus utilization rpcout ttReduceSlots ServiceFileserverDownNotRunningAlarm ServiceCLDBDownNotRunningAlarm dtotal jt-heartbeat ttReduceUsed dwriteK ServiceTTDownNotRunningAlarm ServiceJTDownNotRunningAlarm ttmapSlots dused uptime hostname health disks faileddisks fs-heartbeat rpcin ip dreadK dwrites ServiceWebserverDownNotRunningAlarm rpcs LogLevelAlarm ServiceHBRegionDownNotRunningAlarm bytesReceived service topo(rack) MapRfs disks ServiceMiscDownNotRunningAlarm VersionMismatchAlarm 8300 0 269 0 0 0 0 75 0 4058 6394230189818826805 7749 4 3 141 50 0 0 286 2 10 32 0 0 100 16 Thu Jan 15 16:58:57 PST 1970 whatsup 0 1 0 0 51 10.250.1.48 0 2 0 0 0 0 8236 /third/rack/whatsup 1 0 0 Fields For definitions of the output fields, and short names for use with filters, see the Fields table on the node command page. Examples List all nodes: CLI maprcli node list REST https://r1n1.sj.us:8443/rest/node/list List the health of all nodes: CLI maprcli node list -columns service,health REST https://r1n1.sj.us:8443/rest/node/list?columns=service,health List the number of slots on all nodes: CLI maprcli node list -columns ip,ttmapSlots,ttmapUsed,ttReduceSlots,ttReduceUsed REST https://r1n1.sj.us:8443/rest/node/list?columns=ip,ttmapSlots,ttmapUsed,ttReduceSlots,ttReduceUsed List nodes on a particular subnet: CLI maprcli node list -filter [ip==192.168.10.*] REST https://r1n1.sj.us:8443/rest/node/list?filter=%5Bip%3D%3D192.168.10.*%5D node listcldbs The node listcldbs API returns the hostnames of the nodes in the cluster that are running the CLDB service. Syntax CLI REST maprcli node listcldbs [ -cluster ] [ -cldb ] http[s]://:/rest/node/listcldbs? Parameters Parameter Description cluster name The name of the cluster for which to return the list of CLDB node hostnames. cldb hostname|ip:port The hostname or IP address and port number of a CLDB node. Examples Return the list of CLDB nodes for the cluster my.cluster.com: CLI REST maprcli node listcldbs -cluster my.cluster.com https://r1n1.sj.us:8443/rest/node/listcldbs?cluster=my.cluster.com node listcldbzks The node listcldbzks API returns the hostnames of the nodes in the cluster that are running the CLDB service and the IP addresses and port numbers for the nodes in the cluster that are running the ZooKeeper service. Syntax CLI maprcli node listcldbzks [ -cluster ] [ -cldb ] REST http[s]://:/rest/node/listcldbzks? Parameters Parameter Description cluster name The name of the cluster for which to return the CLDB and ZooKeeper information. cldb hostname|ip:port The hostname or IP address and port number of a CLDB node. Examples Return CLDB and ZooKeeper node information for the cluster my.cluster.com: CLI REST maprcli node listcldbzks -cluster my.cluster.com https://r1n1.sj.us:8443/rest/node/listcldbzks?cluster=my.cluster.com node listzookeepers The node listzookeepers API returns the hostnames of the nodes in the cluster that are running the zookeeper service. Syntax CLI REST maprcli node listzookeepers [ -cluster ] [ -cldb ] http[s]://:/rest/node/listzookeepers? Parameters Parameter Description cluster name The name of the cluster for which to return the list of zookeeper node hostnames. cldb hostname|ip:port The hostname or IP address and port number of a valid CLDB node. The other CLDB nodes and zookeeper nodes can be discovered from this node. Examples Return the list of zookeeper nodes for the cluster my.cluster.com If you know that the CLDB service is running on a node with hostname host1, you can enter: CLI REST maprcli node listzookeepers -cluster my.cluster.com -cldb host1 https://r1n1.sj.us:8443/rest/node/listzookeepers?cluster=my.cluster.com&cldb=host1 node maintenance Places a node into a maintenance mode for a specified timeout duration. For the duration of the timeout, the cluster's CLDB does not consider this node's data as lost and does not trigger a resync of the data on this node. See Nodes for more information. Syntax CLI maprcli node maintenance [ -cluster ] [ -serverids ] -nodes -timeoutminutes minutes REST http[s]://:/rest/node/maintenance? Parameters Parameter Description cluster The cluster on which to run the command. serverids List of server IDs nodes List of nodes timeoutminutes Duration of timeout in minutes node metrics Retrieves metrics information for nodes in a cluster. Use the node metrics API to retrieve node data for your job. Syntax CLI maprcli node metrics -nodes -start start_time -end end_time [ -json ] [ -interval interval ] [ -events ] [ -columns columns ] [ -cluster cluster name ] Parameters Parameter Description nodes A space-separated list of node names. start The start of the time range. Can be a UTC timestamp or a UTC date in MM/DD/YY format. end The end of the time range. Can be a UTC timestamp or a UTC date in MM/DD/YY format. json Specify this flag to return data as a JSON object. interval Data measurement interval in seconds. The minimum value is 10 seconds. events Specify TRUE to return node events only. The default value of this parameter is FALSE. columns Comma-separated list of column names to return. cluster Cluster name. Column Name Parameters The node metrics API always returns the NODE (node name), TIMESTAMPSTR (timestamp string), and TIMESTAMP (integer timestamp) columns. Use the -columns flag to specify a comma-separated list of column names to return. The CPUNICE, CPUUSER, and CPUSYSTEM parameters return information in jiffies. This unit measures one tick of the system timer interrupt and is usually equivalent to 10 milliseconds, but may vary depending on your particular node configuration. Call sysconf(_SC _CLK_TCK) to determine the exact value for your node. Parameter Description CPUNICE Amount of CPU time used by processes with a positive nice value. CPUUSER Amount of CPU time used by user processes. CPUSYSTEM Amount of CPU time used by system processes. LOAD5PERCENT Percentage of time this node spent at load 5 or below LOAD1PERCENT Percentage of time this node spent at load 1 or below MEMORYCACHED Memory cache size in bytes MEMORYSHARED Shared memory size in bytes MEMORYBUFFERS Memory buffer size in bytes MEMORYUSED Memory used in megabytes PROCRUN Number of processes running RPCCOUNT Number of RPC calls RPCINBYTES Number of bytes passed in by RPC calls RPCOUTBYTES Number of bytes passed out by RPC calls SERVAVAILSIZEMB Server storage available in megabytes SERVUSEDSIZEMB Server storage used in megabytes SWAPFREE Free swap space in bytes TTMAPUSED Number of TaskTracker slots used for map tasks TTREDUCEUSED Number of TaskTracker slots used for reduce tasks Notes Three column name parameters return data that is too granular to display in a standard table. Use the -json option to return this information as a JSON object. Parameter Description Metrics Returned CPUS Activity on this node's CPUs. Each CPU on the node is numbered from zero, cpu0 to cpuN. Metrics returned are for each CPU. CPUIDLE: Amount of CPU time spent idle. Reported as jiffies. CPUIOWAIT: Amount of CPU time spent waiting for I/O operations. Reported as jiffies. CPUTOTAL: Total amount of CPU time. Reported as jiffies. DISKS Activity on this node's disks. Metrics returned are for each partition. READOPS: Number of read operations. READKB: Number of kilobytes read. WRITEOPS: Number of write operations. WRITEKB: Number of kilobytes written. NETWORK Activity on this node's network interfaces. Metrics returned are for each interface. BYTESIN: Number of bytes received. BYTESOUT: Number of bytes sent. PKTSIN: Number of packets received. PKTSOUT: Number of packets sent. Examples To retrieve the percentage of time that a node spent at the 1 and 5 load levels: [user@host ~]# maprcli node metrics -nodes my.node.lab -start 07/25/12 -end 07/26/12 -interval 7200 -columns LOAD1PERCENT,LOAD5PERCENT NODE my.node.lab my.node.lab my.node.lab my.node.lab my.node.lab my.node.lab my.node.lab my.node.lab my.node.lab my.node.lab my.node.lab my.node.lab LOAD5PERCENT 11 10 15 18 28 35 30 36 37 12 22 LOAD1PERCENT 18 23 46 34 70 84 35 62 44 28 38 TIMESTAMPSTR Wed Jul 25 12:52:40 Wed Jul 25 14:52:50 Wed Jul 25 16:52:50 Wed Jul 25 18:52:57 Wed Jul 25 20:52:58 Wed Jul 25 22:53:01 Thu Jul 26 00:53:01 Thu Jul 26 02:53:03 Thu Jul 26 04:53:10 Thu Jul 26 06:53:14 Thu Jul 26 08:53:21 Thu Jul 26 10:53:30 PDT PDT PDT PDT PDT PDT PDT PDT PDT PDT PDT PDT 2012 2012 2012 2012 2012 2012 2012 2012 2012 2012 2012 2012 TIMESTAMP 1343245960047 1343253170000 1343260370000 1343267577000 1343274778000 1343281981000 1343289181000 1343296383000 1343303590000 1343310794000 1343318001000 1343325210000 Sample JSON object This JSON object is returned by the following command: [user@host ~]# maprcli node metrics -nodes my.node.lab -json -start 1343290000000 -end 1343300000000 -interval 28800 -columns LOAD1PERCENT,LOAD5PERCENT,CPUS { "timestamp":1343333063869, "status":"OK", "total":3, "data":[ { "NODE":"my.node.lab", "TIMESTAMPSTR":"Wed Jul 25 18:00:05 PDT 2012", "TIMESTAMP":1343264405000, "LOAD1PERCENT":13, "LOAD5PERCENT":12 "CPUS":{ "cpu0":{ "CPUIDLE":169173957, "CPUIOWAIT":2982912, "CPUTOTAL":173897423 }, "cpu1":{ "CPUIDLE":172217855, "CPUIOWAIT":26760, "CPUTOTAL":174016589 }, "cpu2":{ "CPUIDLE":171071574, "CPUIOWAIT":4051, "CPUTOTAL":173957716 }, }, }, { "NODE":"my.node.lab", "TIMESTAMPSTR":"Thu Jul 26 02:00:08 PDT 2012", "TIMESTAMP":1343293208000, "LOAD1PERCENT":17, "LOAD5PERCENT":13 "CPUS":{ "cpu0":{ "CPUIDLE":169173957, "CPUIOWAIT":2982912, "CPUTOTAL":173897423 }, "cpu1":{ "CPUIDLE":172217855, "CPUIOWAIT":26760, "CPUTOTAL":174016589 }, "cpu2":{ "CPUIDLE":171071574, "CPUIOWAIT":4051, "CPUTOTAL":173957716 }, }, }, { "NODE":"my.node.lab", "TIMESTAMPSTR":"Thu Jul 26 10:00:08 PDT 2012", "TIMESTAMP":1343322008000, "LOAD1PERCENT":18, "LOAD5PERCENT":13 "CPUS":{ "cpu0":{ "CPUIDLE":169173957, "CPUIOWAIT":2982912, "CPUTOTAL":173897423 }, "cpu1":{ "CPUIDLE":172217855, "CPUIOWAIT":26760, "CPUTOTAL":174016589 }, "cpu2":{ "CPUIDLE":171071574, "CPUIOWAIT":4051, "CPUTOTAL":173957716 }, }, } ] } node move Moves one or more nodes to a different topology. Permissions required: fc or a. Syntax CLI maprcli node move [ -cluster ] -serverids -topology REST http[s]://:/rest/node/move? Parameters Parameter Description cluster The cluster on which to run the command. serverids The server IDs of the nodes to move. topology The new topology. To obtain the server ID, run maprcli node list -columns id . Sample output from maprcli node list -columns id is shown below. The resulting server ID(s) can be copied and pasted into the maprcl i node move command. id 547819249997313015 2130988050310536949 8683110801227243688 5056865595028557458 3111141192171195352 hostname node-34.lab node-36.lab node-37.lab node-38.lab node-39.lab ip 10.10.40.34,10.10.88.34 10.10.40.36,10.10.88.36 10.10.40.37,10.10.88.37 10.10.40.38,10.10.88.38 10.10.40.39,10.10.88.39 node remove The node remove command removes one or more server nodes from the system. Permissions required: fc or a After issuing the node remove command, wait several minutes to ensure that the node has been properly and completely removed. Syntax CLI maprcli node remove [ -filter "" ] [ -force true|false ] [ -hostids ] [ -nodes ] [ -zkconnect ] REST http[s]://:/rest/node/remove? Parameters Parameter Description filter A filter specifying nodes on which to start or stop services. See Filters for more information. force Forces the service stop operations. Default: false hostids A list of host IDs, separated by spaces. nodes A list of node names, separated by spaces. zkconnect ZooKeeper Connect String. Example: 'host:port,host:port,host:port,...'. default: localhost:5181 node services Starts, stops, or restarts services on one or more server nodes. Permissions required: ss, fc or a To start or stop services, you must specify the service name, the action (start, stop, or restart), and the nodes on which to perform the action. You can specify the nodes in either of two ways: Use the nodes parameter to specify a space-delimited list of node names. Use the filter parameter to specify all nodes that match a certain pattern. See API Reference for more information about filters. Syntax CLI maprcli node services [ -action start|stop|restart] [ -cldb start|stop|restart] [ -cluster ] [ -fileserver start|stop|restart] [ -filter ] [ -hbmaster start|stop|restart] [ -hbregionserver start|stop|restart] [ -jobtracker start|stop|restart] [ -name ] [ -nfs start|stop|restart] [ -nodes ] [ -tasktracker start|stop|restart] [ -webserver start|stop|restart] [ -zkconnect ] REST http[s]://:/rest/node/services? Parameters To perform an action on a service, on a particular set of nodes, you must specify the following three parameters: action- the action to perform; starting, stopping, or restarting a service node or filter - the nodes on which to perform the action; either a list of nodes, or a filter that matches a set of nodes name - the service on which to perform the action The following table lists the parameters available with the node services comand. Parameter Description action An action to perform on a service specified in the name parameter: Values: start, stop, or restart cldb Starts, stops, or restarts the cldb service. Values: start, stop, or restart cluster The cluster on which to run the command. fileserver Starts, stops, or restarts the fileserver service. Values: start, stop, or restart filter A filter specifying nodes on which to start or stop services. For fields to use with the filter, see the Fields table in the node list command. See Filters for more information about filters. hbmaster Starts, stops, or restarts the hbmaster service. Values: start, stop, or restart hbregionserver Starts, stops, or restarts the hbregionserver service. Values: start, stop, or restart jobtracker Starts, stops, or restarts the jobtracker service. Values: start, stop, or restart name A service on which to perform an action specified by the action parameter. Any service can be specified with this option, but the following services can only be specified with the name option: historyserver, hivemeta, hs2, httpfs, hue, nod emanager, oozie, and resourcemanager. nfs Starts, stops, or restarts the nfs service. Values: start, stop, or restart nodes A list of node names, separated by spaces. tasktracker Starts, stops, or restarts the tasktracker service. Values: start, stop, or restart webserver Starts, stops, or restarts the web server service. Values: start, stop, or restart zkconnect ZooKeeper Connect String Examples Start the Hue Service maprcli node services -name hue -action start -nodes Restart specified TaskTrackers in the Cluster To restart TaskTrackers on specific nodes, provide the list of nodes in the nodes parameter. maprcli node services -tasktracker restart -nodes Restart all TaskTrackers in the Cluster Instead of specifying a list of all nodes in the cluster that are running the TaskTracker service, you can use the filter parameter to restart every ta sktracker by matching only nodes on which the TaskTracker is configured. Remember that csvc is the short name of the configuredservi ce field. See the Fields table on the node list page. maprcli node services -tasktracker restart -filter '[csvc==tasktracker]' node topo Lists cluster topology information. Lists internal nodes only (switches/racks/etc) and not leaf nodes (server nodes). Syntax CLI maprcli node topo [ -cluster ] [ -path ] REST http[s]://:/rest/node/topo? Parameters Parameter Description cluster The cluster on which to run the command. path The path on which to list node topology. Output Node topology information. Sample output { "timestamp":1360704473225, "status":"OK", "total":recordCount, "data": [ { "path":path, }, ...additional structures above for each topology node... ] } Output Fields Field Description path The physical topology path to the node. rlimit The rlimit commands enable you to get and set resource usage limits for your cluster. rlimit get rlimit set rlimit get The rlimit get API returns the resource usage limit for the cluster's disk resource Syntax CLI REST maprcli rlimit get -resource disk [ -cluster ] http[s]://:/rest/rlimit/get? Parameters Parameter Description resource The type of resource to get the usage limit for. Currently only the value disk is supported. cluster name The name of the cluster whose usage limit is being queried. Examples Return the disk usage limit for the cluster my.cluster.com: CLI REST maprcli rlimit get -resource disk -cluster my.cluster.com https://r1n1.sj.us:8443/rest/rlimit/get?cluster=my.cluster.com rlimit set The rlimit set API sets the resource usage limit for the cluster's disk resource Syntax CLI REST Parameters maprcli rlimit set -resource disk [ -cluster ] -value http[s]://:/rest/rlimit/set? Parameter Description resource The type of resource to set the usage limit for. Currently only the value disk is supported. cluster name The name of the cluster whose usage limit is being set. limit The value of the limit being set. You can express the value as KB, MB, GB, or TB. Examples Set the disk usage limit for the cluster my.cluster.com to 80TB: CLI REST maprcli rlimit set -resource disk -cluster my.cluster.com -value 80TB https://r1n1.sj.us:8443/rest/rlimit/get?resource=disk&cluster=my.cluster.com&value=80TB schedule The schedule commands let you work with schedules: schedule create creates a schedule schedule list lists schedules schedule modify modifies the name or rules of a schedule by ID schedule remove removes a schedule by ID A schedule is a JSON object that specifies a single or recurring time for volume snapshot creation or mirror syncing. For a schedule to be useful, it must be associated with at least one volume. See volume create and volume modify. Schedule Fields The schedule object contains the following fields: Field Value id The ID of the schedule. name The name of the schedule. inuse Indicates whether the schedule is associated with an action. rules An array of JSON objects specifying how often the scheduled action occurs. See Rule Fields below. Rule Fields The following table shows the fields to use when creating a rules object. Field Values frequency How often to perform the action: once - Once yearly - Yearly monthly - Monthly weekly - Weekly daily - Daily hourly - Hourly semihourly - Every 30 minutes quarterhourly - Every 15 minutes fiveminutes - Every 5 minutes minute - Every minute retain How long to retain the data resulting from the action. For example, if the schedule creates a snapshot, the retain field sets the snapshot's expiration. The retain field consists of an integer and one of the following units of time: mi - minutes h - hours d - days w - weeks m - months y - years time The time of day to perform the action, in 24-hour format: HH date The date on which to perform the action: For single occurrences, specify month, day and year: MM/DD/YYYY For yearly occurrences, specify the month and day: MM/DD For monthly occurrences occurrences, specify the day: DD Daily and hourly occurrences do not require the date field. Example The following example JSON shows a schedule called "snapshot," with three rules. { "id":8, "name":"snapshot", "inuse":0, "rules":[ { "frequency":"monthly", "date":"8", "time":14, "retain":"1m" }, { "frequency":"weekly", "date":"sat", "time":14, "retain":"2w" }, { "frequency":"hourly", "retain":"1d" } ] } schedule create Creates a schedule. Permissions required: fc or a A schedule can be associated with a volume to automate mirror syncing and snapshot creation. See volume create and volume modify. Syntax CLI maprcli schedule create [ -cluster ] -schedule REST http[s]://:/rest/schedule/create? Parameters Parameter Description cluster The cluster on which to run the command. schedule A JSON object describing the schedule. See Schedule Objects for more information. Examples Scheduling a Single Occurrence CLI REST maprcli schedule create -schedule '{"name":"Schedule-1","rules":[{"frequency":"once","retain":"1w","time":13,"d https://r1n1.sj.us:8443/rest/schedule/create?schedule={"name":"Schedule-1","rules":[{"frequency":"once","retain A Schedule with Several Rules CLI REST maprcli schedule create -schedule '{"name":"Schedule-2","rules":[{"frequency":"weekly","date":"sun","time":7,"r https://r1n1.sj.us:8443/rest/schedule/create?schedule={"name":"Schedule-1","rules":[{"frequency":"weekly","date schedule list Lists the schedules on the cluster. Syntax CLI REST maprcli schedule list [ -cluster ] [ -output terse|verbose ] http[s]://:/rest/schedule/list? Parameters Parameter Description cluster The cluster on which to run the command. output Specifies whether the output should be terse or verbose. Output A list of all schedules on the cluster. See Schedule Objects for more information. Examples List schedules: CLI maprcli schedule list REST https://r1n1.sj.us:8443/rest/schedule/list schedule modify Modifies an existing schedule, specified by ID. Permissions required: fc or a To find a schedule's ID: 1. Use the schedule list command to list the schedules. 2. Select the schedule to modify 3. Pass the selected schedule's ID in the -id parameter to the schedule modify command. Syntax CLI maprcli schedule modify [ -cluster ] -id [ -name ] REST http[s]://:/rest/schedule/modify? Parameters Parameter Description cluster The cluster on which to run the command. id The ID of the schedule to modify. name The new name of the schedule. rules A JSON object describing the rules for the schedule. If specified, replaces the entire existing rules object in the schedule. For information about the fields to use in the JSON object, see Rule Fields. Examples Modify a schedule CLI maprcli schedule modify -id 0 -name Newname -rules '[{"frequency":"weekly","date":"sun","time":7,"retain":"2w"} REST https://r1n1.sj.us:8443/rest/schedule/modify?id=0&name=Newname&rules=[{"frequency":"weekly","date":"sun","time" schedule remove Removes a schedule. A schedule can only be removed if it is not associated with any volumes. See volume modify. Syntax CLI REST maprcli schedule remove [ -cluster ] -id http[s]://:/rest/schedule/remove? Parameters Parameter Description cluster The cluster on which to run the command. id The ID of the schedule to remove. Examples Remove schedule with ID 0: CLI REST maprcli schedule remove -id 0 https://r1n1.sj.us:8443/rest/schedule/remove?id=0 service list Lists all services on the specified node, along with the state and log path for each service. Syntax CLI REST Parameters maprcli service list -node http[s]://:/rest/service/list? Parameter Description node The node on which to list the services Output Information about services on the specified node. For each service, the status is reported numerically: 0 - NOT_CONFIGURED: the package for the service is not installed and/or the service is not configured ( configure.sh has not run) 2 - RUNNING: the service is installed, has been started by the warden, and is currently executing 3 - STOPPED: the service is installed and configure.sh has run, but the service is currently not executing 5 - STAND_BY: the service is installed and is in standby mode, waiting to take over in case of failure of another instance (mainly used for JobTracker warm standby) setloglevel The setloglevel commands set the log level on individual services: setloglevel cldb - Sets the log level for the CLDB. setloglevel hbmaster - Sets the log level for the HB Master. setloglevel hbregionserver - Sets the log level for the HBase RegionServer. setloglevel jobtracker - Sets the log level for the JobTracker. setloglevel fileserver - Sets the log level for the FileServer. setloglevel nfs - Sets the log level for the NFS. setloglevel tasktracker - Sets the log level for the TaskTracker. setloglevel cldb Sets the log level on the CLDB service. Permissions required: fc or a Syntax CLI REST maprcli setloglevel cldb -classname -loglevel DEBUG|ERROR|FATAL|INFO|TRACE|WARN -node -port http[s]://:/rest/setloglevel/cldb? Parameters Parameter Description classname The name of the class for which to set the log level. loglevel The log level to set: DEBUG ERROR FATAL INFO TRACE WARN node The node on which to set the log level. port The CLDB port setloglevel fileserver Sets the log level on the FileServer service. Permissions required: fc or a Syntax CLI REST maprcli setloglevel fileserver -classname -loglevel DEBUG|ERROR|FATAL|INFO|TRACE|WARN -node -port http[s]://:/rest/setloglevel/fileserver? Parameters Parameter Description classname The name of the class for which to set the log level. loglevel The log level to set: DEBUG ERROR FATAL INFO TRACE WARN node The node on which to set the log level. port The MapR-FS port setloglevel hbmaster Sets the log level on the HBase Master service. Permissions required: fc or a Syntax CLI REST Parameters maprcli setloglevel hbmaster -classname -loglevel DEBUG|ERROR|FATAL|INFO|TRACE|WARN -node -port http[s]://:/rest/setloglevel/hbmaster? Parameter Description classname The name of the class for which to set the log level. loglevel The log level to set: DEBUG ERROR FATAL INFO TRACE WARN node The node on which to set the log level. port The HBase Master webserver port setloglevel hbregionserver Sets the log level on the HBase RegionServer service. Permissions required: fc or a Syntax CLI REST maprcli setloglevel hbregionserver -classname -loglevel DEBUG|ERROR|FATAL|INFO|TRACE|WARN -node -port http[s]://:/rest/setloglevel/hbregionserver? Parameters Parameter Description classname The name of the class for which to set the log level. loglevel The log level to set: DEBUG ERROR FATAL INFO TRACE WARN node The node on which to set the log level. port The Hbase Region Server webserver port setloglevel jobtracker Sets the log level on the JobTracker service. Permissions required: fc or a Syntax CLI REST maprcli setloglevel jobtracker -classname -loglevel DEBUG|ERROR|FATAL|INFO|TRACE|WARN -node -port http[s]://:/rest/setloglevel/jobtracker? Parameters Parameter Description classname The name of the class for which to set the log level. loglevel The log level to set: DEBUG ERROR FATAL INFO TRACE WARN node The node on which to set the log level. port The JobTracker webserver port setloglevel nfs Sets the log level on the NFS service. Permissions required: fc or a Syntax CLI REST maprcli setloglevel nfs -classname -loglevel DEBUG|ERROR|FATAL|INFO|TRACE|WARN -node -port http[s]://:/rest/setloglevel/nfs? Parameters Parameter Description classname The name of the class for which to set the log level. loglevel The log level to set: DEBUG ERROR FATAL INFO TRACE WARN node The node on which to set the log level. port The NFS port setloglevel tasktracker Sets the log level on the TaskTracker service. Permissions required: fc or a Syntax CLI maprcli setloglevel tasktracker -classname -loglevel DEBUG|ERROR|FATAL|INFO|TRACE|WARN -node -port REST http[s]://:/rest/setloglevel/tasktracker? Parameters Parameter Description classname The name of the class for which to set the log level. loglevel The log level to set: DEBUG ERROR FATAL INFO TRACE WARN node The node on which to set the log level. port The TaskTracker port table The table commands perform functions related to MapR tables: table attr table cf table create table delete table listrecent table region table attr The table attr commands enable you to list and edit attributes for MapR tables. table attr list - lists the attributes of an existing MapR table table attr edit - edits the attributes of an existing MapR table Table Attributes Name Field Value autoSplit Boolean. Defines whether or not this table is automatically split. The default value is true. A value of false indicates this table will not be automatically split. bulkloadperm Boolean. Defines whether or not this table supports bulk loading. The default value is false. A value of true indicates that this table supports bulk loading. table attr edit Edits the attributes of a specified MapR table. Syntax CLI maprcli table attr edit -path -attrname -attrvalue REST http[s]://:/rest/table/attr/edit?path=&attrname=&attrvalue= Parameters Parameter Description path Path to the table. attrname The attribute to edit. attrvalue The new value for the attribute. Table Attributes Name Field Value autoSplit Boolean. Defines whether or not this table is automatically split. The default value is true. A value of false indicates this table will not be automatically split. bulkloadperm Boolean. Defines whether or not this table supports bulk loading. The default value is false. A value of true indicates that this table supports bulk loading. sizethreshpcntforpack Indicates the size of a segment expressed as a percentage of the base spill. When a segment exceeds the specified size threshold, that segment is packed. When set to 0, this attribute is ignored. The default value is 0. Examples Editing a table's attributes This example changes the value of the autoSplit attribute for the table mytable01 from true to false. CLI maprcli table attr edit -path /my.cluster.com/user/user01/mytable01 -attrname autoSplit -attrvalue false REST https://r1n1.sj.us:8443/rest/table/attr/list?path=%2Fmy.cluster.com%2Fuser%2Fuser01%2Fmytable01&attrname=autoS This example changes the value of the sizethreshpcntforpack attribute for the table mytable01 from 0 to 140. When this attribute is set to 140, a segment is packed whenever it reaches 140 percent of its original size. CLI maprcli table attr edit -path /my.cluster.com/user/user01/mytable01 -attrname sizethreshpcntforpack -attrvalue REST https://r1n1.sj.us:8443/rest/table/attr/list?path=%2Fmy.cluster.com%2Fuser%2Fuser01%2Fmytable01&attrname=sizet table attr list Lists the attributes of a specified MapR table. Syntax CLI maprcli table attr list -path REST http[s]://:/rest/table/attr/list?path=& Parameters Parameter Description path Path to the table. Output fields Name Field Value autoSplit Boolean. Defines whether or not this table is automatically split. The default value is true. A value of false indicates this table will not be automatically split. bulkloadperm Boolean. Defines whether or not this table supports bulk loading. The default value is false. A value of true indicates that this table supports bulk loading. Examples Listing a table's attributes This example lists the attribute information for the table mytable01. CLI maprcli table attr list -path /my.cluster.com/user/user01/mytable01 REST https://r1n1.sj.us:8443/rest/table/attr/list?path=%2Fmy.cluster.com%2Fuser%2Fuser01%2Fmytable01 Example Output [user@node]# maprcli table attr list -path /mapr/my.cluster.com/user/user01/mytable01 name value autoSplit true table cf The table cf commands deal with creating and managing column families for MapR tables. table cf create - creates a new column family table cf edit - edits the properties of an existing column family table cf delete - deletes a column family table cf list - lists information about column families table cf create Creates a column family for a MapR table. Syntax CLI maprcli table cf create -path -cfname [ -compression off|lzf|lz4|zlib ] [ -minversions ] [ -maxversions ] [ -ttl ] [ -inmemory true|false ] REST http[s]://:/rest/table/cf/create?path=&cfname=& Parameters Parameter Description path Path to the MapR table. cfname The name of the new column family. compression The compression setting to use for the column family. Valid options are off, lzf, lz4, and zlib. The default setting is equal to the table's compression setting. minversions Minimum number of versions to keep. The default is zero. maxversions Maximum number of versions to keep. The default is three. ttl Time to live. The default value is forever. When the age of the data in this column family exceeds the value of the ttl param eter, the data is purged from the column family. Setting the value of ttl to 0 is equivalent to the default value of forever. inmemory Boolean. Whether or not to keep this column family in memory. The default value is false. Examples Creating a new column family for a table, keeping four versions in memory CLI maprcli table cf create -path /volume1/mytable -cfname mynewcf -maxversions 4 -inmemory true REST https://r1n1.sj.us:8443/rest/table/cf/create?path=%2Fvolume1%2Fmytable&cfname=mynewcf&maxversions=4&inmemory=t table cf delete Delete a column family from a MapR table, removing all records in the column family. Deletion cannot be undone. Syntax CLI maprcli table cf delete -path -cfname REST http[s]://:/rest/table/cf/delete?path=&cfname= Parameters Parameter Description path Path to the MapR table. cfname The name of the column family to delete. Examples Deleting a column family CLI REST maprcli table cf delete -path /volume1/thetable -cfname mycf https://r1n1.sj.us:8443/rest/table/cf/edit?path=%2Fvolume1%2Fthetable&cfname=mycf table cf edit Edits a column family definition. You can alter a column family's name, minimum and maximum versions, time to live, compression, and memory residence status. Syntax CLI maprcli table cf edit -path -cfname [ -newcfname ] [ -minversions ] [ -maxversions ] [ -ttl ] [ -inmemory true|false ] [ -compression off|lzf|lz4|zlib ] REST http[s]://:/rest/table/cf/create?path=&cfname=& Parameters Parameter Description path Path to the table. cfname The name of the column family to edit. newcfname The new name of the column family. minversions Minimum number of versions to keep. The default is zero. maxversions Maximum number of versions to keep. The default is three. ttl Time to live. The default value is forever. When the age of the data in this column family exceeds the value of the ttl param eter, the data is purged from the column family. Setting the value of ttl to 0 is equivalent to the default value of forever. inmemory Boolean. Whether or not to keep this column family in memory. The default value is false. compression The compression setting to use for the column family. Valid options are off, lzf, lz4, and zlib. The default setting is equal to the table's compression setting. Examples Changing a column family's name and time to live CLI maprcli table cf edit -path /my.cluster.com/volume1/newtable -cfname mynewcf -newcfname mynewcfname -ttl 3 REST https://r1n1.sj.us:8443/rest/table/cf/edit?path=%2Fmy.cluster.com%2Fvolume1%2Fnewtable&cfname=mynewcf&newcfnam table cf list Lists a MapR table's column families. Syntax CLI maprcli table cf list -path [ -cfname ] [ -output verbose|terse ] REST http[s]://:/rest/table/cf/list?path=& Parameters Parameter Description path Path to the table. cfname The name of the column family to edit. output Valid options are verbose or terse. Verbose output lists full names for column headers. Terse output lists abbreviated column header names. The default value is verbose. Output fields Verbose Field Name Terse Field Name Field Value inmemory inmem Whether or not this column value resides in memory cfname n The column family name maxversions vmax Maximum number of versions for this column family minversions vmin Minimum number of versions for this column family compression comp Compression scheme used for this column family ttl ttl Time to live for this column family Examples Tersely listing the column families for a table This example lists all column families for the table newtable. CLI maprcli table cf list -path /my.cluster.com/volume1/newtable -output terse REST https://r1n1.sj.us:8443/rest/table/cf/list?path=%2Fmy.cluster.com%2Fvolume1%2Fnewtable&output=terse Example Output [user@node]# terse comp inmem lz4 false lz4 false lz4 false maprcli table cf list -path /mapr/default/user/user/newtable -output vmax 3 3 3 n dine nahashchid wollachee ttl 2147483647 2147483647 2147483647 vmin 0 0 0 table create Creates a new MapR table. Optionally, you can specify several permissions levels for this table and the column families in the table with Access Control Expressions (ACEs). Syntax CLI maprcli table create -path [ -bulkloadperm Bulk Load Permission settings ] [ -purgeperm Force Purge Permission settings ] [ -autosplitperm Auto-Split Permission settings ] [ -regionsizeperm Modify Region Size Permission settings ] [ -modifyfamilyperm Add/Rename Family Permission settings ] [ -deletefamilyperm Delete Family Permission settings ] [ -versionsperm CF Versions Default Permission ] [ -compressionperm CF Compression Default Permission ] [ -memoryperm CF Memory Default Permission ] [ -readperm CF Read Default Permission ] [ -writeperm CF Write Default Permission ] [ -appendperm CF Append Default Permission ] [ -encryptperm CF Encrypt Default Permission ] REST http[s]://:/rest/table/create?path= Parameters Parameter Description path Path to the new MapR table. purgeperm Contains an ACE inside quotation marks that defines access to purge permissions for this table. autosplitperm Contains an ACE inside quotation marks that defines access to autosplit permissions for this table. bulkloadperm Users with this permission can load this table with bulk loads if the table was created with bulk load support. Defaults to creating user's UID. regionalsplitperm Contains an ACE inside quotation marks that defines access to regional split permissions for this table. modifyfamiilyperm Contains an ACE inside quotation marks that defines access to modify permissions for this table. deletefamilyperm Contains an ACE inside quotation marks that defines access to delete permissions for this table. versionperm Contains an ACE inside quotation marks that defines access to minimum and maximum versions permissions for this table. compressionperm Contains an ACE inside quotation marks that defines access to compression permissions for this table. memoryperm Contains an ACE inside quotation marks that defines access to in-memory permissions for this table. readperm Contains an ACE inside quotation marks that defines access to read permissions for this table. writeperm Contains an ACE inside quotation marks that defines access to write permissions for this table. appendperm Contains an ACE inside quotation marks that defines access to append permissions for this table. encryptperm Contains an ACE inside quotation marks that defines access to encryption permissions for this table. Examples Creating a new MapR table with restricted read permissions This example creates a table named newtable with read permissions restricted to the user with UID 1010 or any user in the admin group. CLI maprcli table create -path /my.cluster.com/volume1/newtable -readperm "u:1010 | g:admin" REST https://r1n1.sj.us:8443/rest/table/create?path=%2Fmy.cluster.com%2Fvolume1%2Fnewtable&readperm="u%3A1010%20%7C table delete Deletes a MapR table. Syntax CLI maprcli table delete -path REST http[s]://:/rest/table/create?path= Parameters Parameter Description path Path to the MapR table to delete. Examples Deleting a table CLI REST maprcli table delete -path /my.cluster.com/volume1/table https://r1n1.sj.us:8443/rest/table/delete?path=%2Fmy.cluster.com%2Fvolume1%2Ftable table listrecent MapR keeps track of the 50 most recently-accessed tables by each user. When the path argument is specified, maprcli table listrecent verifies if a table exists at that path. When used without path, maprcli table listrecent lists the user's recently-accessed tables. The paths to recently-accessed tables are written to a file, maprfs:///user//.recent_tables. Referencing a table with any of the maprcli table commands or the MapR Control System will log the table path in .recent_tables. If a user doesn't have a home directory on the cluster, maprcli table listrecent throws an error. Syntax CLI maprcli table listrecent [ -path ] [ -output verbose|terse ] REST http[s]://:/rest/table/listrecent? Parameters Parameter Description path Specifies a path to verify if a table exists at that path output Valid options are verbose or terse. Verbose output lists full names for column headers. Terse output lists abbreviated column header names. The default value is verbose. Output fields Verbose Field Name Terse Field Name Field Value path p Path to the table Examples Listing recently-accessed tables CLI maprcli table listrecent REST https://r1n1.sj.us:8443/rest/table/listrecent Listing tables verbosely CLI REST maprcli table listrecent -path /my.cluster.com/volume1/ -output verbose https://r1n1.sj.us:8443/rest/table/listrecent?path=%2Fmy.cluster.com%2Fvolume1&output=verbose table region The table region list command lists the regions associated with a specified MapR table. The table region merge command merges regions in a specified MapR table. The table region split command splits a region in a specified MapR table. table region list Lists the regions that make up a specified table. Syntax CLI maprcli table region list -path [ -output verbose|terse ] [ -start offset ] [ -limit number ] REST http[s]://:/rest/table/region/list?path=& Parameters Parameter Description path Path to the table. output Valid options are verbose or terse. Verbose output lists full names for column headers. Terse output lists abbreviated column header names. The default value is verbose. start The offset from the starting region. The default value is 0. limit The number of regions to return, counting from the starting region. The default value is 2147483647. Output fields Verbose Field Name Terse Field Name Field Value numberofrows nr Number of rows in the region fid fid The region's FID. primarynode pn Host name of the primary node for this region secondarynodes sn Host names of the secondary nodes where this region is replicated numberofrowswithdelete nrd Number of rows in the region, counting deleted rows startkey sk Value of the start key for this region endkey ek Value of the end key for this region lastheartbeat lhb Time since last heartbeat from the region's primary node logicalsize ls Logical size of the region physicalsize ps Physical size of the region Examples Tersely listing the region information for a table This example lists the region information for the table newtable. CLI maprcli table region list -path /my.cluster.com/volume1/newtable -output terse REST https://r1n1.sj.us:8443/rest/table/region/list?path=%2Fmy.cluster.com%2Fvolume1%2Fnewtable&output=terse Example Output [user@node]# maprcli table region list -path /mapr/default/user/user/newtable -output terse fid sk sn nr ls ek pn nrd ps lhb 2119.32.131282 -INFINITY 0 0 INFINITY myhost01 0 0 0 table region merge Merges regions of a table together to reduce the number of regions that table occupies. Users must have the splitmergeperm permission on this table to merge table regions. Syntax CLI maprcli table region merge -fid -path REST http[s]://:/rest/table/region/merge?fid=&path= Parameters Parameter Description fid The FID for the table region to merge. The output of maprcli table region list lists the FIDs for the table. path The path to the table whose regions are being merged. table region split Splits a region in a table. Syntax CLI maprcli table region split -path -fid REST http[s]://:/rest/table/region/split?path=&fid= Parameters Parameter Description path Path to the table. fid The FID of the region to split. The output of maprcli table region list lists the FIDs for the table's regions. Examples Splitting a table region This example splits a region in the table newtable. CLI maprcli table region split -path /my.cluster.com/volume1/newtable -fid $FID REST https://r1n1.sj.us:8443/rest/table/region/split?path=%2Fmy.cluster.com%2Fvolume1%2Fnewtable&fid=$FID task The task commands enable you to manipulate information about the Hadoop jobs that are running on your cluster: task killattempt - Kills a specific task attempt. task failattempt - Ends a specific task attempt as failed. task table - Retrieves detailed information about the task attempts associated with a job running on the cluster. task failattempt The task failattempt API ends the specified task attempt as failed. Syntax CLI REST maprcli task failattempt [ -cluster cluster name ] -taskattemptid task attempt ID http[s]://:/rest/task/failattempt?[cluster=cluster_name&]taskattemptid=task_attempt_ID Parameters Parameter Description cluster Cluster name taskattemptid Task attempt ID Examples Ending a Task Attempt as Failed CLI REST maprcli task failattempt -taskattemptid attempt_201187941846_1077_300_7707 https://r1n1.sj.us:8443/rest/task/failattempt?taskattemptid=attempt_201187941846_1077_300_7707 task killattempt The task killattempt API kills the specified task attempt. Syntax CLI REST maprcli task killattempt [ -cluster cluster name ] -taskattemptid task attempt ID http[s]://:/rest/task/killattempt?[cluster=cluster_name&]taskattemptid=task_attempt_ID Parameters Parameter Description cluster Cluster name taskattemptid Task attempt ID Examples Killing a Task Attempt CLI REST maprcli task killattempt -taskattemptid attempt_201187941846_1077_300_7707 https://r1n1.sj.us:8443/rest/task/killattempt?taskattemptid=attempt_201187941846_1077_300_7707 task table Retrieves histograms and line charts for task metrics. Use the task table API to retrieve task analytics data for your job. The metrics data can be formatted for histogram display or line chart display. Syntax REST http[s]://:/rest/task/table?output=terse&filter=string&chart=chart_type&columns=list_of_columns&sca Parameters Parameter Description filter Filters results to match the value of a specified string. chart Chart type to use: line for a line chart, bar for a histogram. columns Comma-separated list of column names to return. bincount Number of histogram bins. scale Scale to use for the histogram. Specify linear for a linear scale and log for a logarithmic scale. Column Names The following table lists the terse short names for particular metrics regarding task attempts. Parameter Description tacir Combine Task Attempt Input Records tacor Combine Task Attempt Output Records tamib Map Task Attempt Input Bytes tamir Map Task Attempt Input Records tamob Map Task Attempt Output Bytes tamor Map Task Attempt Output Records tamsr Map Task Attempt Skipped Records tarig Reduce Task Attempt Input Groups tarir Reduce Task Attempt Input Records taror Reduce Task Attempt Output Records tarsb Reduce Task Attempt Shuffle Bytes tarsr Reduce Task Attempt Skipped Records tacput Task Attempt CPU Time talbr Task Attempt Local Bytes Read talbw Task Attempt Local Bytes Written tambr Task Attempt MapR-FS Bytes Read tambw Task Attempt MapR-FS Bytes Written tapmem Task Attempt Physical Memory Bytes taspr Task Attempt Spilled Records tavmem Task Attempt Virtual Memory Bytes tad Task Attempt Duration (histogram only) tagct Task Attempt Garbage Collection Time (histogram only) td Task Duration (histogram only) taid Task Attempt ID (filter only) tat Task Attempt Type (filter only) tas Task Attempt Status (filter only) tapro Task Attempt Progress (filter only) tast Task Attempt Start Time (filter only) taft Task Attempt Finish Time (filter only) tashe Task Attempt Shuffle End tase Task Attempt Sort End tah Task Attempt Host Location talog Location of logs, stderr, and stdout for this task attempt. tadi Freeform information about this task attempt used for diagnosing behaviors. tamor Map Task Attempt Output Records tarsg Reduce Task Attempt Skipped Groups (filter only) tasrb Reduce Task Attempt Shuffle Bytes tamirps Map Task Attempt Input Records per Second tarirps Reduce Task Attempt Input Records per Second tamorps Map Task Attempt Output Records per Second tarorps Reduce Task Attempt Output Records per Second tamibps Map Task Attempt Input Bytes per Second tamobps Map Output Bytes per Second tarsbps Reduce Task Attempt Shuffle Bytes per Second ts Task Status (filter only) tid Task Duration tt Task Type (filter only) tsta Primary Task Attempt ID (filter only) tst Task Start Time (filter only) tft Task End Time (filter only) th Task Host Location (filter only) thl Task Host Locality (filter only) Example Retrieve a Task Histogram: REST CURL https://r1n1.sj.us:8443/rest/task/table?chart=bar&filter=%5Btt!=JOB_SETUP%5Dand%5Btt!=JOB_CLEANUP%5Dand%5Bjid= curl -d @json https://r1n1.sj.us:8443/api/task/table In the curl example above, the json file contains a URL-encoded version of the information in the Request section below. Request GENERAL_PARAMS: { [chart: "bar"|"line"], columns: , [filter: "[{operator}]and[...]",] [output: terse,] [start: int,] [limit: int] } REQUEST_PARAMS_HISTOGRAM: { chart:bar columns:td filter: } REQUEST_PARAMS_LINE: { chart:line, columns:tapmem, filter: NOT PARSED, UNUSED IN BACKEND } REQUEST_PARAMS_GRID: { columns:tid,tt,tsta,tst,tft filter: output:terse, start:0, limit:50 } Response RESPONSE_SUCCESS_HISTOGRAM: { "status" : "OK", "total" : 15, "columns" : ["td"], "binlabels" : ["0-5s","5-10s","10-30s","30-60s","60-90s","90s-2m","2m-5m","5m-10m","10m-30m","30m-1h ","1h-2h","2h-6h","6h-12h","12h-24h",">24h"], "binranges" : [ [0,5000], [5000,10000], [10000,30000], [30000,60000], [60000,90000], [90000,120000], [120000,300000], [300000,600000], [600000,1800000], [1800000,3600000], [3600000,7200000], [7200000,21600000], [21600000,43200000], [43200000,86400000], [86400000] ], "data" : [33,919,1,133,9820,972,39,2,44,80,11,93,31,0,0] } RESPONSE_SUCCESS_GRID: { "status": "OK", "total" : 67, "columns" : ["ts","tid","tt","tsta","tst","tft","td","th","thl"], "data" : [ ["FAILED","task_201204837529_1284_9497_4858","REDUCE","attempt_201204837529_1284_9497_ 4858_3680", 1301066803229,1322663797292,21596994063,"newyork-rack00-8","remote"], ["PENDING","task_201204837529_1284_9497_4858","MAP","attempt_201204837529_1284_9497_48 58_8178", 1334918721349,1341383566992,6464845643,"newyork-rack00-7","unknown"], ["RUNNING","task_201204837529_1284_9497_4858","JOB_CLEANUP","attempt_201204837529_1284 _9497_4858_1954", 1335088225728,1335489232319,401006591,"newyork-rack00-8","local"], ]} RESPONSE_SUCCESS_LINE: { "status" : "OK", "total" : 22, "columns" : ["tapmem"], "data" : [ [1329891055016,0], [1329891060016,8], [1329891065016,16], [1329891070016,1024], [1329891075016,2310], [1329891080016,3243], [1329891085016,4345], [1329891090016,7345], [1329891095016,7657], [1329891100016,8758], [1329891105016,9466], [1329891110016,10345], [1329891115016,235030], [1329891120016,235897], [1329891125016,287290], [1329891130016,298390], [1329891135016,301355], [1329891140016,302984], [1329891145016,303985], [1329891150016,304403], [1329891155016,503030], [1329891160016,983038] ] } trace The trace commands let you view and modify the trace buffer, and the trace levels for the system modules. The valid trace levels are: DEBUG INFO ERROR WARN FATAL The following pages provide information about the trace commands: trace dump trace info trace print trace reset trace resize trace setlevel trace setmode trace dump Dumps the contents of the trace buffer into the MapR-FS log. Syntax CLI maprcli trace dump [ -host ] [ -port ] None. REST Parameters Parameter Description host The IP address of the node from which to dump the trace buffer. Default: localhost port The port to use when dumping the trace buffer. Default: 5660 Examples Dump the trace buffer to the MapR-FS log: CLI maprcli trace dump trace info Displays the trace level of each module. Syntax CLI maprcli trace info [ -host ] [ -port ] None. REST Parameters Parameter Description host The IP address of the node on which to display the trace level of each module. Default: localhost port The port to use. Default: 5660 Output A list of all modules and their trace levels. Sample Output RPC Client Initialize **Trace is in DEFAULT mode. **Allowed Trace Levels are: FATAL ERROR WARN INFO DEBUG **Trace buffer size: 2097152 **Modules and levels: Global : INFO RPC : ERROR MessageQueue : ERROR CacheMgr : INFO IOMgr : INFO Transaction : ERROR Log : INFO Cleaner : ERROR Allocator : ERROR BTreeMgr : ERROR BTree : ERROR BTreeDelete : ERROR BTreeOwnership : INFO MapServerFile : ERROR MapServerDir : INFO Container : INFO Snapshot : INFO Util : ERROR Replication : INFO PunchHole : ERROR KvStore : ERROR Truncate : ERROR Orphanage : INFO FileServer : INFO Defer : ERROR ServerCommand : INFO NFSD : INFO Cidcache : ERROR Client : ERROR Fidcache : ERROR Fidmap : ERROR Inode : ERROR JniCommon : ERROR Shmem : ERROR Table : ERROR Fctest : ERROR DONE Examples Display trace info: CLI maprcli trace info trace print Manually dumps the trace buffer to stdout. Syntax CLI maprcli trace print [ -host ] [ -port ] -size None. REST Parameters Parameter Description host The IP address of the node from which to dump the trace buffer to stdout. Default: localhost port The port to use. Default: 5660 size The number of kilobytes of the trace buffer to print. Maximum: 64 Output The most recent bytes of the trace buffer. ----------------------------------------------------2010-10-04 13:59:31,0000 Program: mfs on Host: fakehost IP: 0.0.0.0, Port: 0, PID: 0 ----------------------------------------------------DONE Examples Display the trace buffer: CLI maprcli trace print trace reset Resets the in-memory trace buffer. Syntax CLI maprcli trace reset [ -host ] [ -port ] None. REST Parameters Parameter Description host The IP address of the node on which to reset the trace parameters. Default: localhost port The port to use. Default: 5660 Examples Reset trace parameters: CLI maprcli trace reset trace resize Resizes the trace buffer. Syntax CLI maprcli trace resize [ -host ] [ -port ] -size None. REST Parameters Parameter Description host The IP address of the node on which to resize the trace buffer. Default: localhost port The port to use. Default: 5660 size The size of the trace buffer, in kilobytes. Default: 2097152 Minimum: 1 Examples Resize the trace buffer to 1000 CLI maprcli trace resize -size 1000 trace setlevel Sets the trace level on one or more modules. Syntax CLI maprcli trace setlevel [ -host ] -level -module [ -port ] None. REST Parameters Parameter Description host The node on which to set the trace level. Default: localhost module The module on which to set the trace level. If set to all, sets the trace level on all modules. level The new trace level. If set to default, sets the trace level to its default. port The port to use. Default: 5660 Examples Set the trace level of the log module to INFO: CLI maprcli trace setlevel -module log -level info Set the trace levels of all modules to their defaults: CLI maprcli trace setlevel -module all -level default trace setmode Sets the trace mode. There are two modes: Default Continuous In default mode, all trace messages are saved in a memory buffer. If there is an error, the buffer is dumped to stdout. In continuous mode, every allowed trace message is dumped to stdout in real time. Syntax CLI REST maprcli trace setmode [ -host ] -mode default|continuous [ -port ] None. Parameters Parameter Description host The IP address of the host on which to set the trace mode mode The trace mode. port The port to use. Examples Set the trace mode to continuous: CLI maprcli trace setmode -mode continuous urls The urls command displays the status page URL for the specified service. Syntax CLI REST maprcli urls [ -cluster ] -name [ -zkconnect ] http[s]://:/rest/urls/ Parameters Parameter Description cluster The name of the cluster on which to save the configuration. name The name of the service for which to get the status page: cldb jobtracker tasktracker zkconnect ZooKeeper Connect String Examples Display the URL of the status page for the CLDB service: CLI maprcli urls -name cldb https://r1n1.sj.us:8443/rest/maprcli/urls/cldb REST userconfig The userconfig load command displays information about the current user. userconfig load Loads the configuration for the specified user. Syntax CLI maprcli userconfig load -username REST http[s]://:/rest/userconfig/load? Parameters Parameter Description username The username for which to load the configuration. Output The configuration for the specified user. Sample Output username root fsadmin 1 mradmin 1 Output Fields Field Description username The username for the specified user. email The email address for the user. fsadmin Indicates whether the user is a MapR-FS Administrator: 0 = no 1 = yes mradmin Indicates whether the user is a MapReduce Administrator: 0 = no 1 = yes helpUrl URL pattern for locating help files on the server. Example: http://www.mapr.com/doc/display/MapR-/# helpVersion Version of the help content corresponding to this build of MapR. Note that this is different from the build version. Examples View the root user's configuration: CLI maprcli userconfig load -username root REST https://r1n1.sj.us:8443/rest/userconfig/load?username=root virtualip The virtualip commands let you work with virtual IP addresses for NFS nodes: virtualip add adds a range of virtual IP addresses virtualip edit edits a range of virtual IP addresses virtualip list lists virtual IP addresses virtualip move reassigns a range of virtual IP addresses to a MAC virtualip remove removes a range of virtual IP addresses Virtual IP Fields Field Description macaddress The MAC address of the virtual IP. netmask The netmask of the virtual IP. virtualipend The virtual IP range end. virtualip add Adds a virtual IP address. Permissions required: fc or a Syntax CLI maprcli virtualip add [ -cluster ] [ -gateway ] [ -macs ] -netmask -virtualip [ -virtualipend ] [ -preferredmac ] REST http[s]://:/rest/virtualip/add? Parameters Parameter Description cluster The cluster on which to run the command. gateway The NFS gateway IP or address macs A list of the MAC addresses that represent the NICs on the nodes that the VIPs in the VIP range can be associated with. Use this list to limit VIP assignment to NICs on a particular subnet when your NFS server is part of multiple subnets. netmask The netmask of the virtual IP. virtualip The virtual IP, or the start of the virtual IP range. virtualipend The end of the virtual IP range. preferredmac The preferred MAC for this virtual IP. When an NFS server restarts, the MapR system attempts to move all of the virtual IP addresses that list a MAC address on this node as a preferred MAC to this node. If the new value is null, this parameter resets the preferred MAC value. virtualip edit Edits a virtual IP (VIP) range. Permissions required: fc or a Syntax CLI REST maprcli virtualip edit [ -cluster ] [ -macs ] -netmask -virtualip [ -virtualipend ] [ -preferredmac ] http[s]://:/rest/virtualip/edit? Parameters Parameter Description cluster The cluster on which to run the command. macs A list of the MAC addresses that represent the NICs on the nodes that the VIPs in the VIP range can be associated with. Use this list to limit VIP assignment to NICs on a particular subnet when your NFS server is part of multiple subnets. netmask The netmask of the virtual IP. virtualip The virtual IP, or the start of the virtual IP range. virtualipend The end of the virtual IP range. preferredmac The preferred MAC for this virtual IP. When an NFS server restarts, the MapR system attempts to move all of the virtual IP addresses that list a MAC address on this node as a preferred MAC to this node. If the new value is null, this parameter resets the preferred MAC value. virtualip list Lists the virtual IP addresses in the cluster. Syntax CLI REST maprcli virtualip list [ -cluster ] [ -columns ] [ -filter ] [ -limit ] [ -nfsmacs ] [ -output ] [ -range ] [ -start ] http[s]://:/rest/virtualip/list? Parameters Parameter Description cluster The cluster on which to run the command. columns The columns to display. filter A filter specifying VIPs to list. See Filters for more information. limit The number of records to return. nfsmacs The MAC addresses of servers running NFS. output Whether the output should be terse or verbose. range The VIP range. start The index of the first record to return. virtualip move The virtualip move API reassigns a virtual IP or a range of virtual IP addresses to a specified Media Access Control (MAC) address. Syntax CLI REST maprcli virtualip move [ -cluster ] -virtualip [ -virtualipend -tomac http[s]://:/rest/virtualip/move? Parameters Parameter Description cluster name The name of the cluster where the virtual IP addresses are being moved. virtualip A virtual IP address. If you provide a value for -virtualipend, this virtual IP address defines the beginning of the range. virtualip end range A virtual IP address that defines the end of a virtual IP address range. mac The MAC address that the virtual IP addresses are being assigned. Examples Move a range of three virtual IP addresses to a MAC address for the cluster my.cluster.com: CLI maprcli virtualip move -cluster my.cluster.com -virtualip 192.168.0.8 -virtualipend 192.168.0.10 -tomac 00:FE:E REST https://r1n1.sj.us:8443/rest/virtualip/move?cluster=my.cluster.com&virtualip=192.168.0.8&virtualipend=192.168.0 virtualip remove Removes a virtual IP (VIP) or a VIP range. Permissions required: fc or a Syntax CLI maprcli virtualip remove [ -cluster ] -virtualip [ -virtualipend ] REST http[s]://:/rest/virtualip/remove? Parameters Parameter Description cluster The cluster on which to run the command. virtualip The virtual IP or the start of the VIP range to remove. virtualipend The end of the VIP range to remove. volume The volume commands let you work with volumes, snapshots and mirrors: volume create creates a volume volume dump create creates a volume dump volume dump restore restores a volume from a volume dump volume info displays information about a volume volume link create creates a symbolic link volume link remove removes a symbolic link volume list lists volumes in the cluster volume mirror push pushes a volume's changes to its local mirrors volume mirror start starts mirroring a volume volume mirror stop stops mirroring a volume volume modify modifies a volume volume mount mounts a volume volume move moves a volume volume remove removes a volume volume rename renames a volume volume showmounts shows the mount points for a volume volume snapshot create creates a volume snapshot volume snapshot list lists volume snapshots volume snapshot preserve prevents a volume snapshot from expiring volume snapshot remove removes a volume snapshot volume unmount unmounts a volume Fields The following table lists the data fields that provide information about each volume. Each field has two names: Field name - displayed in the output of the node list command Short name - used to specify the columns displayed using the columns parameter The short name is also used when specifying rows with a filter, for example when specifying a set of volumes about which to get information. Verbose Terse Description acl acl A JSON object that contains the Access Control List for the volume. actualreplication arf The actual current replication factor by percentage of the volume, as a zero-based array of integers from 0 to 100. For each position in the array, the value is the percentage of the volume that is replicated index number of times. Example: arf=5,10,85 means that 5% is not replicated, 10% is replicated once, 85% is replicated twice. advisoryquota aqt A value of 0 indicates there are no soft or advisory quotas for this volume aename aen Accountable entity name aetype aet Accountable entity type: creator on Name of the user that created the volume logicalUsed dlu Logical size of disk used by this volume maxinodesalarmthreshold miath The threshold of inodes in use that will set off the VOLUME_ALARM_INODES_EXCEEDED alarm minreplicas mrf Minimum number of replicas before re-replication starts. mountdir p The path the volume is mounted on mounted mt A value of 1 indicates the volume is mounted nameContainerSizeMB ncsmb needsGfsck nfsck A value of TRUE indicates this volume requires a filesystem check numreplicas drf Desired number of replicas. Containers with this amount of replicas are not re-replicated. partlyOutOfTopology poot A value of 1 indicates this volume is partly out of its topology quota qta A value of 0 indicates there are no hard quotas for this volume rackpath rp The rack path for this volume readonly ro A value of 1 indicates the volume is read-only replicationtype dcr Replication type scheduleid sid The ID of the schedule, if any, used by this volume schedulename sn The name of the schedule, if any, used by this volume snapshotcount sc The number of snapshots for this volume snapshotused ssu Disk space used for all snapshots, in MB totalused Total space used for volume and snapshots, in MB used Disk space used, in MB, not including snapshots volumeid id The volume ID volumename n The name of the volume volumetype The volume type volume container move Moves a container. Permissions required: cv, fc, or a The volume container move command moves a specified container (cid) from a source file server (fromfileserver) to a destination file server (tofileserver). If the tofileserver parameter is not specified, a destination file server is chosen by the CLDB. If the tofileserver is specified but does not exist, the command fails with an error. If the fromfileserver does not exist or is down, the container move occurs once the source file server comes back up. Syntax CLI volume container move -cid -fromfileserverid fromfileserverid [ -tofileserverid tofileserverid ] Parameters Parameter Description cid The container ID. fromfileserverid The ID of the file server on which the container to be moved currently resides. The ID is available from the maprcli node list command. tofileserverid The ID of the file server to which to move the container. If not specified, a file server is chosen by the CLDB. The ID is available from the maprcli node list command. Examples : CLI maprcli volume container move -cid 2316 -fromfileserverid 5227152973904547710 -tofileserverid 875290643748357753 volume create Creates a volume. Permissions required: cv, fc, or a Syntax CLI REST maprcli volume create -name -type 0|1 [ -advisoryquota ] [ -ae ] [ -aetype ] [ -cluster ] [ -createparent 0|1 ] [ -group ] [ -localvolumehost ] [ -localvolumeport ] [ -maxinodesalarmthreshold ] [ -minreplication ] [ -mount 0|1 ] [ -path ] [ -quota ] [ -readonly ] [ -replication ] [ -replicationtype ] [ -rereplicationtimeoutsec ] [ -rootdirperms ] [ -schedule ] [ -source ] [ -topology ] [ -user ] http[s]://:/rest/volume/create? Parameters Parameter Description advisoryquota The advisory quota for the volume as integer plus unit. Example: quota=500G; Units: B, K, M, G, T, P ae The accounting entity that owns the volume. aetype The type of accounting entity: 0=user 1=group cluster The cluster on which to create the volume. createparent Specifies whether or not to create a parent volume: 0 = Do not create a parent volume. 1 = Create a parent volume. group Space-separated list of group:permission pairs. localvolumehost The local volume host. localvolumeport The local volume port. Default: 5660 maxinodesalarmthreshold Threshold for the INODES_EXCEEDED alarm. minreplication The minimum replication level. Default: 2 When the replication factor falls below this minimum, re-replication occurs as aggressively as possible to restore the replication level. If any containers in the CLDB volume fall below the minimum replication factor, writes are disabled until aggressive re-replication restores the minimum level of replication. mount Specifies whether the volume is mounted at creation time. name The name of the volume to create. path The path at which to mount the volume. quota The quota for the volume as integer plus unit. Example: quota=500G; Units: B, K, M, G, T, P readonly Specifies whether or not the volume is read-only: 0 = Volume is read/write. 1 = Volume is read-only. replication The desired replication level. Default: 3 When the number of copies falls below the desired replication factor, but remains equal to or above the minimum replication factor, re-replication occurs after the timeout specified in the cldb.fs.mark.rereplicate.sec parameter. replicationtype The desired replication type. You can specify low_latency (star replication) or high_throughput (chain replication). The default setting is high_throughput. rereplicationtimeoutsec The re-replication timeout, in seconds. rootdirperms Permissions on the volume root directory. schedule The ID of a schedule. If a schedule ID is provided, then the volume will automatically create snapshots (normal volume) or sync with its source volume (mirror volume) on the specified schedule. Use the schedule list command to find the ID of the named schedule you wish to apply to the volume. source For mirror volumes, the source volume to mirror, in the format @ (Required when creating a mirror volume). topology The rack path to the volume. user Space-separated list of user:permission pairs. type The type of volume to create: 0 - standard volume 1 - mirror volume Examples Create the volume "test-volume" mounted at "/test/test-volume": CLI REST maprcli volume create -name test-volume -path /test/test-volume https://r1n1.sj.us:8443/rest/volume/create?name=test-volume&path=/test/test-volume Create Volume with a Quota and an Advisory Quota This example creates a volume with the following parameters: advisoryquota: 100M name: volumename path: /volumepath quota: 500M replication: 3 schedule: 2 topology: /East Coast type: 0 CLI maprcli volume create -name volumename -path /volumepath -advisoryquota 100M -quota 500M -replication 3 -schedu REST https://r1n1.sj.us:8443/rest/volume/create?advisoryquota=100M&name=volumename&path=/volumepath"a=500M&repli Create the mirror volume "test-volume.mirror" from source volume "test-volume" and mount at "/test/test-volume-mirror": CLI maprcli volume create -name test-volume.mirror -source test-volume@src-cluster-name -path /test/test-volume-mir REST https://r1n1.sj.us:8443/rest/volume/create?name=test-volume.mirror&sourcetest-volume@src-cluster-name&path=/tes volume dump create The volume dump create command creates a volume dump file containing data from a volume for distribution or restoration. Permissions required: dump, fc, or a You can use volume dump create to create two types of files: full dump files containing all data in a volume incremental dump files that contain changes to a volume between two points in time A full dump file is useful for restoring a volume from scratch. An incremental dump file contains the changes necessary to take an existing (or restored) volume from one point in time to another. Along with the dump file, a full or incremental dump operation can produce a state file (specified by the ?-e parameter) that contains a table of the version number of every container in the volume at the time the dump file was created. This represents the end point of the dump file, which is used as the start point of the next incremental dump. The main difference between creating a full dump and creating an incremental dump is whether the -s parameter is specified; if -s is not specified, the volume create command includes all volume data and creates a full dump file. If you create a full dump followed by a series of incremental dumps, the result is a sequence of dump files and their accompanying state files: dumpfile1 statefile1 dumpfile2 statefile2 dumpfile3 statefile3 ... To maintain an up-to-date dump of a volume: 1. Create a full dump file. Example: maprcli volume dump create -name cli-created -dumpfile fulldump1 -e statefile1 2. Periodically, add an incremental dump file. Examples: maprcli volume dump create -s statefile1 -e statefile2 -name cli-created -dumpfile incrdump1 maprcli volume dump create -s statefile2 -e statefile3 -name cli-created -dumpfile incrdump2 maprcli volume dump create -s statefile3 -e statefile4 -name cli-created -dumpfile incrdump3 ...and so on. You can restore the volume from scratch, using the volume dump restore command with the full dump file, followed by each dump file in sequence. Syntax CLI maprcli volume dump create [ -cluster ] [ -s ] [ -e ] [ -o ] [ -dumpfile ] -name volumename {anchor:cli-syntax-end} None. REST Parameters Parameter Description cluster The cluster on which to run the command. dumpfile The name of the dump file (ignored if -o is used). e The name of the state file to create for the end point of the dump. name A volume name. o This option dumps the volume to stdout instead of to a file. s The start point for an incremental dump. Examples Create a full dump: CLI maprcli volume create -e statefile1 -dumpfile fulldump1 -name volume -n Create an incremental dump: CLI maprcli volume dump -s statefile1 -e statefile2 -name volume -dumpfile incrdump1 volume dump restore The volume dump restore command restores or updates a volume from a dump file. Permissions required: dump, fc, or a There are two ways to use volume dump restore: With a full dump file, volume dump restore recreates a volume from scratch from volume data stored in the dump file. With an incremental dump file, volume dump restore updates a volume using incremental changes stored in the dump file. The volume that results from a volume dump restore operation is a mirror volume whose source is the volume from which the dump was created. After the operation, this volume can perform mirroring from the source volume. When you are updating a volume from an incremental dump file, you must specify an existing volume and an incremental dump file. To restore from a sequence of previous dump files would involve first restoring from the volume's full dump file, then applying each subsequent incremental dump file. A restored volume may contain mount points that represent volumes that were mounted under the original source volume from which the dump was created. In the restored volume, these mount points have no meaning and do not provide access to any volumes that were mounted under the source volume. If the source volume still exists, then the mount points in the restored volume will work if the restored volume is associated with the source volume as a mirror. To restore from a full dump plus a sequence of incremental dumps: 1. Restore from the full dump file, using the -n option to create a new mirror volume and the -name option to specify the name. Example: maprcli volume dump restore -dumpfile fulldump1 -name restore1 -n 2. Restore from each incremental dump file in order, specifying the same volume name. Examples: maprcli volume dump restore -dumpfile incrdump1 -name restore1 maprcli volume dump restore -dumpfile incrdump2 -name restore1 maprcli volume dump restore -dumpfile incrdump3 -name restore1 ...and so on. Syntax CLI maprcli volume dump restore [ -cluster ] [ -dumpfile dumpfilename ] [ -i ] [ -n ] -name None. REST Parameters Parameter Description cluster The cluster on which to run the command. dumpfile The name of the dumpfile (ignored if -i is used). i This option reads the dump file from stdin. n This option creates a new volume if it doesn't exist. name A volume name, in the form volumename Examples Restore a volume from a full dump file: CLI maprcli volume dump restore -name volume -dumpfile fulldump1 Apply an incremental dump file to a volume: CLI maprcli volume dump restore -name volume -dumpfile incrdump1 volume fixmountpath Corrects the mount path of a volume. Permissions required: fc or a The CLDB maintains information about the mount path of every volume. If a directory in a volume's path is renamed (by a hadoop fs command, for example) the information in the CLDB will be out of date. The volume fixmountpath command does a reverse path walk from the volume and corrects the mount path information in the CLDB. Syntax CLI REST maprcli volume fixmountpath -name [ -cluster ] http[s]://:/rest/volume/fixmountpath? Parameters Parameter Description name The volume name. clustername The cluster name Examples Fix the mount path of volume v1: CLI REST maprcli volume fixmountpath -name v1 https://r1n1.sj.us:8443/rest/volume/fixmountpath?name=v1 volume info Displays information about the specified volume. Syntax CLI maprcli volume info [ -cluster ] [ -name ] [ -output terse|verbose ] [ -path ] REST http[s]://:/rest/volume/info? Parameters You must specify either name or path. Parameter Description cluster The cluster on which to run the command. name The volume for which to retrieve information. output Whether the output should be terse or verbose. path The mount path of the volume for which to retrieve information. Fields For definitions of the output fields, and short names for use with filters, see the Fields table on the volume command page. volume link create Creates a link to a volume. Permissions required: fc or a Syntax CLI REST maprcli volume link create [ -cluster ] -path -type -volume http[s]://:/rest/volume/link/remove? Parameters Parameter Description path The path parameter specifies the link path and other information, using the following syntax: /link/[maprfs::][volume::]:: link - the link path maprfs - a keyword to indicate a special MapR-FS link volume - a keyword to indicate a link to a volume volume type - writeable or mirror volume name - the name of the volume Example: /abc/maprfs::mirror::abc type The volume type: writeable or mirror. volume The volume name. clustername The cluster name. Examples Create a link to v1 at the path v1. mirror: CLI REST maprcli volume link create -volume v1 -type mirror -path /v1.mirror https://r1n1.sj.us:8443/rest/volume/link/create?path=/v1.mirror&type=mirror&volume=v1 volume link remove Removes the specified symbolic link. Permissions required: fc or a Syntax CLI maprcli volume link remove -path [ -cluster ] REST http[s]://:/rest/volume/link/remove? Parameters Parameter Description path The symbolic link to remove. The path parameter specifies the link path and other information about the symbolic link, using the following syntax: /link/[maprfs::][volume::]:: link - the symbolic link path *maprfs - a keyword to indicate a special MapR-FS link volume - a keyword to indicate a link to a volume volume type - writeable or mirror volume name - the name of the volume Example: /abc/maprfs::mirror::abc clustername The cluster name. Examples Remove the link /abc: CLI REST maprcli volume link remove -path /abc/maprfs::mirror::abc https://r1n1.sj.us:8443/rest/volume/link/remove?path=/abc/maprfs::mirror::abc volume list Lists information about volumes specified by name, path, or filter. See the Fields table on the volume page for the fields available to filter. See the API Reference page for information on filters. Syntax CLI maprcli volume list [ -alarmedvolumes 1 ] [ -cluster ] [ -columns ] [ -filter ] [ -limit ] [ -nodes ] [ -output terse | verbose ] [ -start ] REST http[s]://:/rest/volume/list? Parameters Parameter Description alarmedvolumes Specifies whether to list alarmed volumes only. cluster The cluster on which to run the command. columns A comma-separated list of fields to return in the query. See the Fields table on the volume page. filter A filter specifying volumes to list. See API Reference for more information. limit The number of rows to return, beginning at start. Default: 0 nodes A list of nodes. If specified, volume list only lists volumes on the specified nodes. output Specifies whether the output should be terse or verbose. start The offset from the starting row according to sort. Default: 0 Fields For definitions of the output fields, and short names for use with filters, see the Fields table on the volume command page. Output Information about the specified volumes. mirrorstatus QuotaExceededAlarm numreplicas schedulename DataUnavailableAlarm volumeid rackpath volumename used volumetype SnapshotFailureAlarm mirrorDataSrcVolumeId advisoryquota aetype creator snapshotcount quota mountdir scheduleid snapshotused MirrorFailureAlarm AdvisoryQuotaExceededAlarm minreplicas mirrorDataSrcCluster actualreplication aename mirrorSrcVolumeId mirrorId mirrorSrcCluster lastSuccessfulMirrorTime nextMirrorId mirrorDataSrcVolume mirrorSrcVolume mounted logicalUsed readonly totalused DataUnderReplicatedAlarm mirror-percent-complete 0 0 3 every15min 0 362 / ATS-Run-2011-01-31-160018 864299 0 0 0 0 0 root 3 0 /ATS-Run-2011-01-31-160018 4 1816201 0 0 1 ... root 0 0 0 0 1 2110620 0 2680500 0 0 0 0 3 0 12 / mapr.cluster.internal 0 0 0 0 0 0 root 0 0 /var/mapr/cluster 0 0 0 0 1 ... root 0 0 0 0 1 0 0 0 0 0 0 0 3 0 11 / mapr.cluster.root 1 0 0 0 0 0 root 0 0 / 0 0 0 0 1 ... root 0 0 0 0 1 1 0 1 0 0 0 0 10 0 21 / mapr.jobtracker.volume 1 0 0 0 0 0 root 0 0 /var/mapr/cluster/mapred/jobTracker 0 0 0 0 1 ... root 0 0 0 0 1 1 0 1 0 0 0 0 3 0 1 / mapr.kvstore.table 0 0 0 0 0 0 root 0 0 0 0 0 0 1 ... root 0 0 0 0 0 0 0 0 0 0 Examples List all volumes CLI maprcli volume list REST https://r1n1.sj.us:8443/rest/volume/list List the first ten volumes The start and limit parameters are useful for windowing the results. You can list the first ten volumes, then the next ten, and so on. CLI maprcli volume list -start 0 -limit 10 REST https://r1n1.sj.us:8443/rest/node/list?start=0&limit=10 volume mirror push Pushes the changes in a volume to all of its mirror volumes in the same cluster, and waits for each mirroring operation to complete. Use this command when you need to push recent changes. Syntax CLI maprcli volume mirror push [ -cluster ] -name [ -verbose true|false ] REST None. Parameters Parameter Description cluster The cluster on which to run the command. name The volume to push. verbose Specifies whether the command output should be verbose. Default: true Output Sample Output Starting mirroring of volume mirror1 Mirroring complete for volume mirror1 Successfully completed mirror push to all local mirrors of volume volume1 Examples Push changes from the volume "volume1" to its local mirror volumes: maprcli volume mirror push -name volume1 -cluster mycluster CLI volume mirror start Starts mirroring on the specified volume from its source volume. License required: M5 Permissions required: fc or a When a mirror is started, the mirror volume is synchronized from a hidden internal snapshot so that the mirroring process is not affected by any concurrent changes to the source volume. The volume mirror start command does not wait for mirror completion, but returns immediately. The changes to the mirror volume occur atomically at the end of the mirroring process; deltas transmitted from the source volume do not appear until mirroring is complete. To provide rollback capability for the mirror volume, the mirroring process creates a snapshot of the mirror volume before starting the mirror, with the following naming format: .mirrorsnap..