JustPaste.it

Fri Dec  2 19:13:57 IST 2016 Starting regionserver on hscale-dev1-dn1
core file size          (blocks, -c) 0
data seg size           (kbytes, -d) unlimited
scheduling priority             (-e) 0
file size               (blocks, -f) unlimited
pending signals                 (-i) 62057
max locked memory       (kbytes, -l) 64
max memory size         (kbytes, -m) unlimited
open files                      (-n) 32000
pipe size            (512 bytes, -p) 8
POSIX message queues     (bytes, -q) 819200
real-time priority              (-r) 0
stack size              (kbytes, -s) 10240
cpu time               (seconds, -t) unlimited
max user processes              (-u) 16000
virtual memory          (kbytes, -v) unlimited
file locks                      (-x) unlimited
2016-12-02 19:13:58,477 INFO  [main] util.VersionInfo: HBase 1.1.2.2.4.2.0-258
2016-12-02 19:13:58,477 INFO  [main] util.VersionInfo: Source code repository file:///grid/0/jenkins/workspace/HDP-build-centos6/bigtop/build/hbase/rpm/BUILD/hbase-1.1.2.2.4.2.0 revision=Unknown
2016-12-02 19:13:58,477 INFO  [main] util.VersionInfo: Compiled by jenkins on Mon Apr 25 06:36:21 UTC 2016
2016-12-02 19:13:58,477 INFO  [main] util.VersionInfo: From source with checksum 4f661ee4f9f148ce7bfcad5b0d667c27
2016-12-02 19:13:58,755 INFO  [main] util.ServerCommandLine: env:PATH=/usr/sbin:/sbin:/usr/lib/ambari-server/*:/usr/lib64/qt-3.3/bin:/usr/local/sbin:/usr/local/bin:/sbin:/bin:/usr/sbin:/usr/bin:/usr/jdk64/jdk1.8.0_60/bin:/usr/jdk64/jdk1.8.0_60/jre/bin:/root/bin:/var/lib/ambari-agent
2016-12-02 19:13:58,755 INFO  [main] util.ServerCommandLine: env:HISTCONTROL=ignoredups
2016-12-02 19:13:58,755 INFO  [main] util.ServerCommandLine: env:HBASE_PID_DIR=/var/run/hbase
2016-12-02 19:13:58,755 INFO  [main] util.ServerCommandLine: env:HBASE_REGIONSERVER_OPTS= -Xmn512m -Xms3072m -Xmx3072m  -XX:+HeapDumpOnOutOfMemoryError -XX:MaxDirectMemorySize=2g -XX:+AlwaysPreTouch -XX:+UseG1GC -XX:MaxNewSize=4g -XX:InitiatingHeapOccupancyPercent=60 -XX:ParallelGCThreads=24 -XX:+ParallelRefProcEnabled -XX:MaxGCPauseMillis=5000 -Dcom.sun.management.jmxremote.authenticate=false -Dcom.sun.management.jmxremote.ssl=false -Dcom.sun.management.jmxremote.port=10102
2016-12-02 19:13:58,755 INFO  [main] util.ServerCommandLine: env:HBASE_CONF_DIR=/usr/hdp/current/hbase-regionserver/conf
2016-12-02 19:13:58,755 INFO  [main] util.ServerCommandLine: env:JRE_HOME=/usr/jdk64/jdk1.8.0_60/jre
2016-12-02 19:13:58,755 INFO  [main] util.ServerCommandLine: env:MAIL=/var/spool/mail/hbase
2016-12-02 19:13:58,756 INFO  [main] util.ServerCommandLine: env:LD_LIBRARY_PATH=::/usr/hdp/2.4.2.0-258/hadoop/lib/native/Linux-amd64-64:/usr/hdp/2.4.2.0-258/hadoop/lib/native
2016-12-02 19:13:58,756 INFO  [main] util.ServerCommandLine: env:LOGNAME=hbase
2016-12-02 19:13:58,756 INFO  [main] util.ServerCommandLine: env:HBASE_REST_OPTS=
2016-12-02 19:13:58,756 INFO  [main] util.ServerCommandLine: env:PWD=/home/hbase
2016-12-02 19:13:58,756 INFO  [main] util.ServerCommandLine: env:HBASE_ROOT_LOGGER=INFO,RFA
2016-12-02 19:13:58,756 INFO  [main] util.ServerCommandLine: env:LESSOPEN=||/usr/bin/lesspipe.sh %s
2016-12-02 19:13:58,756 INFO  [main] util.ServerCommandLine: env:SHELL=/bin/bash
2016-12-02 19:13:58,756 INFO  [main] util.ServerCommandLine: env:HBASE_ENV_INIT=true
2016-12-02 19:13:58,756 INFO  [main] util.ServerCommandLine: env:QTINC=/usr/lib64/qt-3.3/include
2016-12-02 19:13:58,756 INFO  [main] util.ServerCommandLine: env:HBASE_MASTER_OPTS= -Xms4096m -Xmx4096m  -XX:+HeapDumpOnOutOfMemoryError -XX:MaxDirectMemorySize=2g -XX:+AlwaysPreTouch -XX:+UseParNewGC -XX:+UseConcMarkSweepGC -XX:CMSInitiatingOccupancyFraction=70 -XX:+CMSParallelRemarkEnabled -Dcom.sun.management.jmxremote.authenticate=false -Dcom.sun.management.jmxremote.ssl=false -Dcom.sun.management.jmxremote.port=10101 -Dsplice.spark.enabled=true -Dsplice.spark.app.name=SpliceMachine -Dsplice.spark.master=yarn-client -Dsplice.spark.logConf=true -Dsplice.spark.broadcast.factory=org.apache.spark.broadcast.HttpBroadcastFactory -Dsplice.spark.driver.maxResultSize=1g -Dsplice.spark.driver.memory=1g -Dsplice.spark.dynamicAllocation.enabled=true -Dsplice.spark.dynamicAllocation.executorIdleTimeout=600 -Dsplice.spark.dynamicAllocation.minExecutors=0 -Dsplice.spark.io.compression.lz4.blockSize=32k -Dsplice.spark.kryo.referenceTracking=false -Dsplice.spark.kryo.registrator=com.splicemachine.derby.impl.SpliceSparkKryoRegistrator -Dsplice.spark.kryoserializer.buffer.max=512m -Dsplice.spark.kryoserializer.buffer=4m -Dsplice.spark.locality.wait=100 -Dsplice.spark.scheduler.mode=FAIR -Dsplice.spark.serializer=org.apache.spark.serializer.KryoSerializer -Dsplice.spark.shuffle.compress=false -Dsplice.spark.shuffle.file.buffer=128k -Dsplice.spark.shuffle.memoryFraction=0.7 -Dsplice.spark.shuffle.service.enabled=true -Dsplice.spark.storage.memoryFraction=0.1 -Dsplice.spark.yarn.am.extraLibraryPath=/usr/hdp/current/hadoop-client/lib/native -Dsplice.spark.yarn.am.waitTime=10s -Dsplice.spark.yarn.executor.memoryOverhead=2048 -Dsplice.spark.driver.extraJavaOptions=-Dlog4j.configuration=file:/etc/spark/conf/log4j.properties -Dsplice.spark.driver.extraLibraryPath=/usr/hdp/current/hadoop-client/lib/native -Dsplice.spark.driver.extraClassPath=/usr/hdp/current/hbase-regionserver/conf:/usr/hdp/current/hbase-regionserver/lib/htrace-core-3.1.0-incubating.jar -Dsplice.spark.executor.extraJavaOptions=-Dlog4j.configuration=file:/etc/spark/conf/log4j.properties -Dsplice.spark.executor.extraLibraryPath=/usr/hdp/current/hadoop-client/lib/native -Dsplice.spark.executor.extraClassPath=/usr/hdp/current/hbase-regionserver/conf:/usr/hdp/current/hbase-regionserver/lib/htrace-core-3.1.0-incubating.jar -Dsplice.spark.ui.retainedJobs=100 -Dsplice.spark.ui.retainedStages=100 -Dsplice.spark.worker.ui.retainedExecutors=100 -Dsplice.spark.worker.ui.retainedDrivers=100 -Dsplice.spark.streaming.ui.retainedBatches=100 -Dsplice.spark.executor.cores=4 -Dsplice.spark.executor.memory=8g -Dspark.compaction.reserved.slots=4 -Dsplice.spark.eventLog.enabled=true -Dsplice.spark.eventLog.dir=hdfs:///user/splice/history -Dsplice.spark.local.dir=/diska/tmp,/diskb/tmp,/diskc/tmp,/diskd/tmp
2016-12-02 19:13:58,756 INFO  [main] util.ServerCommandLine: env:HBASE_MANAGES_ZK=false
2016-12-02 19:13:58,756 INFO  [main] util.ServerCommandLine: env:HBASE_REGIONSERVERS=/usr/hdp/current/hbase-regionserver/conf/regionservers
2016-12-02 19:13:58,756 INFO  [main] util.ServerCommandLine: env:HADOOP_HOME=/usr/hdp/2.4.2.0-258/hadoop
2016-12-02 19:13:58,756 INFO  [main] util.ServerCommandLine: env:HBASE_NICENESS=0
2016-12-02 19:13:58,756 INFO  [main] util.ServerCommandLine: env:HBASE_OPTS=-Dhdp.version=2.4.2.0-258  -XX:ErrorFile=/var/log/hbase/hs_err_pid%p.log -Djava.io.tmpdir=/tmp -verbose:gc -XX:+PrintGCDetails -XX:+PrintGCDateStamps -Xloggc:/var/log/hbase/gc.log-201612021913  -Xmn512m -Xms3072m -Xmx3072m  -XX:+HeapDumpOnOutOfMemoryError -XX:MaxDirectMemorySize=2g -XX:+AlwaysPreTouch -XX:+UseG1GC -XX:MaxNewSize=4g -XX:InitiatingHeapOccupancyPercent=60 -XX:ParallelGCThreads=24 -XX:+ParallelRefProcEnabled -XX:MaxGCPauseMillis=5000 -Dcom.sun.management.jmxremote.authenticate=false -Dcom.sun.management.jmxremote.ssl=false -Dcom.sun.management.jmxremote.port=10102  -Dhbase.log.dir=/var/log/hbase -Dhbase.log.file=hbase-hbase-regionserver-hscale-dev1-dn1.log -Dhbase.home.dir=/usr/hdp/current/hbase-regionserver/bin/.. -Dhbase.id.str=hbase -Dhbase.root.logger=INFO,RFA -Djava.library.path=:/usr/hdp/2.4.2.0-258/hadoop/lib/native/Linux-amd64-64:/usr/hdp/2.4.2.0-258/hadoop/lib/native -Dhbase.security.logger=INFO,RFAS
2016-12-02 19:13:58,756 INFO  [main] util.ServerCommandLine: env:HBASE_START_FILE=/var/run/hbase/hbase-hbase-regionserver.autorestart
2016-12-02 19:13:58,756 INFO  [main] util.ServerCommandLine: env:HBASE_SECURITY_LOGGER=INFO,RFAS
2016-12-02 19:13:58,756 INFO  [main] util.ServerCommandLine: env:SHLVL=3
2016-12-02 19:13:58,756 INFO  [main] util.ServerCommandLine: env:ZOOKEEPER_HOME=/usr/hdp/2.4.2.0-258/zookeeper
2016-12-02 19:13:58,756 INFO  [main] util.ServerCommandLine: env:HBASE_LOGFILE=hbase-hbase-regionserver-hscale-dev1-dn1.log
2016-12-02 19:13:58,756 INFO  [main] util.ServerCommandLine: env:HISTSIZE=1000
2016-12-02 19:13:58,756 INFO  [main] util.ServerCommandLine: env:JAVA_HOME=/usr/jdk64/jdk1.8.0_60
2016-12-02 19:13:58,756 INFO  [main] util.ServerCommandLine: env:HDP_VERSION=2.4.2.0-258
2016-12-02 19:13:58,756 INFO  [main] util.ServerCommandLine: env:XFILESEARCHPATH=/usr/dt/app-defaults/%L/Dt
2016-12-02 19:13:58,756 INFO  [main] util.ServerCommandLine: env:LANG=en_US.UTF-8
2016-12-02 19:13:58,756 INFO  [main] util.ServerCommandLine: env:G_BROKEN_FILENAMES=1
2016-12-02 19:13:58,756 INFO  [main] util.ServerCommandLine: env:HBASE_CLASSPATH=/usr/hdp/2.4.2.0-258/hadoop/conf:/usr/hdp/2.4.2.0-258/hadoop/*:/usr/hdp/2.4.2.0-258/hadoop/lib/*:/usr/hdp/2.4.2.0-258/zookeeper/*:/usr/hdp/2.4.2.0-258/zookeeper/lib/*::/opt/splice/default/lib/concurrentlinkedhashmap-lru-1.4.2.jar:/opt/splice/default/lib/db-client-2.0.1.28.jar:/opt/splice/default/lib/db-drda-2.0.1.28.jar:/opt/splice/default/lib/db-engine-2.0.1.28.jar:/opt/splice/default/lib/db-shared-2.0.1.28.jar:/opt/splice/default/lib/db-tools-i18n-2.0.1.28.jar:/opt/splice/default/lib/db-tools-ij-2.0.1.28.jar:/opt/splice/default/lib/disruptor-3.2.1.jar:/opt/splice/default/lib/gson-2.2.2.jar:/opt/splice/default/lib/hbase_pipeline-hdp2.4.2-2.0.1.28.jar:/opt/splice/default/lib/hbase_sql-hdp2.4.2-2.0.1.28.jar:/opt/splice/default/lib/hbase_storage-hdp2.4.2-2.0.1.28.jar:/opt/splice/default/lib/hppc-0.5.2.jar:/opt/splice/default/lib/kryo-2.21.jar:/opt/splice/default/lib/kryo-serializers-0.26.jar:/opt/splice/default/lib/lucene-core-4.3.1.jar:/opt/splice/default/lib/opencsv-2.3.jar:/opt/splice/default/lib/pipeline_api-2.0.1.28.jar:/opt/splice/default/lib/protobuf-java-2.5.0.jar:/opt/splice/default/lib/spark-assembly-hadoop2.7.1.2.4.2.0-258-1.6.2.jar:/opt/splice/default/lib/splice_access_api-2.0.1.28.jar:/opt/splice/default/lib/splice_auth-2.0.1.28.jar:/opt/splice/default/lib/splice_backup-hdp2.4.2-2.0.1.28.jar:/opt/splice/default/lib/splice_colperms-2.0.1.28.jar:/opt/splice/default/lib/splice_ee-hdp2.4.2-2.0.1.28.jar:/opt/splice/default/lib/splice_encoding-2.0.1.28.jar:/opt/splice/default/lib/splice_encryption-2.0.1.28.jar:/opt/splice/default/lib/splice_machine-2.0.1.28.jar:/opt/splice/default/lib/splice_protocol-2.0.1.28.jar:/opt/splice/default/lib/splice_si_api-2.0.1.28.jar:/opt/splice/default/lib/splice_timestamp_api-2.0.1.28.jar:/opt/splice/default/lib/stats-2.0.1.28.jar:/opt/splice/default/lib/super-csv-2.4.0.jar:/opt/splice/default/lib/utilities-2.0.1.28.jar
2016-12-02 19:13:58,757 INFO  [main] util.ServerCommandLine: env:CVS_RSH=ssh
2016-12-02 19:13:58,757 INFO  [main] util.ServerCommandLine: env:HBASE_IDENT_STRING=hbase
2016-12-02 19:13:58,757 INFO  [main] util.ServerCommandLine: env:HBASE_ZNODE_FILE=/var/run/hbase/hbase-hbase-regionserver.znode
2016-12-02 19:13:58,757 INFO  [main] util.ServerCommandLine: env:HBASE_LOG_PREFIX=hbase-hbase-regionserver-hscale-dev1-dn1
2016-12-02 19:13:58,757 INFO  [main] util.ServerCommandLine: env:HBASE_LOG_DIR=/var/log/hbase
2016-12-02 19:13:58,757 INFO  [main] util.ServerCommandLine: env:USER=hbase
2016-12-02 19:13:58,757 INFO  [main] util.ServerCommandLine: env:CLASSPATH=/usr/hdp/current/hbase-regionserver/conf:/usr/jdk64/jdk1.8.0_60/lib/tools.jar:/usr/hdp/current/hbase-regionserver/bin/..:/usr/hdp/current/hbase-regionserver/bin/../lib/activation-1.1.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/aopalliance-1.0.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/apacheds-i18n-2.0.0-M15.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/apacheds-kerberos-codec-2.0.0-M15.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/api-asn1-api-1.0.0-M20.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/api-util-1.0.0-M20.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/asm-3.1.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/avro-1.7.4.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/commons-beanutils-1.7.0.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/commons-beanutils-core-1.8.0.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/commons-cli-1.2.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/commons-codec-1.9.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/commons-collections-3.2.2.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/commons-compress-1.4.1.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/commons-configuration-1.6.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/commons-daemon-1.0.13.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/commons-digester-1.8.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/commons-el-1.0.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/commons-httpclient-3.1.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/commons-io-2.4.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/commons-lang-2.6.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/commons-logging-1.2.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/commons-math-2.2.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/commons-math3-3.1.1.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/commons-net-3.1.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/curator-client-2.7.1.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/curator-framework-2.7.1.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/curator-recipes-2.7.1.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/disruptor-3.3.0.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/findbugs-annotations-1.3.9-1.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/gson-2.2.4.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/guava-12.0.1.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/guice-3.0.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/guice-servlet-3.0.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/hbase-annotations-1.1.2.2.4.2.0-258.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/hbase-annotations-1.1.2.2.4.2.0-258-tests.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/hbase-annotations.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/hbase-client-1.1.2.2.4.2.0-258.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/hbase-client.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/hbase-common-1.1.2.2.4.2.0-258.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/hbase-common-1.1.2.2.4.2.0-258-tests.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/hbase-common.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/hbase-examples-1.1.2.2.4.2.0-258.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/hbase-examples.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/hbase-hadoop2-compat-1.1.2.2.4.2.0-258.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/hbase-hadoop2-compat.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/hbase-hadoop-compat-1.1.2.2.4.2.0-258.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/hbase-hadoop-compat.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/hbase-it-1.1.2.2.4.2.0-258.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/hbase-it-1.1.2.2.4.2.0-258-tests.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/hbase-it.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/hbase-prefix-tree-1.1.2.2.4.2.0-258.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/hbase-prefix-tree.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/hbase-procedure-1.1.2.2.4.2.0-258.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/hbase-procedure.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/hbase-protocol-1.1.2.2.4.2.0-258.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/hbase-protocol.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/hbase-resource-bundle-1.1.2.2.4.2.0-258.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/hbase-resource-bundle.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/hbase-rest-1.1.2.2.4.2.0-258.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/hbase-rest.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/hbase-server-1.1.2.2.4.2.0-258.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/hbase-server-1.1.2.2.4.2.0-258-tests.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/hbase-server.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/hbase-shell-1.1.2.2.4.2.0-258.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/hbase-shell.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/hbase-thrift-1.1.2.2.4.2.0-258.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/hbase-thrift.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/htrace-core-3.1.0-incubating.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/httpclient-4.2.5.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/httpcore-4.2.5.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/jackson-core-2.2.3.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/jackson-core-asl-1.9.13.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/jackson-jaxrs-1.9.13.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/jackson-mapper-asl-1.9.13.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/jackson-xc-1.9.13.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/jamon-runtime-2.3.1.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/jasper-compiler-5.5.23.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/jasper-runtime-5.5.23.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/javax.inject-1.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/java-xmlbuilder-0.4.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/jaxb-api-2.2.2.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/jaxb-impl-2.2.3-1.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/jcodings-1.0.8.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/jersey-client-1.9.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/jersey-core-1.9.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/jersey-guice-1.9.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/jersey-json-1.9.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/jersey-server-1.9.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/jets3t-0.9.0.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/jettison-1.3.3.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/jetty-6.1.26.hwx.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/jetty-sslengine-6.1.26.hwx.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/jetty-util-6.1.26.hwx.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/joni-2.1.2.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/jruby-complete-1.6.8.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/jsch-0.1.42.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/jsp-2.1-6.1.14.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/jsp-api-2.1-6.1.14.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/jsr305-1.3.9.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/junit-4.11.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/leveldbjni-all-1.8.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/libthrift-0.9.0.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/log4j-1.2.17.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/metrics-core-2.2.0.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/microsoft-windowsazure-storage-sdk-0.6.0.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/netty-3.2.4.Final.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/netty-all-4.0.23.Final.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/ojdbc6.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/okhttp-2.4.0.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/okio-1.4.0.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/paranamer-2.3.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/phoenix-4.8.0-HBase-1.1-server.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/phoenix-server.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/protobuf-java-2.5.0.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/ranger-hbase-plugin-shim-0.5.0.2.4.2.0-258.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/ranger-plugin-classloader-0.5.0.2.4.2.0-258.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/servlet-api-2.5-6.1.14.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/servlet-api-2.5.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/slf4j-api-1.7.7.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/snappy-java-1.0.4.1.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/spymemcached-2.11.6.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/xercesImpl-2.9.1.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/xml-apis-1.3.04.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/xmlenc-0.52.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/xz-1.0.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/zookeeper.jar:/usr/hdp/2.4.2.0-258/hadoop/conf:/usr/hdp/2.4.2.0-258/hadoop/lib/*:/usr/hdp/2.4.2.0-258/hadoop/.//*:/usr/hdp/2.4.2.0-258/hadoop-hdfs/./:/usr/hdp/2.4.2.0-258/hadoop-hdfs/lib/*:/usr/hdp/2.4.2.0-258/hadoop-hdfs/.//*:/usr/hdp/2.4.2.0-258/hadoop-yarn/lib/*:/usr/hdp/2.4.2.0-258/hadoop-yarn/.//*:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/lib/*:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//*::mysql-connector-java-5.1.17.jar:mysql-connector-java-5.1.17.jar_bkp:mysql-connector-java.jar:/usr/hdp/2.4.2.0-258/tez/*:/usr/hdp/2.4.2.0-258/tez/lib/*:/usr/hdp/2.4.2.0-258/tez/conf:/usr/hdp/2.4.2.0-258/hadoop/conf:/usr/hdp/2.4.2.0-258/hadoop/*:/usr/hdp/2.4.2.0-258/hadoop/lib/*:/usr/hdp/2.4.2.0-258/zookeeper/*:/usr/hdp/2.4.2.0-258/zookeeper/lib/*::/opt/splice/default/lib/concurrentlinkedhashmap-lru-1.4.2.jar:/opt/splice/default/lib/db-client-2.0.1.28.jar:/opt/splice/default/lib/db-drda-2.0.1.28.jar:/opt/splice/default/lib/db-engine-2.0.1.28.jar:/opt/splice/default/lib/db-shared-2.0.1.28.jar:/opt/splice/default/lib/db-tools-i18n-2.0.1.28.jar:/opt/splice/default/lib/db-tools-ij-2.0.1.28.jar:/opt/splice/default/lib/disruptor-3.2.1.jar:/opt/splice/default/lib/gson-2.2.2.jar:/opt/splice/default/lib/hbase_pipeline-hdp2.4.2-2.0.1.28.jar:/opt/splice/default/lib/hbase_sql-hdp2.4.2-2.0.1.28.jar:/opt/splice/default/lib/hbase_storage-hdp2.4.2-2.0.1.28.jar:/opt/splice/default/lib/hppc-0.5.2.jar:/opt/splice/default/lib/kryo-2.21.jar:/opt/splice/default/lib/kryo-serializers-0.26.jar:/opt/splice/default/lib/lucene-core-4.3.1.jar:/opt/splice/default/lib/opencsv-2.3.jar:/opt/splice/default/lib/pipeline_api-2.0.1.28.jar:/opt/splice/default/lib/protobuf-java-2.5.0.jar:/opt/splice/default/lib/spark-assembly-hadoop2.7.1.2.4.2.0-258-1.6.2.jar:/opt/splice/default/lib/splice_access_api-2.0.1.28.jar:/opt/splice/default/lib/splice_auth-2.0.1.28.jar:/opt/splice/default/lib/splice_backup-hdp2.4.2-2.0.1.28.jar:/opt/splice/default/lib/splice_colperms-2.0.1.28.jar:/opt/splice/default/lib/splice_ee-hdp2.4.2-2.0.1.28.jar:/opt/splice/default/lib/splice_encoding-2.0.1.28.jar:/opt/splice/default/lib/splice_encryption-2.0.1.28.jar:/opt/splice/default/lib/splice_machine-2.0.1.28.jar:/opt/splice/default/lib/splice_protocol-2.0.1.28.jar:/opt/splice/default/lib/splice_si_api-2.0.1.28.jar:/opt/splice/default/lib/splice_timestamp_api-2.0.1.28.jar:/opt/splice/default/lib/stats-2.0.1.28.jar:/opt/splice/default/lib/super-csv-2.4.0.jar:/opt/splice/default/lib/utilities-2.0.1.28.jar
2016-12-02 19:13:58,757 INFO  [main] util.ServerCommandLine: env:SERVER_GC_OPTS=-verbose:gc -XX:+PrintGCDetails -XX:+PrintGCDateStamps -Xloggc:/var/log/hbase/gc.log-201612021913
2016-12-02 19:13:58,757 INFO  [main] util.ServerCommandLine: env:HADOOP_CONF=/usr/hdp/2.4.2.0-258/hadoop/conf
2016-12-02 19:13:58,757 INFO  [main] util.ServerCommandLine: env:HOSTNAME=hscale-dev1-dn1
2016-12-02 19:13:58,757 INFO  [main] util.ServerCommandLine: env:QTDIR=/usr/lib64/qt-3.3
2016-12-02 19:13:58,757 INFO  [main] util.ServerCommandLine: env:NLSPATH=/usr/dt/lib/nls/msg/%L/%N.cat
2016-12-02 19:13:58,757 INFO  [main] util.ServerCommandLine: env:HBASE_THRIFT_OPTS=
2016-12-02 19:13:58,757 INFO  [main] util.ServerCommandLine: env:HBASE_HOME=/usr/hdp/current/hbase-regionserver/bin/..
2016-12-02 19:13:58,757 INFO  [main] util.ServerCommandLine: env:QTLIB=/usr/lib64/qt-3.3/lib
2016-12-02 19:13:58,757 INFO  [main] util.ServerCommandLine: env:HOME=/home/hbase
2016-12-02 19:13:58,757 INFO  [main] util.ServerCommandLine: env:MALLOC_ARENA_MAX=4
2016-12-02 19:13:58,758 INFO  [main] util.ServerCommandLine: vmName=Java HotSpot(TM) 64-Bit Server VM, vmVendor=Oracle Corporation, vmVersion=25.60-b23
2016-12-02 19:13:58,758 INFO  [main] util.ServerCommandLine: vmInputArguments=[-Dproc_regionserver, -XX:OnOutOfMemoryError=kill -9 %p, -Dhdp.version=2.4.2.0-258, -XX:ErrorFile=/var/log/hbase/hs_err_pid%p.log, -Djava.io.tmpdir=/tmp, -verbose:gc, -XX:+PrintGCDetails, -XX:+PrintGCDateStamps, -Xloggc:/var/log/hbase/gc.log-201612021913, -Xmn512m, -Xms3072m, -Xmx3072m, -XX:+HeapDumpOnOutOfMemoryError, -XX:MaxDirectMemorySize=2g, -XX:+AlwaysPreTouch, -XX:+UseG1GC, -XX:MaxNewSize=4g, -XX:InitiatingHeapOccupancyPercent=60, -XX:ParallelGCThreads=24, -XX:+ParallelRefProcEnabled, -XX:MaxGCPauseMillis=5000, -Dcom.sun.management.jmxremote.authenticate=false, -Dcom.sun.management.jmxremote.ssl=false, -Dcom.sun.management.jmxremote.port=10102, -Dhbase.log.dir=/var/log/hbase, -Dhbase.log.file=hbase-hbase-regionserver-hscale-dev1-dn1.log, -Dhbase.home.dir=/usr/hdp/current/hbase-regionserver/bin/.., -Dhbase.id.str=hbase, -Dhbase.root.logger=INFO,RFA, -Djava.library.path=:/usr/hdp/2.4.2.0-258/hadoop/lib/native/Linux-amd64-64:/usr/hdp/2.4.2.0-258/hadoop/lib/native, -Dhbase.security.logger=INFO,RFAS]
2016-12-02 19:13:58,956 INFO  [main] regionserver.RSRpcServices: regionserver/hscale-dev1-dn1/10.60.70.11:16020 server-side HConnection retries=50
2016-12-02 19:13:59,033 INFO  [main] ipc.SimpleRpcScheduler: Using deadline as user call queue, count=10
2016-12-02 19:13:59,042 INFO  [main] ipc.RpcServer: regionserver/hscale-dev1-dn1/10.60.70.11:16020: started 10 reader(s).
2016-12-02 19:13:59,075 INFO  [main] impl.MetricsConfig: loaded properties from hadoop-metrics2-hbase.properties
2016-12-02 19:13:59,091 INFO  [main] timeline.HadoopTimelineMetricsSink: Initializing Timeline metrics sink.
2016-12-02 19:13:59,091 INFO  [main] timeline.HadoopTimelineMetricsSink: Identified hostname = hscale-dev1-dn1, serviceName = hbase
2016-12-02 19:13:59,094 INFO  [main] timeline.HadoopTimelineMetricsSink: Collector Uri: http://hscale-dev1-nn:6188/ws/v1/timeline/metrics
2016-12-02 19:13:59,100 INFO  [main] impl.MetricsSinkAdapter: Sink timeline started
2016-12-02 19:13:59,109 INFO  [main] impl.MetricsSystemImpl: Scheduled snapshot period at 60 second(s).
2016-12-02 19:13:59,109 INFO  [main] impl.MetricsSystemImpl: HBase metrics system started
2016-12-02 19:13:59,758 INFO  [main] fs.HFileSystem: Added intercepting call to namenode#getBlockLocations so can do block reordering using class class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks
2016-12-02 19:13:59,907 INFO  [main] zookeeper.RecoverableZooKeeper: Process identifier=regionserver:16020 connecting to ZooKeeper ensemble=hscale-dev1-dn1:2181,hscale-dev1-dn3:2181,hscale-dev1-dn2:2181,hscale-dev1-dn4:2181
2016-12-02 19:13:59,912 INFO  [main] zookeeper.ZooKeeper: Client environment:zookeeper.version=3.4.6-258--1, built on 04/25/2016 05:22 GMT
2016-12-02 19:13:59,912 INFO  [main] zookeeper.ZooKeeper: Client environment:host.name=hscale-dev1-dn1
2016-12-02 19:13:59,912 INFO  [main] zookeeper.ZooKeeper: Client environment:java.version=1.8.0_60
2016-12-02 19:13:59,912 INFO  [main] zookeeper.ZooKeeper: Client environment:java.vendor=Oracle Corporation
2016-12-02 19:13:59,912 INFO  [main] zookeeper.ZooKeeper: Client environment:java.home=/usr/jdk64/jdk1.8.0_60/jre
2016-12-02 19:13:59,913 INFO  [main] zookeeper.ZooKeeper: Client environment:java.class.path=/usr/hdp/current/hbase-regionserver/conf:/usr/jdk64/jdk1.8.0_60/lib/tools.jar:/usr/hdp/current/hbase-regionserver/bin/..:/usr/hdp/current/hbase-regionserver/bin/../lib/activation-1.1.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/aopalliance-1.0.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/apacheds-i18n-2.0.0-M15.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/apacheds-kerberos-codec-2.0.0-M15.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/api-asn1-api-1.0.0-M20.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/api-util-1.0.0-M20.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/asm-3.1.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/avro-1.7.4.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/commons-beanutils-1.7.0.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/commons-beanutils-core-1.8.0.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/commons-cli-1.2.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/commons-codec-1.9.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/commons-collections-3.2.2.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/commons-compress-1.4.1.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/commons-configuration-1.6.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/commons-daemon-1.0.13.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/commons-digester-1.8.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/commons-el-1.0.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/commons-httpclient-3.1.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/commons-io-2.4.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/commons-lang-2.6.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/commons-logging-1.2.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/commons-math-2.2.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/commons-math3-3.1.1.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/commons-net-3.1.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/curator-client-2.7.1.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/curator-framework-2.7.1.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/curator-recipes-2.7.1.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/disruptor-3.3.0.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/findbugs-annotations-1.3.9-1.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/gson-2.2.4.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/guava-12.0.1.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/guice-3.0.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/guice-servlet-3.0.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/hbase-annotations-1.1.2.2.4.2.0-258.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/hbase-annotations-1.1.2.2.4.2.0-258-tests.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/hbase-annotations.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/hbase-client-1.1.2.2.4.2.0-258.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/hbase-client.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/hbase-common-1.1.2.2.4.2.0-258.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/hbase-common-1.1.2.2.4.2.0-258-tests.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/hbase-common.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/hbase-examples-1.1.2.2.4.2.0-258.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/hbase-examples.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/hbase-hadoop2-compat-1.1.2.2.4.2.0-258.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/hbase-hadoop2-compat.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/hbase-hadoop-compat-1.1.2.2.4.2.0-258.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/hbase-hadoop-compat.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/hbase-it-1.1.2.2.4.2.0-258.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/hbase-it-1.1.2.2.4.2.0-258-tests.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/hbase-it.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/hbase-prefix-tree-1.1.2.2.4.2.0-258.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/hbase-prefix-tree.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/hbase-procedure-1.1.2.2.4.2.0-258.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/hbase-procedure.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/hbase-protocol-1.1.2.2.4.2.0-258.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/hbase-protocol.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/hbase-resource-bundle-1.1.2.2.4.2.0-258.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/hbase-resource-bundle.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/hbase-rest-1.1.2.2.4.2.0-258.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/hbase-rest.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/hbase-server-1.1.2.2.4.2.0-258.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/hbase-server-1.1.2.2.4.2.0-258-tests.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/hbase-server.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/hbase-shell-1.1.2.2.4.2.0-258.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/hbase-shell.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/hbase-thrift-1.1.2.2.4.2.0-258.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/hbase-thrift.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/htrace-core-3.1.0-incubating.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/httpclient-4.2.5.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/httpcore-4.2.5.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/jackson-core-2.2.3.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/jackson-core-asl-1.9.13.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/jackson-jaxrs-1.9.13.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/jackson-mapper-asl-1.9.13.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/jackson-xc-1.9.13.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/jamon-runtime-2.3.1.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/jasper-compiler-5.5.23.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/jasper-runtime-5.5.23.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/javax.inject-1.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/java-xmlbuilder-0.4.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/jaxb-api-2.2.2.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/jaxb-impl-2.2.3-1.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/jcodings-1.0.8.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/jersey-client-1.9.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/jersey-core-1.9.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/jersey-guice-1.9.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/jersey-json-1.9.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/jersey-server-1.9.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/jets3t-0.9.0.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/jettison-1.3.3.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/jetty-6.1.26.hwx.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/jetty-sslengine-6.1.26.hwx.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/jetty-util-6.1.26.hwx.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/joni-2.1.2.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/jruby-complete-1.6.8.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/jsch-0.1.42.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/jsp-2.1-6.1.14.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/jsp-api-2.1-6.1.14.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/jsr305-1.3.9.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/junit-4.11.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/leveldbjni-all-1.8.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/libthrift-0.9.0.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/log4j-1.2.17.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/metrics-core-2.2.0.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/microsoft-windowsazure-storage-sdk-0.6.0.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/netty-3.2.4.Final.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/netty-all-4.0.23.Final.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/ojdbc6.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/okhttp-2.4.0.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/okio-1.4.0.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/paranamer-2.3.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/phoenix-4.8.0-HBase-1.1-server.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/phoenix-server.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/protobuf-java-2.5.0.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/ranger-hbase-plugin-shim-0.5.0.2.4.2.0-258.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/ranger-plugin-classloader-0.5.0.2.4.2.0-258.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/servlet-api-2.5-6.1.14.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/servlet-api-2.5.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/slf4j-api-1.7.7.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/snappy-java-1.0.4.1.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/spymemcached-2.11.6.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/xercesImpl-2.9.1.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/xml-apis-1.3.04.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/xmlenc-0.52.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/xz-1.0.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/zookeeper.jar:/usr/hdp/2.4.2.0-258/hadoop/conf:/usr/hdp/2.4.2.0-258/hadoop/lib/jackson-core-asl-1.9.13.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/jackson-databind-2.2.3.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/httpclient-4.2.5.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/commons-collections-3.2.2.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/jaxb-api-2.2.2.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/log4j-1.2.17.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/jettison-1.1.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/jackson-jaxrs-1.9.13.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/hamcrest-core-1.3.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/servlet-api-2.5.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/gson-2.2.4.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/spark-yarn-shuffle.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/aws-java-sdk-1.7.4.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/htrace-core-3.1.0-incubating.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/ranger-hdfs-plugin-shim-0.5.0.2.4.2.0-258.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/paranamer-2.3.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/jersey-core-1.9.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/commons-codec-1.4.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/avro-1.7.4.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/commons-math3-3.1.1.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/commons-logging-1.1.3.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/microsoft-windowsazure-storage-sdk-0.6.0.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/apacheds-i18n-2.0.0-M15.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/jetty-6.1.26.hwx.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/httpcore-4.2.5.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/api-util-1.0.0-M20.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/jackson-annotations-2.2.3.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/commons-compress-1.4.1.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/commons-digester-1.8.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/commons-beanutils-1.7.0.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/jackson-mapper-asl-1.9.13.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/slf4j-log4j12-1.7.10.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/curator-framework-2.7.1.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/jets3t-0.9.0.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/commons-httpclient-3.1.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/apacheds-kerberos-codec-2.0.0-M15.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/snappy-java-1.0.4.1.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/junit-4.11.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/azure-storage-2.2.0.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/ojdbc6.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/ranger-plugin-classloader-0.5.0.2.4.2.0-258.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/activation-1.1.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/jackson-xc-1.9.13.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/jersey-server-1.9.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/mockito-all-1.8.5.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/guava-11.0.2.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/netty-3.6.2.Final.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/ranger-yarn-plugin-shim-0.5.0.2.4.2.0-258.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/protobuf-java-2.5.0.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/xmlenc-0.52.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/api-asn1-api-1.0.0-M20.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/commons-configuration-1.6.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/jsp-api-2.1.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/xz-1.0.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/zookeeper-3.4.6.2.4.2.0-258.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/commons-io-2.4.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/curator-client-2.7.1.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/commons-cli-1.2.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/jersey-json-1.9.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/jsch-0.1.42.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/commons-net-3.1.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/java-xmlbuilder-0.4.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/curator-recipes-2.7.1.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/asm-3.2.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/slf4j-api-1.7.10.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/jaxb-impl-2.2.3-1.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/jsr305-3.0.0.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/stax-api-1.0-2.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/commons-lang-2.6.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/commons-beanutils-core-1.8.0.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/jetty-util-6.1.26.hwx.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/jackson-core-2.2.3.jar:/usr/hdp/2.4.2.0-258/hadoop/.//hadoop-nfs-2.7.1.2.4.2.0-258.jar:/usr/hdp/2.4.2.0-258/hadoop/.//hadoop-common-2.7.1.2.4.2.0-258-tests.jar:/usr/hdp/2.4.2.0-258/hadoop/.//hadoop-annotations-2.7.1.2.4.2.0-258.jar:/usr/hdp/2.4.2.0-258/hadoop/.//hadoop-auth-2.7.1.2.4.2.0-258.jar:/usr/hdp/2.4.2.0-258/hadoop/.//hadoop-azure-2.7.1.2.4.2.0-258.jar:/usr/hdp/2.4.2.0-258/hadoop/.//hadoop-aws-2.7.1.2.4.2.0-258.jar:/usr/hdp/2.4.2.0-258/hadoop/.//hadoop-common-tests.jar:/usr/hdp/2.4.2.0-258/hadoop/.//hadoop-common.jar:/usr/hdp/2.4.2.0-258/hadoop/.//hadoop-auth.jar:/usr/hdp/2.4.2.0-258/hadoop/.//hadoop-aws.jar:/usr/hdp/2.4.2.0-258/hadoop/.//hadoop-common-2.7.1.2.4.2.0-258.jar:/usr/hdp/2.4.2.0-258/hadoop/.//hadoop-annotations.jar:/usr/hdp/2.4.2.0-258/hadoop/.//hadoop-azure.jar:/usr/hdp/2.4.2.0-258/hadoop/.//hadoop-nfs.jar:/usr/hdp/2.4.2.0-258/hadoop-hdfs/./:/usr/hdp/2.4.2.0-258/hadoop-hdfs/lib/jackson-core-asl-1.9.13.jar:/usr/hdp/2.4.2.0-258/hadoop-hdfs/lib/okio-1.4.0.jar:/usr/hdp/2.4.2.0-258/hadoop-hdfs/lib/log4j-1.2.17.jar:/usr/hdp/2.4.2.0-258/hadoop-hdfs/lib/servlet-api-2.5.jar:/usr/hdp/2.4.2.0-258/hadoop-hdfs/lib/htrace-core-3.1.0-incubating.jar:/usr/hdp/2.4.2.0-258/hadoop-hdfs/lib/jersey-core-1.9.jar:/usr/hdp/2.4.2.0-258/hadoop-hdfs/lib/commons-codec-1.4.jar:/usr/hdp/2.4.2.0-258/hadoop-hdfs/lib/commons-logging-1.1.3.jar:/usr/hdp/2.4.2.0-258/hadoop-hdfs/lib/jetty-6.1.26.hwx.jar:/usr/hdp/2.4.2.0-258/hadoop-hdfs/lib/jackson-mapper-asl-1.9.13.jar:/usr/hdp/2.4.2.0-258/hadoop-hdfs/lib/xercesImpl-2.9.1.jar:/usr/hdp/2.4.2.0-258/hadoop-hdfs/lib/leveldbjni-all-1.8.jar:/usr/hdp/2.4.2.0-258/hadoop-hdfs/lib/jersey-server-1.9.jar:/usr/hdp/2.4.2.0-258/hadoop-hdfs/lib/guava-11.0.2.jar:/usr/hdp/2.4.2.0-258/hadoop-hdfs/lib/netty-3.6.2.Final.jar:/usr/hdp/2.4.2.0-258/hadoop-hdfs/lib/protobuf-java-2.5.0.jar:/usr/hdp/2.4.2.0-258/hadoop-hdfs/lib/xmlenc-0.52.jar:/usr/hdp/2.4.2.0-258/hadoop-hdfs/lib/xml-apis-1.3.04.jar:/usr/hdp/2.4.2.0-258/hadoop-hdfs/lib/netty-all-4.0.23.Final.jar:/usr/hdp/2.4.2.0-258/hadoop-hdfs/lib/commons-io-2.4.jar:/usr/hdp/2.4.2.0-258/hadoop-hdfs/lib/commons-cli-1.2.jar:/usr/hdp/2.4.2.0-258/hadoop-hdfs/lib/asm-3.2.jar:/usr/hdp/2.4.2.0-258/hadoop-hdfs/lib/jsr305-3.0.0.jar:/usr/hdp/2.4.2.0-258/hadoop-hdfs/lib/commons-lang-2.6.jar:/usr/hdp/2.4.2.0-258/hadoop-hdfs/lib/jetty-util-6.1.26.hwx.jar:/usr/hdp/2.4.2.0-258/hadoop-hdfs/lib/okhttp-2.4.0.jar:/usr/hdp/2.4.2.0-258/hadoop-hdfs/lib/commons-daemon-1.0.13.jar:/usr/hdp/2.4.2.0-258/hadoop-hdfs/.//hadoop-hdfs-2.7.1.2.4.2.0-258-tests.jar:/usr/hdp/2.4.2.0-258/hadoop-hdfs/.//hadoop-hdfs-tests.jar:/usr/hdp/2.4.2.0-258/hadoop-hdfs/.//hadoop-hdfs-nfs-2.7.1.2.4.2.0-258.jar:/usr/hdp/2.4.2.0-258/hadoop-hdfs/.//hadoop-hdfs-nfs.jar:/usr/hdp/2.4.2.0-258/hadoop-hdfs/.//hadoop-hdfs-2.7.1.2.4.2.0-258.jar:/usr/hdp/2.4.2.0-258/hadoop-hdfs/.//hadoop-hdfs.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/lib/jackson-core-asl-1.9.13.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/lib/jackson-databind-2.2.3.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/lib/httpclient-4.2.5.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/lib/commons-collections-3.2.2.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/lib/jaxb-api-2.2.2.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/lib/log4j-1.2.17.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/lib/jettison-1.1.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/lib/jackson-jaxrs-1.9.13.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/lib/guice-3.0.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/lib/servlet-api-2.5.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/lib/gson-2.2.4.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/lib/htrace-core-3.1.0-incubating.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/lib/paranamer-2.3.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/lib/jersey-core-1.9.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/lib/commons-codec-1.4.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/lib/avro-1.7.4.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/lib/commons-math3-3.1.1.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/lib/commons-logging-1.1.3.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/lib/jersey-guice-1.9.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/lib/microsoft-windowsazure-storage-sdk-0.6.0.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/lib/apacheds-i18n-2.0.0-M15.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/lib/jetty-6.1.26.hwx.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/lib/httpcore-4.2.5.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/lib/spark-assembly-hadoop2.7.1.2.4.2.0-258-1.6.2.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/lib/api-util-1.0.0-M20.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/lib/jackson-annotations-2.2.3.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/lib/commons-compress-1.4.1.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/lib/guice-servlet-3.0.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/lib/commons-digester-1.8.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/lib/javassist-3.18.1-GA.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/lib/commons-beanutils-1.7.0.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/lib/jackson-mapper-asl-1.9.13.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/lib/curator-framework-2.7.1.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/lib/jets3t-0.9.0.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/lib/commons-httpclient-3.1.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/lib/apacheds-kerberos-codec-2.0.0-M15.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/lib/snappy-java-1.0.4.1.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/lib/jersey-client-1.9.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/lib/activation-1.1.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/lib/jackson-xc-1.9.13.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/lib/leveldbjni-all-1.8.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/lib/jersey-server-1.9.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/lib/zookeeper-3.4.6.2.4.2.0-258-tests.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/lib/guava-11.0.2.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/lib/netty-3.6.2.Final.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/lib/protobuf-java-2.5.0.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/lib/xmlenc-0.52.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/lib/api-asn1-api-1.0.0-M20.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/lib/commons-configuration-1.6.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/lib/jsp-api-2.1.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/lib/fst-2.24.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/lib/xz-1.0.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/lib/aopalliance-1.0.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/lib/zookeeper-3.4.6.2.4.2.0-258.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/lib/commons-io-2.4.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/lib/curator-client-2.7.1.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/lib/commons-cli-1.2.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/lib/jersey-json-1.9.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/lib/jsch-0.1.42.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/lib/commons-net-3.1.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/lib/objenesis-2.1.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/lib/java-xmlbuilder-0.4.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/lib/curator-recipes-2.7.1.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/lib/asm-3.2.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/lib/jaxb-impl-2.2.3-1.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/lib/jsr305-3.0.0.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/lib/stax-api-1.0-2.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/lib/commons-lang-2.6.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/lib/javax.inject-1.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/lib/commons-beanutils-core-1.8.0.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/lib/jetty-util-6.1.26.hwx.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/lib/jackson-core-2.2.3.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/.//hadoop-yarn-server-common.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/.//hadoop-yarn-applications-distributedshell-2.7.1.2.4.2.0-258.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/.//hadoop-yarn-server-nodemanager-2.7.1.2.4.2.0-258.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/.//hadoop-yarn-server-common-2.7.1.2.4.2.0-258.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/.//hadoop-yarn-server-sharedcachemanager-2.7.1.2.4.2.0-258.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/.//hadoop-yarn-server-timeline-plugins-2.7.1.2.4.2.0-258.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/.//hadoop-yarn-registry-2.7.1.2.4.2.0-258.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/.//hadoop-yarn-server-applicationhistoryservice.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/.//hadoop-yarn-server-sharedcachemanager.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/.//hadoop-yarn-common-2.7.1.2.4.2.0-258.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/.//hadoop-yarn-server-web-proxy-2.7.1.2.4.2.0-258.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/.//hadoop-yarn-client.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/.//hadoop-yarn-common.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/.//hadoop-yarn-client-2.7.1.2.4.2.0-258.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/.//hadoop-yarn-registry.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/.//hadoop-yarn-api.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/.//hadoop-yarn-server-tests-2.7.1.2.4.2.0-258.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/.//hadoop-yarn-server-nodemanager.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/.//hadoop-yarn-server-resourcemanager.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/.//hadoop-yarn-server-resourcemanager-2.7.1.2.4.2.0-258.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/.//hadoop-yarn-applications-distributedshell.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/.//hadoop-yarn-applications-unmanaged-am-launcher-2.7.1.2.4.2.0-258.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/.//hadoop-yarn-applications-unmanaged-am-launcher.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/.//hadoop-yarn-server-applicationhistoryservice-2.7.1.2.4.2.0-258.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/.//hadoop-yarn-server-web-proxy.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/.//hadoop-yarn-api-2.7.1.2.4.2.0-258.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/.//hadoop-yarn-server-tests.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/.//hadoop-yarn-server-timeline-plugins.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/lib/jackson-core-asl-1.9.13.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/lib/log4j-1.2.17.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/lib/hamcrest-core-1.3.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/lib/guice-3.0.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/lib/paranamer-2.3.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/lib/jersey-core-1.9.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/lib/avro-1.7.4.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/lib/jersey-guice-1.9.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/lib/commons-compress-1.4.1.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/lib/guice-servlet-3.0.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/lib/jackson-mapper-asl-1.9.13.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/lib/snappy-java-1.0.4.1.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/lib/junit-4.11.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/lib/leveldbjni-all-1.8.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/lib/jersey-server-1.9.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/lib/netty-3.6.2.Final.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/lib/protobuf-java-2.5.0.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/lib/xz-1.0.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/lib/aopalliance-1.0.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/lib/commons-io-2.4.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/lib/asm-3.2.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/lib/javax.inject-1.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//jackson-core-asl-1.9.13.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//hadoop-openstack.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//httpclient-4.2.5.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//joda-time-2.9.3.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//commons-collections-3.2.2.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//hadoop-sls.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//hadoop-mapreduce-client-core.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//jaxb-api-2.2.2.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//hadoop-mapreduce-examples-2.7.1.2.4.2.0-258.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//hadoop-mapreduce-client-app.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//log4j-1.2.17.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//jettison-1.1.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//jackson-jaxrs-1.9.13.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//hamcrest-core-1.3.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//servlet-api-2.5.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//hadoop-ant-2.7.1.2.4.2.0-258.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//hadoop-mapreduce-client-jobclient-2.7.1.2.4.2.0-258.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//hadoop-auth-2.7.1.2.4.2.0-258.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//hadoop-sls-2.7.1.2.4.2.0-258.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//gson-2.2.4.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//htrace-core-3.1.0-incubating.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//hadoop-mapreduce-client-core-2.7.1.2.4.2.0-258.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//paranamer-2.3.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//jersey-core-1.9.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//hadoop-rumen.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//commons-codec-1.4.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//avro-1.7.4.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//commons-math3-3.1.1.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//hadoop-datajoin-2.7.1.2.4.2.0-258.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//commons-logging-1.1.3.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//hadoop-datajoin.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//hadoop-streaming.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//microsoft-windowsazure-storage-sdk-0.6.0.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//apacheds-i18n-2.0.0-M15.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//jetty-6.1.26.hwx.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//httpcore-4.2.5.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//hadoop-mapreduce-client-hs-plugins.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//api-util-1.0.0-M20.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//commons-compress-1.4.1.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//commons-digester-1.8.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//commons-beanutils-1.7.0.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//jackson-mapper-asl-1.9.13.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//curator-framework-2.7.1.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//jets3t-0.9.0.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//commons-httpclient-3.1.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//apacheds-kerberos-codec-2.0.0-M15.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//hadoop-mapreduce-client-hs-plugins-2.7.1.2.4.2.0-258.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//snappy-java-1.0.4.1.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//junit-4.11.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//hadoop-mapreduce-client-common.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//hadoop-auth.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//metrics-core-3.0.1.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//activation-1.1.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//jackson-xc-1.9.13.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//jersey-server-1.9.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//mockito-all-1.8.5.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//hadoop-mapreduce-client-hs.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//guava-11.0.2.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//netty-3.6.2.Final.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//protobuf-java-2.5.0.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//xmlenc-0.52.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//api-asn1-api-1.0.0-M20.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//hadoop-archives-2.7.1.2.4.2.0-258.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//commons-configuration-1.6.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//hadoop-mapreduce-client-hs-2.7.1.2.4.2.0-258.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//jsp-api-2.1.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//xz-1.0.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//zookeeper-3.4.6.2.4.2.0-258.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//commons-io-2.4.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//curator-client-2.7.1.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//commons-cli-1.2.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//jersey-json-1.9.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//jsch-0.1.42.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//hadoop-mapreduce-client-jobclient-tests.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//commons-net-3.1.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//hadoop-extras-2.7.1.2.4.2.0-258.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//hadoop-rumen-2.7.1.2.4.2.0-258.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//hadoop-mapreduce-client-app-2.7.1.2.4.2.0-258.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//hadoop-openstack-2.7.1.2.4.2.0-258.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//java-xmlbuilder-0.4.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//hadoop-archives.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//hadoop-gridmix.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//hadoop-ant.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//curator-recipes-2.7.1.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//hadoop-mapreduce-client-shuffle.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//asm-3.2.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//hadoop-gridmix-2.7.1.2.4.2.0-258.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//jaxb-impl-2.2.3-1.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//jsr305-3.0.0.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//hadoop-streaming-2.7.1.2.4.2.0-258.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//hadoop-distcp.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//stax-api-1.0-2.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//commons-lang-2.6.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//hadoop-mapreduce-client-jobclient.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//commons-beanutils-core-1.8.0.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//jetty-util-6.1.26.hwx.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//commons-lang3-3.3.2.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//hadoop-mapreduce-examples.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//jackson-core-2.2.3.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//hadoop-mapreduce-client-shuffle-2.7.1.2.4.2.0-258.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//hadoop-mapreduce-client-common-2.7.1.2.4.2.0-258.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//hadoop-mapreduce-client-jobclient-2.7.1.2.4.2.0-258-tests.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//hadoop-distcp-2.7.1.2.4.2.0-258.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//hadoop-extras.jar::mysql-connector-java-5.1.17.jar:mysql-connector-java-5.1.17.jar_bkp:mysql-connector-java.jar:/usr/hdp/2.4.2.0-258/tez/tez-yarn-timeline-history-with-acls-0.7.0.2.4.2.0-258.jar:/usr/hdp/2.4.2.0-258/tez/tez-runtime-library-0.7.0.2.4.2.0-258.jar:/usr/hdp/2.4.2.0-258/tez/tez-common-0.7.0.2.4.2.0-258.jar:/usr/hdp/2.4.2.0-258/tez/tez-dag-0.7.0.2.4.2.0-258.jar:/usr/hdp/2.4.2.0-258/tez/tez-api-0.7.0.2.4.2.0-258.jar:/usr/hdp/2.4.2.0-258/tez/tez-tests-0.7.0.2.4.2.0-258.jar:/usr/hdp/2.4.2.0-258/tez/tez-examples-0.7.0.2.4.2.0-258.jar:/usr/hdp/2.4.2.0-258/tez/tez-runtime-internals-0.7.0.2.4.2.0-258.jar:/usr/hdp/2.4.2.0-258/tez/tez-mapreduce-0.7.0.2.4.2.0-258.jar:/usr/hdp/2.4.2.0-258/tez/tez-yarn-timeline-history-0.7.0.2.4.2.0-258.jar:/usr/hdp/2.4.2.0-258/tez/tez-history-parser-0.7.0.2.4.2.0-258.jar:/usr/hdp/2.4.2.0-258/tez/tez-yarn-timeline-history-with-fs-0.7.0.2.4.2.0-258.jar:/usr/hdp/2.4.2.0-258/tez/tez-yarn-timeline-cache-plugin-0.7.0.2.4.2.0-258.jar:/usr/hdp/2.4.2.0-258/tez/lib/commons-collections-3.2.2.jar:/usr/hdp/2.4.2.0-258/tez/lib/servlet-api-2.5.jar:/usr/hdp/2.4.2.0-258/tez/lib/hadoop-annotations-2.7.1.2.4.2.0-258.jar:/usr/hdp/2.4.2.0-258/tez/lib/hadoop-yarn-server-timeline-plugins-2.7.1.2.4.2.0-258.jar:/usr/hdp/2.4.2.0-258/tez/lib/slf4j-api-1.7.5.jar:/usr/hdp/2.4.2.0-258/tez/lib/hadoop-mapreduce-client-core-2.7.1.2.4.2.0-258.jar:/usr/hdp/2.4.2.0-258/tez/lib/commons-codec-1.4.jar:/usr/hdp/2.4.2.0-258/tez/lib/commons-math3-3.1.1.jar:/usr/hdp/2.4.2.0-258/tez/lib/hadoop-azure-2.7.1.2.4.2.0-258.jar:/usr/hdp/2.4.2.0-258/tez/lib/hadoop-aws-2.7.1.2.4.2.0-258.jar:/usr/hdp/2.4.2.0-258/tez/lib/jetty-6.1.26.hwx.jar:/usr/hdp/2.4.2.0-258/tez/lib/hadoop-yarn-server-web-proxy-2.7.1.2.4.2.0-258.jar:/usr/hdp/2.4.2.0-258/tez/lib/jersey-client-1.9.jar:/usr/hdp/2.4.2.0-258/tez/lib/commons-collections4-4.1.jar:/usr/hdp/2.4.2.0-258/tez/lib/guava-11.0.2.jar:/usr/hdp/2.4.2.0-258/tez/lib/protobuf-java-2.5.0.jar:/usr/hdp/2.4.2.0-258/tez/lib/jettison-1.3.4.jar:/usr/hdp/2.4.2.0-258/tez/lib/commons-io-2.4.jar:/usr/hdp/2.4.2.0-258/tez/lib/commons-cli-1.2.jar:/usr/hdp/2.4.2.0-258/tez/lib/jersey-json-1.9.jar:/usr/hdp/2.4.2.0-258/tez/lib/jsr305-2.0.3.jar:/usr/hdp/2.4.2.0-258/tez/lib/commons-lang-2.6.jar:/usr/hdp/2.4.2.0-258/tez/lib/jetty-util-6.1.26.hwx.jar:/usr/hdp/2.4.2.0-258/tez/lib/hadoop-mapreduce-client-common-2.7.1.2.4.2.0-258.jar:/usr/hdp/2.4.2.0-258/tez/conf:/usr/hdp/2.4.2.0-258/hadoop/conf:/usr/hdp/2.4.2.0-258/hadoop/hadoop-nfs-2.7.1.2.4.2.0-258.jar:/usr/hdp/2.4.2.0-258/hadoop/hadoop-common-2.7.1.2.4.2.0-258-tests.jar:/usr/hdp/2.4.2.0-258/hadoop/hadoop-annotations-2.7.1.2.4.2.0-258.jar:/usr/hdp/2.4.2.0-258/hadoop/hadoop-auth-2.7.1.2.4.2.0-258.jar:/usr/hdp/2.4.2.0-258/hadoop/hadoop-azure-2.7.1.2.4.2.0-258.jar:/usr/hdp/2.4.2.0-258/hadoop/hadoop-aws-2.7.1.2.4.2.0-258.jar:/usr/hdp/2.4.2.0-258/hadoop/hadoop-common-tests.jar:/usr/hdp/2.4.2.0-258/hadoop/hadoop-common.jar:/usr/hdp/2.4.2.0-258/hadoop/hadoop-auth.jar:/usr/hdp/2.4.2.0-258/hadoop/hadoop-aws.jar:/usr/hdp/2.4.2.0-258/hadoop/hadoop-common-2.7.1.2.4.2.0-258.jar:/usr/hdp/2.4.2.0-258/hadoop/hadoop-annotations.jar:/usr/hdp/2.4.2.0-258/hadoop/hadoop-azure.jar:/usr/hdp/2.4.2.0-258/hadoop/hadoop-nfs.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/jackson-core-asl-1.9.13.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/jackson-databind-2.2.3.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/httpclient-4.2.5.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/commons-collections-3.2.2.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/jaxb-api-2.2.2.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/log4j-1.2.17.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/jettison-1.1.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/jackson-jaxrs-1.9.13.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/hamcrest-core-1.3.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/servlet-api-2.5.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/gson-2.2.4.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/spark-yarn-shuffle.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/aws-java-sdk-1.7.4.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/htrace-core-3.1.0-incubating.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/ranger-hdfs-plugin-shim-0.5.0.2.4.2.0-258.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/paranamer-2.3.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/jersey-core-1.9.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/commons-codec-1.4.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/avro-1.7.4.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/commons-math3-3.1.1.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/commons-logging-1.1.3.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/microsoft-windowsazure-storage-sdk-0.6.0.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/apacheds-i18n-2.0.0-M15.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/jetty-6.1.26.hwx.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/httpcore-4.2.5.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/api-util-1.0.0-M20.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/jackson-annotations-2.2.3.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/commons-compress-1.4.1.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/commons-digester-1.8.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/commons-beanutils-1.7.0.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/jackson-mapper-asl-1.9.13.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/slf4j-log4j12-1.7.10.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/curator-framework-2.7.1.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/jets3t-0.9.0.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/commons-httpclient-3.1.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/apacheds-kerberos-codec-2.0.0-M15.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/snappy-java-1.0.4.1.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/junit-4.11.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/azure-storage-2.2.0.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/ojdbc6.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/ranger-plugin-classloader-0.5.0.2.4.2.0-258.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/activation-1.1.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/jackson-xc-1.9.13.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/jersey-server-1.9.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/mockito-all-1.8.5.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/guava-11.0.2.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/netty-3.6.2.Final.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/ranger-yarn-plugin-shim-0.5.0.2.4.2.0-258.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/protobuf-java-2.5.0.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/xmlenc-0.52.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/api-asn1-api-1.0.0-M20.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/commons-configuration-1.6.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/jsp-api-2.1.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/xz-1.0.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/zookeeper-3.4.6.2.4.2.0-258.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/commons-io-2.4.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/curator-client-2.7.1.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/commons-cli-1.2.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/jersey-json-1.9.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/jsch-0.1.42.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/commons-net-3.1.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/java-xmlbuilder-0.4.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/curator-recipes-2.7.1.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/asm-3.2.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/slf4j-api-1.7.10.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/jaxb-impl-2.2.3-1.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/jsr305-3.0.0.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/stax-api-1.0-2.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/commons-lang-2.6.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/commons-beanutils-core-1.8.0.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/jetty-util-6.1.26.hwx.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/jackson-core-2.2.3.jar:/usr/hdp/2.4.2.0-258/zookeeper/zookeeper.jar:/usr/hdp/2.4.2.0-258/zookeeper/zookeeper-3.4.6.2.4.2.0-258.jar:/usr/hdp/2.4.2.0-258/zookeeper/lib/netty-3.7.0.Final.jar:/usr/hdp/2.4.2.0-258/zookeeper/lib/plexus-interpolation-1.11.jar:/usr/hdp/2.4.2.0-258/zookeeper/lib/nekohtml-1.9.6.2.jar:/usr/hdp/2.4.2.0-258/zookeeper/lib/commons-codec-1.6.jar:/usr/hdp/2.4.2.0-258/zookeeper/lib/maven-plugin-registry-2.2.1.jar:/usr/hdp/2.4.2.0-258/zookeeper/lib/jline-0.9.94.jar:/usr/hdp/2.4.2.0-258/zookeeper/lib/wagon-http-2.4.jar:/usr/hdp/2.4.2.0-258/zookeeper/lib/wagon-http-shared-1.0-beta-6.jar:/usr/hdp/2.4.2.0-258/zookeeper/lib/jsoup-1.7.1.jar:/usr/hdp/2.4.2.0-258/zookeeper/lib/maven-artifact-manager-2.2.1.jar:/usr/hdp/2.4.2.0-258/zookeeper/lib/maven-profile-2.2.1.jar:/usr/hdp/2.4.2.0-258/zookeeper/lib/commons-io-2.2.jar:/usr/hdp/2.4.2.0-258/zookeeper/lib/ant-1.8.0.jar:/usr/hdp/2.4.2.0-258/zookeeper/lib/log4j-1.2.16.jar:/usr/hdp/2.4.2.0-258/zookeeper/lib/maven-model-2.2.1.jar:/usr/hdp/2.4.2.0-258/zookeeper/lib/maven-ant-tasks-2.1.3.jar:/usr/hdp/2.4.2.0-258/zookeeper/lib/maven-error-diagnostics-2.2.1.jar:/usr/hdp/2.4.2.0-258/zookeeper/lib/httpclient-4.2.3.jar:/usr/hdp/2.4.2.0-258/zookeeper/lib/slf4j-api-1.6.1.jar:/usr/hdp/2.4.2.0-258/zookeeper/lib/wagon-http-shared4-2.4.jar:/usr/hdp/2.4.2.0-258/zookeeper/lib/backport-util-concurrent-3.1.jar:/usr/hdp/2.4.2.0-258/zookeeper/lib/plexus-utils-3.0.8.jar:/usr/hdp/2.4.2.0-258/zookeeper/lib/maven-settings-2.2.1.jar:/usr/hdp/2.4.2.0-258/zookeeper/lib/slf4j-log4j12-1.6.1.jar:/usr/hdp/2.4.2.0-258/zookeeper/lib/xercesMinimal-1.9.6.2.jar:/usr/hdp/2.4.2.0-258/zookeeper/lib/plexus-container-default-1.0-alpha-9-stable-1.jar:/usr/hdp/2.4.2.0-258/zookeeper/lib/ant-launcher-1.8.0.jar:/usr/hdp/2.4.2.0-258/zookeeper/lib/wagon-file-1.0-beta-6.jar:/usr/hdp/2.4.2.0-258/zookeeper/lib/maven-artifact-2.2.1.jar:/usr/hdp/2.4.2.0-258/zookeeper/lib/commons-logging-1.1.1.jar:/usr/hdp/2.4.2.0-258/zookeeper/lib/classworlds-1.1-alpha-2.jar:/usr/hdp/2.4.2.0-258/zookeeper/lib/wagon-http-lightweight-1.0-beta-6.jar:/usr/hdp/2.4.2.0-258/zookeeper/lib/wagon-provider-api-2.4.jar:/usr/hdp/2.4.2.0-258/zookeeper/lib/maven-repository-metadata-2.2.1.jar:/usr/hdp/2.4.2.0-258/zookeeper/lib/maven-project-2.2.1.jar:/usr/hdp/2.4.2.0-258/zookeeper/lib/httpcore-4.2.3.jar::/opt/splice/default/lib/concurrentlinkedhashmap-lru-1.4.2.jar:/opt/splice/default/lib/db-client-2.0.1.28.jar:/opt/splice/default/lib/db-drda-2.0.1.28.jar:/opt/splice/default/lib/db-engine-2.0.1.28.jar:/opt/splice/default/lib/db-shared-2.0.1.28.jar:/opt/splice/default/lib/db-tools-i18n-2.0.1.28.jar:/opt/splice/default/lib/db-tools-ij-2.0.1.28.jar:/opt/splice/default/lib/disruptor-3.2.1.jar:/opt/splice/default/lib/gson-2.2.2.jar:/opt/splice/default/lib/hbase_pipeline-hdp2.4.2-2.0.1.28.jar:/opt/splice/default/lib/hbase_sql-hdp2.4.2-2.0.1.28.jar:/opt/splice/default/lib/hbase_storage-hdp2.4.2-2.0.1.28.jar:/opt/splice/default/lib/hppc-0.5.2.jar:/opt/splice/default/lib/kryo-2.21.jar:/opt/splice/default/lib/kryo-serializers-0.26.jar:/opt/splice/default/lib/lucene-core-4.3.1.jar:/opt/splice/default/lib/opencsv-2.3.jar:/opt/splice/default/lib/pipeline_api-2.0.1.28.jar:/opt/splice/default/lib/protobuf-java-2.5.0.jar:/opt/splice/default/lib/spark-assembly-hadoop2.7.1.2.4.2.0-258-1.6.2.jar:/opt/splice/default/lib/splice_access_api-2.0.1.28.jar:/opt/splice/default/lib/splice_auth-2.0.1.28.jar:/opt/splice/default/lib/splice_backup-hdp2.4.2-2.0.1.28.jar:/opt/splice/default/lib/splice_colperms-2.0.1.28.jar:/opt/splice/default/lib/splice_ee-hdp2.4.2-2.0.1.28.jar:/opt/splice/default/lib/splice_encoding-2.0.1.28.jar:/opt/splice/default/lib/splice_encryption-2.0.1.28.jar:/opt/splice/default/lib/splice_machine-2.0.1.28.jar:/opt/splice/default/lib/splice_protocol-2.0.1.28.jar:/opt/splice/default/lib/splice_si_api-2.0.1.28.jar:/opt/splice/default/lib/splice_timestamp_api-2.0.1.28.jar:/opt/splice/default/lib/stats-2.0.1.28.jar:/opt/splice/default/lib/super-csv-2.4.0.jar:/opt/splice/default/lib/utilities-2.0.1.28.jar
2016-12-02 19:13:59,914 INFO  [main] zookeeper.ZooKeeper: Client environment:java.library.path=:/usr/hdp/2.4.2.0-258/hadoop/lib/native/Linux-amd64-64:/usr/hdp/2.4.2.0-258/hadoop/lib/native
2016-12-02 19:13:59,914 INFO  [main] zookeeper.ZooKeeper: Client environment:java.io.tmpdir=/tmp
2016-12-02 19:13:59,914 INFO  [main] zookeeper.ZooKeeper: Client environment:java.compiler=<NA>
2016-12-02 19:13:59,914 INFO  [main] zookeeper.ZooKeeper: Client environment:os.name=Linux
2016-12-02 19:13:59,914 INFO  [main] zookeeper.ZooKeeper: Client environment:os.arch=amd64
2016-12-02 19:13:59,914 INFO  [main] zookeeper.ZooKeeper: Client environment:os.version=2.6.32-504.el6.x86_64
2016-12-02 19:13:59,914 INFO  [main] zookeeper.ZooKeeper: Client environment:user.name=hbase
2016-12-02 19:13:59,914 INFO  [main] zookeeper.ZooKeeper: Client environment:user.home=/home/hbase
2016-12-02 19:13:59,914 INFO  [main] zookeeper.ZooKeeper: Client environment:user.dir=/home/hbase
2016-12-02 19:13:59,914 INFO  [main] zookeeper.ZooKeeper: Initiating client connection, connectString=hscale-dev1-dn1:2181,hscale-dev1-dn3:2181,hscale-dev1-dn2:2181,hscale-dev1-dn4:2181 sessionTimeout=120000 watcher=regionserver:160200x0, quorum=hscale-dev1-dn1:2181,hscale-dev1-dn3:2181,hscale-dev1-dn2:2181,hscale-dev1-dn4:2181, baseZNode=/hbase-unsecure
2016-12-02 19:13:59,929 INFO  [main-SendThread(hscale-dev1-dn3:2181)] zookeeper.ClientCnxn: Opening socket connection to server hscale-dev1-dn3/10.60.70.13:2181. Will not attempt to authenticate using SASL (unknown error)
2016-12-02 19:13:59,935 INFO  [main-SendThread(hscale-dev1-dn3:2181)] zookeeper.ClientCnxn: Socket connection established to hscale-dev1-dn3/10.60.70.13:2181, initiating session
2016-12-02 19:13:59,960 INFO  [main-SendThread(hscale-dev1-dn3:2181)] zookeeper.ClientCnxn: Session establishment complete on server hscale-dev1-dn3/10.60.70.13:2181, sessionid = 0x358bf3f651f001c, negotiated timeout = 120000
2016-12-02 19:13:59,970 INFO  [RpcServer.responder] ipc.RpcServer: RpcServer.responder: starting
2016-12-02 19:13:59,971 INFO  [RpcServer.listener,port=16020] ipc.RpcServer: RpcServer.listener,port=16020: starting
2016-12-02 19:14:00,024 INFO  [main] mortbay.log: Logging to org.slf4j.impl.Log4jLoggerAdapter(org.mortbay.log) via org.mortbay.log.Slf4jLog
2016-12-02 19:14:00,026 INFO  [main] http.HttpRequestLog: Http request log for http.requests.regionserver is not defined
2016-12-02 19:14:00,033 INFO  [main] http.HttpServer: Added global filter 'safety' (class=org.apache.hadoop.hbase.http.HttpServer$QuotingInputFilter)
2016-12-02 19:14:00,035 INFO  [main] http.HttpServer: Added filter static_user_filter (class=org.apache.hadoop.hbase.http.lib.StaticUserWebFilter$StaticUserFilter) to context regionserver
2016-12-02 19:14:00,035 INFO  [main] http.HttpServer: Added filter static_user_filter (class=org.apache.hadoop.hbase.http.lib.StaticUserWebFilter$StaticUserFilter) to context static
2016-12-02 19:14:00,035 INFO  [main] http.HttpServer: Added filter static_user_filter (class=org.apache.hadoop.hbase.http.lib.StaticUserWebFilter$StaticUserFilter) to context logs
2016-12-02 19:14:00,044 INFO  [main] http.HttpServer: Jetty bound to port 16030
2016-12-02 19:14:00,044 INFO  [main] mortbay.log: jetty-6.1.26.hwx
2016-12-02 19:14:00,280 INFO  [main] mortbay.log: Started SelectChannelConnector@0.0.0.0:16030
2016-12-02 19:14:00,322 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] zookeeper.RecoverableZooKeeper: Process identifier=hconnection-0x59159a3b connecting to ZooKeeper ensemble=hscale-dev1-dn1:2181,hscale-dev1-dn3:2181,hscale-dev1-dn2:2181,hscale-dev1-dn4:2181
2016-12-02 19:14:00,322 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] zookeeper.ZooKeeper: Initiating client connection, connectString=hscale-dev1-dn1:2181,hscale-dev1-dn3:2181,hscale-dev1-dn2:2181,hscale-dev1-dn4:2181 sessionTimeout=120000 watcher=hconnection-0x59159a3b0x0, quorum=hscale-dev1-dn1:2181,hscale-dev1-dn3:2181,hscale-dev1-dn2:2181,hscale-dev1-dn4:2181, baseZNode=/hbase-unsecure
2016-12-02 19:14:00,323 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020-SendThread(hscale-dev1-dn1:2181)] zookeeper.ClientCnxn: Opening socket connection to server hscale-dev1-dn1/10.60.70.11:2181. Will not attempt to authenticate using SASL (unknown error)
2016-12-02 19:14:00,323 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020-SendThread(hscale-dev1-dn1:2181)] zookeeper.ClientCnxn: Socket connection established to hscale-dev1-dn1/10.60.70.11:2181, initiating session
2016-12-02 19:14:00,343 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020-SendThread(hscale-dev1-dn1:2181)] zookeeper.ClientCnxn: Session establishment complete on server hscale-dev1-dn1/10.60.70.11:2181, sessionid = 0x158bf3f64e00029, negotiated timeout = 120000
2016-12-02 19:14:00,405 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] regionserver.HRegionServer: ClusterId : e53c756c-5d17-4349-aaa8-6d5c497d937d
2016-12-02 19:14:00,451 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] regionserver.MemStoreFlusher: globalMemStoreLimit=768 M, globalMemStoreLimitLowMark=691.2 M, maxHeap=3 G
2016-12-02 19:14:00,454 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] regionserver.HRegionServer: CompactionChecker runs every 10sec
2016-12-02 19:14:00,479 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] regionserver.RegionServerCoprocessorHost: System coprocessor loading is enabled
2016-12-02 19:14:00,479 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] regionserver.RegionServerCoprocessorHost: Table coprocessor loading is enabled
2016-12-02 19:14:00,552 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config: Created Splice configuration.
2016-12-02 19:14:00,555 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  LOG = [org.apache.log4j.Logger@aaa57b3]
2016-12-02 19:14:00,556 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  activeTransactionCacheSize = [4096]
2016-12-02 19:14:00,556 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  authentication = [NATIVE]
2016-12-02 19:14:00,556 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  authenticationCustomProvider = [null]
2016-12-02 19:14:00,556 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  authenticationLdapSearchauthdn = [null]
2016-12-02 19:14:00,556 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  authenticationLdapSearchauthpw = [null]
2016-12-02 19:14:00,556 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  authenticationLdapSearchbase = [null]
2016-12-02 19:14:00,556 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  authenticationLdapSearchfilter = [null]
2016-12-02 19:14:00,556 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  authenticationLdapServer = [null]
2016-12-02 19:14:00,556 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  authenticationNativeAlgorithm = [SHA-512]
2016-12-02 19:14:00,556 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  authenticationNativeCreateCredentialsDatabase = [true]
2016-12-02 19:14:00,556 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  backupParallelism = [16]
2016-12-02 19:14:00,556 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  backupPath = [/backup]
2016-12-02 19:14:00,556 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  batchOnceBatchSize = [50000]
2016-12-02 19:14:00,556 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  broadcastRegionMbThreshold = [30]
2016-12-02 19:14:00,556 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  broadcastRegionRowThreshold = [1000000]
2016-12-02 19:14:00,556 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  cardinalityPrecision = [14]
2016-12-02 19:14:00,556 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  clientPause = [1000]
2016-12-02 19:14:00,556 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  compactionReservedSlots = [1]
2016-12-02 19:14:00,556 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  completedTxnCacheSize = [131072]
2016-12-02 19:14:00,556 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  completedTxnConcurrency = [128]
2016-12-02 19:14:00,556 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  compressionAlgorithm = [snappy]
2016-12-02 19:14:00,557 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  configSource = [com.splicemachine.access.HBaseConfigurationSource@1052f90c]
2016-12-02 19:14:00,557 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  coreWriterThreads = [2]
2016-12-02 19:14:00,557 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  ddlDrainingInitialWait = [1000]
2016-12-02 19:14:00,557 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  ddlDrainingMaximumWait = [100000]
2016-12-02 19:14:00,557 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  ddlRefreshInterval = [10000]
2016-12-02 19:14:00,557 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  debugDumpBindTree = [false]
2016-12-02 19:14:00,557 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  debugDumpClassFile = [false]
2016-12-02 19:14:00,557 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  debugDumpOptimizedTree = [false]
2016-12-02 19:14:00,557 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  debugLogStatementContext = [false]
2016-12-02 19:14:00,557 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.block.access.key.update.interval = [600]
2016-12-02 19:14:00,557 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.block.access.token.enable = [true]
2016-12-02 19:14:00,557 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.block.access.token.lifetime = [600]
2016-12-02 19:14:00,557 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.block.scanner.volume.bytes.per.second = [1048576]
2016-12-02 19:14:00,557 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.blockreport.initialDelay = [120]
2016-12-02 19:14:00,557 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.blockreport.intervalMsec = [21600000]
2016-12-02 19:14:00,557 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.blockreport.split.threshold = [1000000]
2016-12-02 19:14:00,557 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.blocksize = [134217728]
2016-12-02 19:14:00,557 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.bytes-per-checksum = [512]
2016-12-02 19:14:00,557 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.cachereport.intervalMsec = [10000]
2016-12-02 19:14:00,557 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.client-write-packet-size = [65536]
2016-12-02 19:14:00,557 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.client.block.write.replace-datanode-on-failure.best-effort = [false]
2016-12-02 19:14:00,557 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.client.block.write.replace-datanode-on-failure.enable = [true]
2016-12-02 19:14:00,557 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.client.block.write.replace-datanode-on-failure.policy = [DEFAULT]
2016-12-02 19:14:00,557 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.client.block.write.retries = [3]
2016-12-02 19:14:00,557 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.client.cached.conn.retry = [3]
2016-12-02 19:14:00,557 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.client.context = [default]
2016-12-02 19:14:00,557 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.client.datanode-restart.timeout = [30]
2016-12-02 19:14:00,558 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.client.domain.socket.data.traffic = [false]
2016-12-02 19:14:00,558 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.client.failover.connection.retries = [0]
2016-12-02 19:14:00,558 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.client.failover.connection.retries.on.timeouts = [0]
2016-12-02 19:14:00,558 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.client.failover.max.attempts = [15]
2016-12-02 19:14:00,558 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.client.failover.sleep.base.millis = [500]
2016-12-02 19:14:00,558 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.client.failover.sleep.max.millis = [15000]
2016-12-02 19:14:00,558 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.client.file-block-storage-locations.num-threads = [10]
2016-12-02 19:14:00,558 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.client.file-block-storage-locations.timeout.millis = [1000]
2016-12-02 19:14:00,558 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.client.https.keystore.resource = [ssl-client.xml]
2016-12-02 19:14:00,558 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.client.https.need-auth = [false]
2016-12-02 19:14:00,558 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.client.mmap.cache.size = [256]
2016-12-02 19:14:00,558 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.client.mmap.cache.timeout.ms = [3600000]
2016-12-02 19:14:00,558 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.client.mmap.enabled = [true]
2016-12-02 19:14:00,558 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.client.mmap.retry.timeout.ms = [300000]
2016-12-02 19:14:00,558 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.client.read.shortcircuit = [true]
2016-12-02 19:14:00,558 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.client.read.shortcircuit.buffer.size = [131072]
2016-12-02 19:14:00,558 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.client.read.shortcircuit.skip.checksum = [false]
2016-12-02 19:14:00,558 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.client.read.shortcircuit.streams.cache.expiry.ms = [300000]
2016-12-02 19:14:00,558 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.client.read.shortcircuit.streams.cache.size = [4096]
2016-12-02 19:14:00,558 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.client.retry.policy.enabled = [false]
2016-12-02 19:14:00,558 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.client.short.circuit.replica.stale.threshold.ms = [1800000]
2016-12-02 19:14:00,558 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.client.slow.io.warning.threshold.ms = [30000]
2016-12-02 19:14:00,558 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.client.use.datanode.hostname = [false]
2016-12-02 19:14:00,558 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.client.use.legacy.blockreader.local = [false]
2016-12-02 19:14:00,558 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.client.write.exclude.nodes.cache.expiry.interval.millis = [600000]
2016-12-02 19:14:00,558 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.cluster.administrators = [ hdfs]
2016-12-02 19:14:00,558 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.content-summary.limit = [5000]
2016-12-02 19:14:00,559 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.datanode.address = [0.0.0.0:50010]
2016-12-02 19:14:00,559 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.datanode.available-space-volume-choosing-policy.balanced-space-preference-fraction = [0.75f]
2016-12-02 19:14:00,559 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.datanode.available-space-volume-choosing-policy.balanced-space-threshold = [10737418240]
2016-12-02 19:14:00,559 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.datanode.balance.bandwidthPerSec = [6250000]
2016-12-02 19:14:00,559 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.datanode.block-pinning.enabled = [false]
2016-12-02 19:14:00,559 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.datanode.block.id.layout.upgrade.threads = [12]
2016-12-02 19:14:00,559 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.datanode.bp-ready.timeout = [20]
2016-12-02 19:14:00,559 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.datanode.cache.revocation.polling.ms = [500]
2016-12-02 19:14:00,559 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.datanode.cache.revocation.timeout.ms = [900000]
2016-12-02 19:14:00,559 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.datanode.data.dir = [/hadoop/hdfs/data]
2016-12-02 19:14:00,559 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.datanode.data.dir.perm = [750]
2016-12-02 19:14:00,559 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.datanode.directoryscan.interval = [21600]
2016-12-02 19:14:00,559 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.datanode.directoryscan.threads = [1]
2016-12-02 19:14:00,559 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.datanode.dns.interface = [default]
2016-12-02 19:14:00,559 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.datanode.dns.nameserver = [default]
2016-12-02 19:14:00,559 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.datanode.drop.cache.behind.reads = [false]
2016-12-02 19:14:00,559 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.datanode.drop.cache.behind.writes = [false]
2016-12-02 19:14:00,559 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.datanode.du.reserved = [1073741824]
2016-12-02 19:14:00,559 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.datanode.failed.volumes.tolerated = [0]
2016-12-02 19:14:00,559 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.datanode.fsdatasetcache.max.threads.per.volume = [4]
2016-12-02 19:14:00,559 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.datanode.handler.count = [20]
2016-12-02 19:14:00,559 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.datanode.hdfs-blocks-metadata.enabled = [false]
2016-12-02 19:14:00,559 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.datanode.http.address = [0.0.0.0:50075]
2016-12-02 19:14:00,559 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.datanode.https.address = [0.0.0.0:50475]
2016-12-02 19:14:00,559 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.datanode.ipc.address = [0.0.0.0:8010]
2016-12-02 19:14:00,559 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.datanode.max.locked.memory = [0]
2016-12-02 19:14:00,559 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.datanode.max.transfer.threads = [1024]
2016-12-02 19:14:00,559 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.datanode.readahead.bytes = [4193404]
2016-12-02 19:14:00,559 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.datanode.scan.period.hours = [504]
2016-12-02 19:14:00,559 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.datanode.shared.file.descriptor.paths = [/dev/shm,/tmp]
2016-12-02 19:14:00,559 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.datanode.slow.io.warning.threshold.ms = [300]
2016-12-02 19:14:00,560 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.datanode.sync.behind.writes = [false]
2016-12-02 19:14:00,560 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.datanode.transfer.socket.recv.buffer.size = [131072]
2016-12-02 19:14:00,560 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.datanode.transfer.socket.send.buffer.size = [131072]
2016-12-02 19:14:00,560 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.datanode.use.datanode.hostname = [false]
2016-12-02 19:14:00,560 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.default.chunk.view.size = [32768]
2016-12-02 19:14:00,560 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.domain.socket.path = [/var/lib/hadoop-hdfs/dn_socket]
2016-12-02 19:14:00,560 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.encrypt.data.transfer = [false]
2016-12-02 19:14:00,560 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.encrypt.data.transfer.cipher.key.bitlength = [128]
2016-12-02 19:14:00,560 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.encrypt.data.transfer.cipher.suites = [AES/CTR/NoPadding]
2016-12-02 19:14:00,560 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.ha.automatic-failover.enabled = [false]
2016-12-02 19:14:00,560 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.ha.fencing.ssh.connect-timeout = [30000]
2016-12-02 19:14:00,560 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.ha.log-roll.period = [120]
2016-12-02 19:14:00,560 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.ha.tail-edits.period = [60]
2016-12-02 19:14:00,560 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.heartbeat.interval = [3]
2016-12-02 19:14:00,560 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.hosts.exclude = [/etc/hadoop/conf/dfs.exclude]
2016-12-02 19:14:00,560 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.http.policy = [HTTP_ONLY]
2016-12-02 19:14:00,560 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.https.port = [50470]
2016-12-02 19:14:00,560 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.https.server.keystore.resource = [ssl-server.xml]
2016-12-02 19:14:00,560 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.image.compress = [false]
2016-12-02 19:14:00,560 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.image.compression.codec = [org.apache.hadoop.io.compress.DefaultCodec]
2016-12-02 19:14:00,560 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.image.transfer.bandwidthPerSec = [0]
2016-12-02 19:14:00,560 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.image.transfer.chunksize = [65536]
2016-12-02 19:14:00,560 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.image.transfer.timeout = [60000]
2016-12-02 19:14:00,560 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.journalnode.edits.dir = [/hadoop/hdfs/journalnode]
2016-12-02 19:14:00,560 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.journalnode.http-address = [0.0.0.0:8480]
2016-12-02 19:14:00,560 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.journalnode.https-address = [0.0.0.0:8481]
2016-12-02 19:14:00,560 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.journalnode.rpc-address = [0.0.0.0:8485]
2016-12-02 19:14:00,560 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.namenode.accesstime.precision = [0]
2016-12-02 19:14:00,560 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.namenode.acls.enabled = [false]
2016-12-02 19:14:00,560 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.namenode.audit.log.async = [true]
2016-12-02 19:14:00,560 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.namenode.audit.loggers = [default]
2016-12-02 19:14:00,560 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.namenode.avoid.read.stale.datanode = [true]
2016-12-02 19:14:00,561 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.namenode.avoid.write.stale.datanode = [true]
2016-12-02 19:14:00,561 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.namenode.backup.address = [0.0.0.0:50100]
2016-12-02 19:14:00,561 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.namenode.backup.http-address = [0.0.0.0:50105]
2016-12-02 19:14:00,561 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.namenode.blocks.per.postponedblocks.rescan = [10000]
2016-12-02 19:14:00,561 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.namenode.checkpoint.check.period = [60]
2016-12-02 19:14:00,561 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.namenode.checkpoint.dir = [/hadoop/hdfs/namesecondary]
2016-12-02 19:14:00,561 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.namenode.checkpoint.edits.dir = [/hadoop/hdfs/namesecondary]
2016-12-02 19:14:00,561 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.namenode.checkpoint.max-retries = [3]
2016-12-02 19:14:00,561 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.namenode.checkpoint.period = [21600]
2016-12-02 19:14:00,561 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.namenode.checkpoint.txns = [1000000]
2016-12-02 19:14:00,561 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.namenode.datanode.registration.ip-hostname-check = [true]
2016-12-02 19:14:00,561 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.namenode.decommission.blocks.per.interval = [500000]
2016-12-02 19:14:00,561 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.namenode.decommission.interval = [30]
2016-12-02 19:14:00,561 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.namenode.decommission.max.concurrent.tracked.nodes = [100]
2016-12-02 19:14:00,561 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.namenode.delegation.key.update-interval = [86400000]
2016-12-02 19:14:00,561 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.namenode.delegation.token.max-lifetime = [604800000]
2016-12-02 19:14:00,561 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.namenode.delegation.token.renew-interval = [86400000]
2016-12-02 19:14:00,561 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.namenode.edit.log.autoroll.check.interval.ms = [300000]
2016-12-02 19:14:00,561 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.namenode.edit.log.autoroll.multiplier.threshold = [2.0]
2016-12-02 19:14:00,561 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.namenode.edits.dir = [/hadoop/hdfs/namenode]
2016-12-02 19:14:00,561 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.namenode.edits.journal-plugin.qjournal = [org.apache.hadoop.hdfs.qjournal.client.QuorumJournalManager]
2016-12-02 19:14:00,561 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.namenode.edits.noeditlogchannelflush = [false]
2016-12-02 19:14:00,561 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.namenode.enable.retrycache = [true]
2016-12-02 19:14:00,561 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.namenode.fs-limits.max-blocks-per-file = [1048576]
2016-12-02 19:14:00,561 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.namenode.fs-limits.max-component-length = [255]
2016-12-02 19:14:00,561 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.namenode.fs-limits.max-directory-items = [1048576]
2016-12-02 19:14:00,561 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.namenode.fs-limits.max-xattr-size = [16384]
2016-12-02 19:14:00,561 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.namenode.fs-limits.max-xattrs-per-inode = [32]
2016-12-02 19:14:00,561 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.namenode.fs-limits.min-block-size = [1048576]
2016-12-02 19:14:00,561 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.namenode.fslock.fair = [false]
2016-12-02 19:14:00,561 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.namenode.full.block.report.lease.length.ms = [300000]
2016-12-02 19:14:00,561 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.namenode.handler.count = [100]
2016-12-02 19:14:00,561 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.namenode.heartbeat.recheck-interval = [300000]
2016-12-02 19:14:00,561 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.namenode.http-address = [hscale-dev1-nn:50070]
2016-12-02 19:14:00,561 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.namenode.https-address = [hscale-dev1-nn:50470]
2016-12-02 19:14:00,561 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.namenode.inotify.max.events.per.rpc = [1000]
2016-12-02 19:14:00,562 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.namenode.invalidate.work.pct.per.iteration = [0.32f]
2016-12-02 19:14:00,562 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.namenode.kerberos.internal.spnego.principal = [${dfs.web.authentication.kerberos.principal}]
2016-12-02 19:14:00,562 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.namenode.kerberos.principal.pattern = [*]
2016-12-02 19:14:00,562 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.namenode.lazypersist.file.scrub.interval.sec = [300]
2016-12-02 19:14:00,562 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.namenode.lifeline.handler.ratio = [0.10]
2016-12-02 19:14:00,562 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.namenode.list.cache.directives.num.responses = [100]
2016-12-02 19:14:00,562 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.namenode.list.cache.pools.num.responses = [100]
2016-12-02 19:14:00,562 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.namenode.list.encryption.zones.num.responses = [100]
2016-12-02 19:14:00,562 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.namenode.max.extra.edits.segments.retained = [10000]
2016-12-02 19:14:00,562 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.namenode.max.full.block.report.leases = [6]
2016-12-02 19:14:00,562 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.namenode.max.objects = [0]
2016-12-02 19:14:00,562 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.namenode.metrics.logger.period.seconds = [600]
2016-12-02 19:14:00,562 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.namenode.name.dir = [/hadoop/hdfs/namenode]
2016-12-02 19:14:00,562 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.namenode.name.dir.restore = [true]
2016-12-02 19:14:00,562 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.namenode.num.checkpoints.retained = [2]
2016-12-02 19:14:00,562 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.namenode.num.extra.edits.retained = [1000000]
2016-12-02 19:14:00,562 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.namenode.path.based.cache.block.map.allocation.percent = [0.25]
2016-12-02 19:14:00,562 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.namenode.path.based.cache.refresh.interval.ms = [30000]
2016-12-02 19:14:00,562 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.namenode.path.based.cache.retry.interval.ms = [30000]
2016-12-02 19:14:00,562 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.namenode.reject-unresolved-dn-topology-mapping = [false]
2016-12-02 19:14:00,562 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.namenode.replication.considerLoad = [true]
2016-12-02 19:14:00,562 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.namenode.replication.interval = [3]
2016-12-02 19:14:00,562 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.namenode.replication.min = [1]
2016-12-02 19:14:00,562 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.namenode.replication.work.multiplier.per.iteration = [2]
2016-12-02 19:14:00,562 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.namenode.resource.check.interval = [5000]
2016-12-02 19:14:00,562 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.namenode.resource.checked.volumes.minimum = [1]
2016-12-02 19:14:00,562 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.namenode.resource.du.reserved = [104857600]
2016-12-02 19:14:00,562 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.namenode.retrycache.expirytime.millis = [600000]
2016-12-02 19:14:00,562 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.namenode.retrycache.heap.percent = [0.03f]
2016-12-02 19:14:00,562 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.namenode.rpc-address = [hscale-dev1-nn:8020]
2016-12-02 19:14:00,562 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.namenode.safemode.extension = [30000]
2016-12-02 19:14:00,562 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.namenode.safemode.min.datanodes = [0]
2016-12-02 19:14:00,562 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.namenode.safemode.threshold-pct = [0.999]
2016-12-02 19:14:00,562 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.namenode.secondary.http-address = [hscale-dev1-dn1:50090]
2016-12-02 19:14:00,562 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.namenode.secondary.https-address = [0.0.0.0:50091]
2016-12-02 19:14:00,562 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.namenode.stale.datanode.interval = [30000]
2016-12-02 19:14:00,562 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.namenode.startup.delay.block.deletion.sec = [3600]
2016-12-02 19:14:00,562 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.namenode.support.allow.format = [true]
2016-12-02 19:14:00,562 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.namenode.top.enabled = [true]
2016-12-02 19:14:00,563 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.namenode.top.num.users = [10]
2016-12-02 19:14:00,563 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.namenode.top.window.num.buckets = [10]
2016-12-02 19:14:00,563 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.namenode.top.windows.minutes = [1,5,25]
2016-12-02 19:14:00,563 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.namenode.write.stale.datanode.ratio = [1.0f]
2016-12-02 19:14:00,563 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.namenode.xattrs.enabled = [true]
2016-12-02 19:14:00,563 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.permissions.enabled = [true]
2016-12-02 19:14:00,563 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.permissions.superusergroup = [hdfs]
2016-12-02 19:14:00,563 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.replication = [3]
2016-12-02 19:14:00,563 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.replication.max = [50]
2016-12-02 19:14:00,563 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.secondary.namenode.kerberos.internal.spnego.principal = [${dfs.web.authentication.kerberos.principal}]
2016-12-02 19:14:00,563 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.short.circuit.shared.memory.watcher.interrupt.check.ms = [60000]
2016-12-02 19:14:00,563 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.storage.policy.enabled = [true]
2016-12-02 19:14:00,563 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.stream-buffer-size = [4096]
2016-12-02 19:14:00,563 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.support.append = [true]
2016-12-02 19:14:00,563 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.user.home.dir.prefix = [/user]
2016-12-02 19:14:00,563 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.webhdfs.enabled = [true]
2016-12-02 19:14:00,563 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.webhdfs.rest-csrf.browser-useragents-regex = [^Mozilla.*,^Opera.*]
2016-12-02 19:14:00,563 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.webhdfs.rest-csrf.custom-header = [X-XSRF-HEADER]
2016-12-02 19:14:00,563 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.webhdfs.rest-csrf.enabled = [false]
2016-12-02 19:14:00,563 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.webhdfs.rest-csrf.methods-to-ignore = [GET,OPTIONS,HEAD,TRACE]
2016-12-02 19:14:00,563 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.webhdfs.ugi.expire.after.access = [600000]
2016-12-02 19:14:00,563 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.webhdfs.user.provider.user.pattern = [^[A-Za-z_][A-Za-z0-9._-]*[$]?$]
2016-12-02 19:14:00,563 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  fallbackLocalLatency = [1]
2016-12-02 19:14:00,563 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  fallbackMinimumRowCount = [20]
2016-12-02 19:14:00,563 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  fallbackNullFraction = [0.1]
2016-12-02 19:14:00,563 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  fallbackOpencloseLatency = [2000]
2016-12-02 19:14:00,563 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  fallbackRegionRowCount = [5000000]
2016-12-02 19:14:00,563 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  fallbackRemoteLatencyRatio = [10]
2016-12-02 19:14:00,563 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  fallbackRowWidth = [170]
2016-12-02 19:14:00,563 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  file.blocksize = [67108864]
2016-12-02 19:14:00,563 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  file.bytes-per-checksum = [512]
2016-12-02 19:14:00,563 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  file.client-write-packet-size = [65536]
2016-12-02 19:14:00,563 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  file.replication = [1]
2016-12-02 19:14:00,563 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  file.stream-buffer-size = [4096]
2016-12-02 19:14:00,564 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  fs.AbstractFileSystem.file.impl = [org.apache.hadoop.fs.local.LocalFs]
2016-12-02 19:14:00,564 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  fs.AbstractFileSystem.ftp.impl = [org.apache.hadoop.fs.ftp.FtpFs]
2016-12-02 19:14:00,564 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  fs.AbstractFileSystem.har.impl = [org.apache.hadoop.fs.HarFs]
2016-12-02 19:14:00,564 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  fs.AbstractFileSystem.hdfs.impl = [org.apache.hadoop.fs.Hdfs]
2016-12-02 19:14:00,564 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  fs.AbstractFileSystem.swebhdfs.impl = [org.apache.hadoop.fs.SWebHdfs]
2016-12-02 19:14:00,564 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  fs.AbstractFileSystem.viewfs.impl = [org.apache.hadoop.fs.viewfs.ViewFs]
2016-12-02 19:14:00,564 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  fs.AbstractFileSystem.webhdfs.impl = [org.apache.hadoop.fs.WebHdfs]
2016-12-02 19:14:00,564 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  fs.automatic.close = [true]
2016-12-02 19:14:00,564 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  fs.client.resolve.remote.symlinks = [true]
2016-12-02 19:14:00,564 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  fs.defaultFS = [hdfs://hscale-dev1-nn:8020]
2016-12-02 19:14:00,564 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  fs.df.interval = [60000]
2016-12-02 19:14:00,564 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  fs.du.interval = [600000]
2016-12-02 19:14:00,564 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  fs.ftp.host = [0.0.0.0]
2016-12-02 19:14:00,564 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  fs.ftp.host.port = [21]
2016-12-02 19:14:00,564 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  fs.har.impl.disable.cache = [true]
2016-12-02 19:14:00,564 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  fs.permissions.umask-mode = [022]
2016-12-02 19:14:00,564 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  fs.s3.block.size = [67108864]
2016-12-02 19:14:00,564 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  fs.s3.buffer.dir = [/tmp/hadoop-hbase/s3]
2016-12-02 19:14:00,564 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  fs.s3.maxRetries = [4]
2016-12-02 19:14:00,564 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  fs.s3.sleepTimeSeconds = [10]
2016-12-02 19:14:00,564 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  fs.s3a.attempts.maximum = [10]
2016-12-02 19:14:00,564 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  fs.s3a.buffer.dir = [/tmp/hadoop-hbase/s3a]
2016-12-02 19:14:00,564 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  fs.s3a.connection.establish.timeout = [5000]
2016-12-02 19:14:00,564 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  fs.s3a.connection.maximum = [15]
2016-12-02 19:14:00,564 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  fs.s3a.connection.ssl.enabled = [true]
2016-12-02 19:14:00,564 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  fs.s3a.connection.timeout = [50000]
2016-12-02 19:14:00,564 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  fs.s3a.fast.buffer.size = [1048576]
2016-12-02 19:14:00,564 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  fs.s3a.fast.upload = [false]
2016-12-02 19:14:00,564 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  fs.s3a.impl = [org.apache.hadoop.fs.s3a.S3AFileSystem]
2016-12-02 19:14:00,564 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  fs.s3a.max.total.tasks = [1000]
2016-12-02 19:14:00,564 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  fs.s3a.multipart.purge = [false]
2016-12-02 19:14:00,564 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  fs.s3a.multipart.purge.age = [86400]
2016-12-02 19:14:00,564 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  fs.s3a.multipart.size = [104857600]
2016-12-02 19:14:00,564 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  fs.s3a.multipart.threshold = [2147483647]
2016-12-02 19:14:00,564 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  fs.s3a.paging.maximum = [5000]
2016-12-02 19:14:00,564 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  fs.s3a.threads.core = [15]
2016-12-02 19:14:00,564 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  fs.s3a.threads.keepalivetime = [60]
2016-12-02 19:14:00,564 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  fs.s3a.threads.max = [256]
2016-12-02 19:14:00,565 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  fs.s3n.block.size = [67108864]
2016-12-02 19:14:00,565 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  fs.s3n.multipart.copy.block.size = [5368709120]
2016-12-02 19:14:00,565 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  fs.s3n.multipart.uploads.block.size = [67108864]
2016-12-02 19:14:00,565 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  fs.s3n.multipart.uploads.enabled = [false]
2016-12-02 19:14:00,565 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  fs.swift.impl = [org.apache.hadoop.fs.swift.snative.SwiftNativeFileSystem]
2016-12-02 19:14:00,565 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  fs.trash.checkpoint.interval = [0]
2016-12-02 19:14:00,565 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  fs.trash.interval = [360]
2016-12-02 19:14:00,565 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  ftp.blocksize = [67108864]
2016-12-02 19:14:00,565 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  ftp.bytes-per-checksum = [512]
2016-12-02 19:14:00,565 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  ftp.client-write-packet-size = [65536]
2016-12-02 19:14:00,565 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  ftp.replication = [3]
2016-12-02 19:14:00,565 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  ftp.stream-buffer-size = [4096]
2016-12-02 19:14:00,565 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  ha.failover-controller.active-standby-elector.zk.op.retries = [120]
2016-12-02 19:14:00,565 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  ha.failover-controller.cli-check.rpc-timeout.ms = [20000]
2016-12-02 19:14:00,565 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  ha.failover-controller.graceful-fence.connection.retries = [1]
2016-12-02 19:14:00,565 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  ha.failover-controller.graceful-fence.rpc-timeout.ms = [5000]
2016-12-02 19:14:00,565 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  ha.failover-controller.new-active.rpc-timeout.ms = [60000]
2016-12-02 19:14:00,565 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  ha.health-monitor.check-interval.ms = [1000]
2016-12-02 19:14:00,565 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  ha.health-monitor.connect-retry-interval.ms = [1000]
2016-12-02 19:14:00,565 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  ha.health-monitor.rpc-timeout.ms = [45000]
2016-12-02 19:14:00,565 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  ha.health-monitor.sleep-after-disconnect.ms = [1000]
2016-12-02 19:14:00,565 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  ha.zookeeper.acl = [world:anyone:rwcda]
2016-12-02 19:14:00,565 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  ha.zookeeper.parent-znode = [/hadoop-ha]
2016-12-02 19:14:00,565 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  ha.zookeeper.session-timeout.ms = [5000]
2016-12-02 19:14:00,565 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hadoop.common.configuration.version = [0.23.0]
2016-12-02 19:14:00,565 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hadoop.fuse.connection.timeout = [300]
2016-12-02 19:14:00,565 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hadoop.fuse.timer.period = [5]
2016-12-02 19:14:00,565 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hadoop.hdfs.configuration.version = [1]
2016-12-02 19:14:00,565 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hadoop.http.authentication.kerberos.keytab = [/home/hbase/hadoop.keytab]
2016-12-02 19:14:00,565 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hadoop.http.authentication.kerberos.principal = [HTTP/_HOST@LOCALHOST]
2016-12-02 19:14:00,565 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hadoop.http.authentication.signature.secret.file = [/home/hbase/hadoop-http-auth-signature-secret]
2016-12-02 19:14:00,565 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hadoop.http.authentication.simple.anonymous.allowed = [true]
2016-12-02 19:14:00,565 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hadoop.http.authentication.token.validity = [36000]
2016-12-02 19:14:00,565 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hadoop.http.authentication.type = [simple]
2016-12-02 19:14:00,565 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hadoop.http.cross-origin.allowed-headers = [X-Requested-With,Content-Type,Accept,Origin]
2016-12-02 19:14:00,565 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hadoop.http.cross-origin.allowed-methods = [GET,POST,HEAD]
2016-12-02 19:14:00,565 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hadoop.http.cross-origin.allowed-origins = [*]
2016-12-02 19:14:00,565 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hadoop.http.cross-origin.enabled = [false]
2016-12-02 19:14:00,565 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hadoop.http.cross-origin.max-age = [1800]
2016-12-02 19:14:00,565 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hadoop.http.filter.initializers = [org.apache.hadoop.http.lib.StaticUserWebFilter,org.apache.hadoop.security.HttpCrossOriginFilterInitializer]
2016-12-02 19:14:00,566 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hadoop.http.staticuser.user = [dr.who]
2016-12-02 19:14:00,566 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hadoop.jetty.logs.serve.aliases = [true]
2016-12-02 19:14:00,566 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hadoop.kerberos.kinit.command = [kinit]
2016-12-02 19:14:00,566 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hadoop.policy.file = [hbase-policy.xml]
2016-12-02 19:14:00,566 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hadoop.proxyuser.hbase.groups = [*]
2016-12-02 19:14:00,566 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hadoop.proxyuser.hbase.hosts = [*]
2016-12-02 19:14:00,566 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hadoop.proxyuser.hcat.groups = [*]
2016-12-02 19:14:00,566 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hadoop.proxyuser.hcat.hosts = [hscale-dev1-dn2]
2016-12-02 19:14:00,566 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hadoop.proxyuser.hdfs.groups = [*]
2016-12-02 19:14:00,566 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hadoop.proxyuser.hdfs.hosts = [*]
2016-12-02 19:14:00,566 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hadoop.proxyuser.hive.groups = [*]
2016-12-02 19:14:00,566 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hadoop.proxyuser.hive.hosts = [hscale-dev1-dn2]
2016-12-02 19:14:00,566 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hadoop.proxyuser.oozie.groups = [*]
2016-12-02 19:14:00,566 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hadoop.proxyuser.oozie.hosts = [hscale-dev1-dn3]
2016-12-02 19:14:00,566 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hadoop.registry.jaas.context = [Client]
2016-12-02 19:14:00,566 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hadoop.registry.rm.enabled = [false]
2016-12-02 19:14:00,566 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hadoop.registry.secure = [false]
2016-12-02 19:14:00,566 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hadoop.registry.system.acls = [sasl:yarn@, sasl:mapred@, sasl:hdfs@]
2016-12-02 19:14:00,566 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hadoop.registry.zk.connection.timeout.ms = [15000]
2016-12-02 19:14:00,566 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hadoop.registry.zk.quorum = [hscale-dev1-dn1:2181,hscale-dev1-dn3:2181,hscale-dev1-dn2:2181,hscale-dev1-dn4:2181]
2016-12-02 19:14:00,566 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hadoop.registry.zk.retry.ceiling.ms = [60000]
2016-12-02 19:14:00,566 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hadoop.registry.zk.retry.interval.ms = [1000]
2016-12-02 19:14:00,566 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hadoop.registry.zk.retry.times = [5]
2016-12-02 19:14:00,566 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hadoop.registry.zk.root = [/registry]
2016-12-02 19:14:00,566 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hadoop.registry.zk.session.timeout.ms = [60000]
2016-12-02 19:14:00,566 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hadoop.rpc.protection = [authentication]
2016-12-02 19:14:00,566 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hadoop.rpc.socket.factory.class.default = [org.apache.hadoop.net.StandardSocketFactory]
2016-12-02 19:14:00,566 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hadoop.security.auth_to_local = [RULE:[1:$1@$0](ambari-qa@HSCALE.COM)s/.*/ambari-qa/
RULE:[1:$1@$0](hbase@HSCALE.COM)s/.*/hbase/
RULE:[1:$1@$0](hdfs@HSCALE.COM)s/.*/hdfs/
RULE:[1:$1@$0](spark@HSCALE.COM)s/.*/spark/
RULE:[1:$1@$0](.*@HSCALE.COM)s/@.*//
RULE:[2:$1@$0](amshbase@HSCALE.COM)s/.*/ams/
RULE:[2:$1@$0](amszk@HSCALE.COM)s/.*/ams/
RULE:[2:$1@$0](dn@HSCALE.COM)s/.*/hdfs/
RULE:[2:$1@$0](hbase@HSCALE.COM)s/.*/hbase/
RULE:[2:$1@$0](hive@HSCALE.COM)s/.*/hive/
RULE:[2:$1@$0](jhs@HSCALE.COM)s/.*/mapred/
RULE:[2:$1@$0](nm@HSCALE.COM)s/.*/yarn/
RULE:[2:$1@$0](nn@HSCALE.COM)s/.*/hdfs/
RULE:[2:$1@$0](oozie@HSCALE.COM)s/.*/oozie/
RULE:[2:$1@$0](rm@HSCALE.COM)s/.*/yarn/
RULE:[2:$1@$0](yarn@HSCALE.COM)s/.*/yarn/
DEFAULT]
2016-12-02 19:14:00,566 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hadoop.security.authentication = [simple]
2016-12-02 19:14:00,566 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hadoop.security.authorization = [false]
2016-12-02 19:14:00,566 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hadoop.security.crypto.buffer.size = [8192]
2016-12-02 19:14:00,566 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hadoop.security.crypto.cipher.suite = [AES/CTR/NoPadding]
2016-12-02 19:14:00,566 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hadoop.security.crypto.codec.classes.aes.ctr.nopadding = [org.apache.hadoop.crypto.OpensslAesCtrCryptoCodec,org.apache.hadoop.crypto.JceAesCtrCryptoCodec]
2016-12-02 19:14:00,566 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hadoop.security.dns.log-slow-lookups.enabled = [false]
2016-12-02 19:14:00,566 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hadoop.security.dns.log-slow-lookups.threshold.ms = [1000]
2016-12-02 19:14:00,566 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hadoop.security.group.mapping = [org.apache.hadoop.security.JniBasedUnixGroupsMappingWithFallback]
2016-12-02 19:14:00,566 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hadoop.security.group.mapping.ldap.directory.search.timeout = [10000]
2016-12-02 19:14:00,566 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hadoop.security.group.mapping.ldap.search.attr.group.name = [cn]
2016-12-02 19:14:00,566 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hadoop.security.group.mapping.ldap.search.attr.member = [member]
2016-12-02 19:14:00,566 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hadoop.security.group.mapping.ldap.search.filter.group = [(objectClass=group)]
2016-12-02 19:14:00,566 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hadoop.security.group.mapping.ldap.search.filter.user = [(&(objectClass=user)(sAMAccountName={0}))]
2016-12-02 19:14:00,566 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hadoop.security.group.mapping.ldap.ssl = [false]
2016-12-02 19:14:00,567 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hadoop.security.groups.cache.secs = [300]
2016-12-02 19:14:00,567 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hadoop.security.groups.cache.warn.after.ms = [5000]
2016-12-02 19:14:00,567 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hadoop.security.groups.negative-cache.secs = [30]
2016-12-02 19:14:00,567 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hadoop.security.instrumentation.requires.admin = [false]
2016-12-02 19:14:00,567 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hadoop.security.java.secure.random.algorithm = [SHA1PRNG]
2016-12-02 19:14:00,567 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hadoop.security.kms.client.authentication.retry-count = [1]
2016-12-02 19:14:00,567 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hadoop.security.kms.client.encrypted.key.cache.expiry = [43200000]
2016-12-02 19:14:00,567 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hadoop.security.kms.client.encrypted.key.cache.low-watermark = [0.3f]
2016-12-02 19:14:00,567 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hadoop.security.kms.client.encrypted.key.cache.num.refill.threads = [2]
2016-12-02 19:14:00,567 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hadoop.security.kms.client.encrypted.key.cache.size = [500]
2016-12-02 19:14:00,567 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hadoop.security.random.device.file.path = [/dev/urandom]
2016-12-02 19:14:00,567 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hadoop.security.uid.cache.secs = [14400]
2016-12-02 19:14:00,567 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hadoop.shell.safely.delete.limit.num.files = [100]
2016-12-02 19:14:00,567 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hadoop.ssl.client.conf = [ssl-client.xml]
2016-12-02 19:14:00,567 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hadoop.ssl.enabled = [false]
2016-12-02 19:14:00,567 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hadoop.ssl.enabled.protocols = [TLSv1]
2016-12-02 19:14:00,567 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hadoop.ssl.hostname.verifier = [DEFAULT]
2016-12-02 19:14:00,567 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hadoop.ssl.keystores.factory.class = [org.apache.hadoop.security.ssl.FileBasedKeyStoresFactory]
2016-12-02 19:14:00,567 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hadoop.ssl.require.client.cert = [false]
2016-12-02 19:14:00,567 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hadoop.ssl.server.conf = [ssl-server.xml]
2016-12-02 19:14:00,567 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hadoop.tmp.dir = [/tmp/hadoop-hbase]
2016-12-02 19:14:00,567 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hadoop.user.group.static.mapping.overrides = [dr.who=;]
2016-12-02 19:14:00,567 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hadoop.util.hash.type = [murmur]
2016-12-02 19:14:00,567 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hadoop.work.around.non.threadsafe.getpwuid = [false]
2016-12-02 19:14:00,567 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.auth.key.update.interval = [86400000]
2016-12-02 19:14:00,567 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.auth.token.max.lifetime = [604800000]
2016-12-02 19:14:00,567 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.balancer.period = [60000]
2016-12-02 19:14:00,567 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.bulkload.retries.number = [10]
2016-12-02 19:14:00,567 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.bulkload.staging.dir = [/apps/hbase/staging]
2016-12-02 19:14:00,567 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.cells.scanned.per.heartbeat.check = [10000]
2016-12-02 19:14:00,567 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.client.ipc.pool.size = [10]
2016-12-02 19:14:00,567 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.client.keyvalue.maxsize = [1048576]
2016-12-02 19:14:00,567 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.client.localityCheck.threadPoolSize = [2]
2016-12-02 19:14:00,567 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.client.max.perregion.tasks = [100]
2016-12-02 19:14:00,567 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.client.max.perserver.tasks = [5]
2016-12-02 19:14:00,567 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.client.max.total.tasks = [100]
2016-12-02 19:14:00,567 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.client.pause = [100]
2016-12-02 19:14:00,567 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.client.retries.number = [5]
2016-12-02 19:14:00,567 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.client.scanner.caching = [1000]
2016-12-02 19:14:00,567 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.client.scanner.timeout.period = [60000]
2016-12-02 19:14:00,567 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.client.write.buffer = [2097152]
2016-12-02 19:14:00,567 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.cluster.distributed = [true]
2016-12-02 19:14:00,567 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.column.max.version = [1]
2016-12-02 19:14:00,568 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.config.read.zookeeper.config = [false]
2016-12-02 19:14:00,568 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.coordinated.state.manager.class = [org.apache.hadoop.hbase.coordination.ZkCoordinatedStateManager]
2016-12-02 19:14:00,568 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.coprocessor.abortonerror = [true]
2016-12-02 19:14:00,568 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.coprocessor.enabled = [true]
2016-12-02 19:14:00,568 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.coprocessor.master.classes = [com.splicemachine.hbase.SpliceMasterObserver]
2016-12-02 19:14:00,568 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.coprocessor.region.classes = [org.apache.hadoop.hbase.security.access.SecureBulkLoadEndpoint,com.splicemachine.hbase.MemstoreAwareObserver,com.splicemachine.derby.hbase.SpliceIndexObserver,com.splicemachine.derby.hbase.SpliceIndexEndpoint,com.splicemachine.hbase.RegionSizeEndpoint,com.splicemachine.si.data.hbase.coprocessor.TxnLifecycleEndpoint,com.splicemachine.si.data.hbase.coprocessor.SIObserver,com.splicemachine.hbase.BackupEndpointObserver]
2016-12-02 19:14:00,568 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.coprocessor.regionserver.classes = [com.splicemachine.hbase.RegionServerLifecycleObserver]
2016-12-02 19:14:00,568 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.coprocessor.user.enabled = [true]
2016-12-02 19:14:00,568 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.data.umask = [000]
2016-12-02 19:14:00,568 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.data.umask.enable = [false]
2016-12-02 19:14:00,568 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.defaults.for.version = [1.1.2.2.4.2.0-258]
2016-12-02 19:14:00,568 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.defaults.for.version.skip = [true]
2016-12-02 19:14:00,568 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.dfs.client.read.shortcircuit.buffer.size = [131072]
2016-12-02 19:14:00,568 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.dynamic.jars.dir = [hdfs://hscale-dev1-nn:8020/apps/hbase/data/lib]
2016-12-02 19:14:00,568 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.fs.tmp.dir = [/user/hbase/hbase-staging]
2016-12-02 19:14:00,568 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.hregion.majorcompaction = [604800000]
2016-12-02 19:14:00,568 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.hregion.majorcompaction.jitter = [0.50]
2016-12-02 19:14:00,568 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.hregion.max.filesize = [10737418240]
2016-12-02 19:14:00,568 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.hregion.memstore.block.multiplier = [4]
2016-12-02 19:14:00,568 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.hregion.memstore.flush.size = [134217728]
2016-12-02 19:14:00,568 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.hregion.memstore.mslab.enabled = [true]
2016-12-02 19:14:00,568 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.hregion.percolumnfamilyflush.size.lower.bound = [16777216]
2016-12-02 19:14:00,568 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.hregion.preclose.flush.size = [5242880]
2016-12-02 19:14:00,568 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.hstore.blockingStoreFiles = [20]
2016-12-02 19:14:00,568 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.hstore.blockingWaitTime = [90000]
2016-12-02 19:14:00,568 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.hstore.bytes.per.checksum = [16384]
2016-12-02 19:14:00,568 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.hstore.checksum.algorithm = [CRC32]
2016-12-02 19:14:00,568 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.hstore.compaction.kv.max = [10]
2016-12-02 19:14:00,568 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.hstore.compaction.max = [10]
2016-12-02 19:14:00,568 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.hstore.compaction.max.size = [260046848]
2016-12-02 19:14:00,568 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.hstore.compaction.min = [5]
2016-12-02 19:14:00,568 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.hstore.compaction.min.size = [16777216]
2016-12-02 19:14:00,568 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.hstore.compactionThreshold = [3]
2016-12-02 19:14:00,568 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.hstore.defaultengine.compactionpolicy.class = [com.splicemachine.compactions.SpliceDefaultCompactionPolicy]
2016-12-02 19:14:00,568 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.hstore.defaultengine.compactor.class = [com.splicemachine.compactions.SpliceDefaultCompactor]
2016-12-02 19:14:00,568 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.hstore.flusher.count = [2]
2016-12-02 19:14:00,568 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.hstore.time.to.purge.deletes = [0]
2016-12-02 19:14:00,568 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.htable.threads.max = [96]
2016-12-02 19:14:00,568 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.http.filter.initializers = [org.apache.hadoop.hbase.http.lib.StaticUserWebFilter]
2016-12-02 19:14:00,568 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.http.max.threads = [10]
2016-12-02 19:14:00,568 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.http.staticuser.user = [dr.stack]
2016-12-02 19:14:00,568 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.ipc.client.fallback-to-simple-auth-allowed = [false]
2016-12-02 19:14:00,568 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.ipc.client.tcpnodelay = [true]
2016-12-02 19:14:00,568 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.ipc.server.callqueue.handler.factor = [0.1]
2016-12-02 19:14:00,568 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.ipc.server.callqueue.read.ratio = [0]
2016-12-02 19:14:00,568 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.ipc.server.callqueue.scan.ratio = [0]
2016-12-02 19:14:00,569 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.ipc.warn.response.size = [-1]
2016-12-02 19:14:00,569 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.ipc.warn.response.time = [-1]
2016-12-02 19:14:00,569 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.lease.recovery.dfs.timeout = [64000]
2016-12-02 19:14:00,569 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.lease.recovery.timeout = [900000]
2016-12-02 19:14:00,569 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.local.dir = [/tmp/hbase-hbase/local]
2016-12-02 19:14:00,569 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.master.catalog.timeout = [600000]
2016-12-02 19:14:00,569 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.master.distributed.log.replay = [false]
2016-12-02 19:14:00,569 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.master.hfilecleaner.plugins = [org.apache.hadoop.hbase.master.cleaner.TimeToLiveHFileCleaner]
2016-12-02 19:14:00,569 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.master.info.bindAddress = [0.0.0.0]
2016-12-02 19:14:00,569 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.master.info.port = [16010]
2016-12-02 19:14:00,569 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.master.infoserver.redirect = [true]
2016-12-02 19:14:00,569 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.master.loadbalance.bytable = [TRUE]
2016-12-02 19:14:00,569 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.master.loadbalancer.class = [org.apache.hadoop.hbase.master.balancer.StochasticLoadBalancer]
2016-12-02 19:14:00,569 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.master.logcleaner.plugins = [org.apache.hadoop.hbase.master.cleaner.TimeToLiveLogCleaner]
2016-12-02 19:14:00,569 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.master.logcleaner.ttl = [600000]
2016-12-02 19:14:00,569 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.master.normalizer.class = [org.apache.hadoop.hbase.master.normalizer.SimpleRegionNormalizer]
2016-12-02 19:14:00,569 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.master.port = [16000]
2016-12-02 19:14:00,569 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.metrics.exposeOperationTimes = [true]
2016-12-02 19:14:00,569 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.metrics.showTableName = [true]
2016-12-02 19:14:00,569 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.mvcc.impl = [org.apache.hadoop.hbase.regionserver.SIMultiVersionConsistencyControl]
2016-12-02 19:14:00,569 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.normalizer.enabled = [false]
2016-12-02 19:14:00,569 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.normalizer.period = [1800000]
2016-12-02 19:14:00,569 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.online.schema.update.enable = [true]
2016-12-02 19:14:00,569 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.region.replica.replication.enabled = [false]
2016-12-02 19:14:00,569 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.regions.slop = [0.01]
2016-12-02 19:14:00,569 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.regionserver.catalog.timeout = [600000]
2016-12-02 19:14:00,569 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.regionserver.checksum.verify = [true]
2016-12-02 19:14:00,569 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.regionserver.dns.interface = [default]
2016-12-02 19:14:00,569 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.regionserver.dns.nameserver = [default]
2016-12-02 19:14:00,569 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.regionserver.global.memstore.size = [0.25]
2016-12-02 19:14:00,569 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.regionserver.global.memstore.size.lower.limit = [0.9]
2016-12-02 19:14:00,569 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.regionserver.handler.abort.on.error.percent = [0.5]
2016-12-02 19:14:00,569 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.regionserver.handler.count = [100]
2016-12-02 19:14:00,569 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.regionserver.hlog.reader.impl = [org.apache.hadoop.hbase.regionserver.wal.ProtobufLogReader]
2016-12-02 19:14:00,569 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.regionserver.hlog.writer.impl = [org.apache.hadoop.hbase.regionserver.wal.ProtobufLogWriter]
2016-12-02 19:14:00,569 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.regionserver.info.bindAddress = [0.0.0.0]
2016-12-02 19:14:00,569 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.regionserver.info.port = [16030]
2016-12-02 19:14:00,569 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.regionserver.info.port.auto = [false]
2016-12-02 19:14:00,569 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.regionserver.lease.period = [1200000]
2016-12-02 19:14:00,569 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.regionserver.logroll.errors.tolerated = [2]
2016-12-02 19:14:00,569 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.regionserver.logroll.period = [3600000]
2016-12-02 19:14:00,569 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.regionserver.maxlogs = [48]
2016-12-02 19:14:00,569 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.regionserver.msginterval = [3000]
2016-12-02 19:14:00,569 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.regionserver.optionalcacheflushinterval = [3600000]
2016-12-02 19:14:00,569 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.regionserver.port = [16020]
2016-12-02 19:14:00,569 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.regionserver.region.split.policy = [org.apache.hadoop.hbase.regionserver.IncreasingToUpperBoundRegionSplitPolicy]
2016-12-02 19:14:00,569 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.regionserver.regionSplitLimit = [1000]
2016-12-02 19:14:00,569 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.regionserver.storefile.refresh.period = [0]
2016-12-02 19:14:00,569 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.regionserver.thread.compaction.large = [1]
2016-12-02 19:14:00,569 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.regionserver.thread.compaction.small = [4]
2016-12-02 19:14:00,570 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.regionserver.thrift.compact = [false]
2016-12-02 19:14:00,570 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.regionserver.thrift.framed = [false]
2016-12-02 19:14:00,570 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.regionserver.thrift.framed.max_frame_size_in_mb = [2]
2016-12-02 19:14:00,570 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.regionserver.wal.codec = [org.apache.hadoop.hbase.regionserver.wal.WALCellCodec]
2016-12-02 19:14:00,570 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.regionserver.wal.enablecompression = [TRUE]
2016-12-02 19:14:00,570 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.replication.rpc.codec = [org.apache.hadoop.hbase.codec.KeyValueCodecWithTags]
2016-12-02 19:14:00,570 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.rest.filter.classes = [org.apache.hadoop.hbase.rest.filter.GzipFilter]
2016-12-02 19:14:00,570 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.rest.port = [8080]
2016-12-02 19:14:00,570 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.rest.readonly = [false]
2016-12-02 19:14:00,570 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.rest.support.proxyuser = [false]
2016-12-02 19:14:00,570 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.rest.threads.max = [100]
2016-12-02 19:14:00,570 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.rest.threads.min = [2]
2016-12-02 19:14:00,570 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.rootdir = [hdfs://hscale-dev1-nn:8020/apps/hbase/data]
2016-12-02 19:14:00,570 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.rootdir.perms = [700]
2016-12-02 19:14:00,570 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.rpc.protection = [authentication]
2016-12-02 19:14:00,570 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.rpc.shortoperation.timeout = [10000]
2016-12-02 19:14:00,570 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.rpc.timeout = [1200000]
2016-12-02 19:14:00,570 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.rs.cacheblocksonwrite = [false]
2016-12-02 19:14:00,570 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.security.authentication = [simple]
2016-12-02 19:14:00,570 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.security.authorization = [false]
2016-12-02 19:14:00,570 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.security.exec.permission.checks = [false]
2016-12-02 19:14:00,570 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.security.visibility.mutations.checkauths = [false]
2016-12-02 19:14:00,570 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.server.compactchecker.interval.multiplier = [1000]
2016-12-02 19:14:00,570 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.server.scanner.max.result.size = [104857600]
2016-12-02 19:14:00,570 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.server.thread.wakefrequency = [10000]
2016-12-02 19:14:00,570 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.server.versionfile.writeattempts = [3]
2016-12-02 19:14:00,570 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.snapshot.enabled = [true]
2016-12-02 19:14:00,570 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.snapshot.master.timeout.millis = [300000]
2016-12-02 19:14:00,570 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.snapshot.region.timeout = [300000]
2016-12-02 19:14:00,570 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.snapshot.restore.failsafe.name = [hbase-failsafe-{snapshot.name}-{restore.timestamp}]
2016-12-02 19:14:00,570 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.snapshot.restore.take.failsafe.snapshot = [true]
2016-12-02 19:14:00,570 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.splitlog.manager.timeout = [3000]
2016-12-02 19:14:00,570 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.status.listener.class = [org.apache.hadoop.hbase.client.ClusterStatusListener$MulticastListener]
2016-12-02 19:14:00,570 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.status.multicast.address.ip = [226.1.1.3]
2016-12-02 19:14:00,570 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.status.multicast.address.port = [16100]
2016-12-02 19:14:00,570 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.status.multicast.port = [16100]
2016-12-02 19:14:00,570 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.status.published = [false]
2016-12-02 19:14:00,570 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.status.publisher.class = [org.apache.hadoop.hbase.master.ClusterStatusPublisher$MulticastPublisher]
2016-12-02 19:14:00,570 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.storescanner.parallel.seek.enable = [false]
2016-12-02 19:14:00,570 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.storescanner.parallel.seek.threads = [10]
2016-12-02 19:14:00,570 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.superuser = [hbase]
2016-12-02 19:14:00,570 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.table.lock.enable = [true]
2016-12-02 19:14:00,570 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.table.max.rowsize = [1073741824]
2016-12-02 19:14:00,570 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.thrift.htablepool.size.max = [1000]
2016-12-02 19:14:00,570 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.thrift.maxQueuedRequests = [1000]
2016-12-02 19:14:00,570 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.thrift.maxWorkerThreads = [1000]
2016-12-02 19:14:00,570 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.thrift.minWorkerThreads = [16]
2016-12-02 19:14:00,570 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.tmp.dir = [/tmp/hbase-hbase]
2016-12-02 19:14:00,570 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.wal.disruptor.batch = [TRUE]
2016-12-02 19:14:00,571 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.wal.provider = [multiwal]
2016-12-02 19:14:00,571 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.wal.regiongrouping.numgroups = [16]
2016-12-02 19:14:00,571 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.zookeeper.dns.interface = [default]
2016-12-02 19:14:00,571 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.zookeeper.dns.nameserver = [default]
2016-12-02 19:14:00,571 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.zookeeper.leaderport = [3888]
2016-12-02 19:14:00,571 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.zookeeper.peerport = [2888]
2016-12-02 19:14:00,571 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.zookeeper.property.clientPort = [2181]
2016-12-02 19:14:00,571 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.zookeeper.property.dataDir = [/tmp/hbase-hbase/zookeeper]
2016-12-02 19:14:00,571 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.zookeeper.property.initLimit = [10]
2016-12-02 19:14:00,571 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.zookeeper.property.maxClientCnxns = [300]
2016-12-02 19:14:00,571 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.zookeeper.property.syncLimit = [5]
2016-12-02 19:14:00,571 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.zookeeper.property.tickTime = [6000]
2016-12-02 19:14:00,571 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.zookeeper.quorum = [hscale-dev1-dn1,hscale-dev1-dn3,hscale-dev1-dn2,hscale-dev1-dn4]
2016-12-02 19:14:00,571 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.zookeeper.useMulti = [true]
2016-12-02 19:14:00,571 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbaseSecurityAuthentication = [false]
2016-12-02 19:14:00,571 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbaseSecurityAuthorization = [simple]
2016-12-02 19:14:00,571 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hfile.block.bloom.cacheonwrite = [TRUE]
2016-12-02 19:14:00,571 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hfile.block.cache.size = [0.25]
2016-12-02 19:14:00,571 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hfile.block.index.cacheonwrite = [false]
2016-12-02 19:14:00,571 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hfile.format.version = [3]
2016-12-02 19:14:00,571 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hfile.index.block.max.size = [131072]
2016-12-02 19:14:00,571 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  ignoreSavePoints = [false]
2016-12-02 19:14:00,571 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  importMaxQuotedColumnLines = [50000]
2016-12-02 19:14:00,571 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  indexBatchSize = [4000]
2016-12-02 19:14:00,571 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  indexFetchSampleSize = [128]
2016-12-02 19:14:00,571 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  indexLookupBlocks = [5]
2016-12-02 19:14:00,571 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  io.compression.codec.bzip2.library = [system-native]
2016-12-02 19:14:00,571 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  io.compression.codecs = [org.apache.hadoop.io.compress.GzipCodec,org.apache.hadoop.io.compress.DefaultCodec,org.apache.hadoop.io.compress.SnappyCodec]
2016-12-02 19:14:00,571 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  io.file.buffer.size = [131072]
2016-12-02 19:14:00,571 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  io.map.index.interval = [128]
2016-12-02 19:14:00,571 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  io.map.index.skip = [0]
2016-12-02 19:14:00,571 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  io.mapfile.bloom.error.rate = [0.005]
2016-12-02 19:14:00,571 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  io.mapfile.bloom.size = [1048576]
2016-12-02 19:14:00,571 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  io.native.lib.available = [true]
2016-12-02 19:14:00,571 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  io.seqfile.compress.blocksize = [1000000]
2016-12-02 19:14:00,571 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  io.seqfile.lazydecompress = [true]
2016-12-02 19:14:00,571 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  io.seqfile.local.dir = [/tmp/hadoop-hbase/io/local]
2016-12-02 19:14:00,571 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  io.seqfile.sorter.recordlimit = [1000000]
2016-12-02 19:14:00,571 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  io.serializations = [org.apache.hadoop.io.serializer.WritableSerialization]
2016-12-02 19:14:00,571 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  io.skip.checksum.errors = [false]
2016-12-02 19:14:00,571 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  io.storefile.bloom.block.size = [131072]
2016-12-02 19:14:00,571 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  io.storefile.bloom.error.rate = [0.005]
2016-12-02 19:14:00,571 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  ipc.client.connect.max.retries = [50]
2016-12-02 19:14:00,571 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  ipc.client.connect.max.retries.on.timeouts = [45]
2016-12-02 19:14:00,571 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  ipc.client.connect.retry.interval = [1000]
2016-12-02 19:14:00,571 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  ipc.client.connect.timeout = [20000]
2016-12-02 19:14:00,571 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  ipc.client.connection.maxidletime = [30000]
2016-12-02 19:14:00,571 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  ipc.client.fallback-to-simple-auth-allowed = [false]
2016-12-02 19:14:00,571 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  ipc.client.idlethreshold = [8000]
2016-12-02 19:14:00,571 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  ipc.client.kill.max = [10]
2016-12-02 19:14:00,571 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  ipc.server.listen.queue.size = [128]
2016-12-02 19:14:00,571 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  ipc.server.log.slow.rpc = [false]
2016-12-02 19:14:00,571 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  ipc.server.max.connections = [0]
2016-12-02 19:14:00,571 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  ipc.server.tcpnodelay = [true]
2016-12-02 19:14:00,571 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  ipcThreads = [100]
2016-12-02 19:14:00,571 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  kryoPoolSize = [1100]
2016-12-02 19:14:00,572 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  map.sort.class = [org.apache.hadoop.util.QuickSort]
2016-12-02 19:14:00,572 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapred.child.java.opts = [-Xmx200m]
2016-12-02 19:14:00,572 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.admin.map.child.java.opts = [-server -XX:NewRatio=8 -Djava.net.preferIPv4Stack=true -Dhdp.version=2.4.2.0-258]
2016-12-02 19:14:00,572 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.admin.reduce.child.java.opts = [-server -XX:NewRatio=8 -Djava.net.preferIPv4Stack=true -Dhdp.version=2.4.2.0-258]
2016-12-02 19:14:00,572 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.admin.user.env = [LD_LIBRARY_PATH=/usr/hdp/2.4.2.0-258/hadoop/lib/native:/usr/hdp/2.4.2.0-258/hadoop/lib/native/Linux-amd64-64]
2016-12-02 19:14:00,572 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.am.max-attempts = [2]
2016-12-02 19:14:00,572 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.app-submission.cross-platform = [false]
2016-12-02 19:14:00,572 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.application.classpath = [$PWD/mr-framework/hadoop/share/hadoop/mapreduce/*:$PWD/mr-framework/hadoop/share/hadoop/mapreduce/lib/*:$PWD/mr-framework/hadoop/share/hadoop/common/*:$PWD/mr-framework/hadoop/share/hadoop/common/lib/*:$PWD/mr-framework/hadoop/share/hadoop/yarn/*:$PWD/mr-framework/hadoop/share/hadoop/yarn/lib/*:$PWD/mr-framework/hadoop/share/hadoop/hdfs/*:$PWD/mr-framework/hadoop/share/hadoop/hdfs/lib/*:$PWD/mr-framework/hadoop/share/hadoop/tools/lib/*:/usr/hdp/2.4.2.0-258/hadoop/lib/hadoop-lzo-0.6.0.2.4.2.0-258.jar:/etc/hadoop/conf/secure]
2016-12-02 19:14:00,572 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.application.framework.path = [/hdp/apps/2.4.2.0-258/mapreduce/mapreduce.tar.gz#mr-framework]
2016-12-02 19:14:00,572 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.client.completion.pollinterval = [5000]
2016-12-02 19:14:00,572 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.client.output.filter = [FAILED]
2016-12-02 19:14:00,572 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.client.progressmonitor.pollinterval = [1000]
2016-12-02 19:14:00,572 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.client.submit.file.replication = [10]
2016-12-02 19:14:00,572 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.cluster.acls.enabled = [false]
2016-12-02 19:14:00,572 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.cluster.administrators = [ hadoop]
2016-12-02 19:14:00,572 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.cluster.local.dir = [/tmp/hadoop-hbase/mapred/local]
2016-12-02 19:14:00,572 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.cluster.temp.dir = [/tmp/hadoop-hbase/mapred/temp]
2016-12-02 19:14:00,572 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.fileoutputcommitter.algorithm.version = [1]
2016-12-02 19:14:00,572 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.framework.name = [yarn]
2016-12-02 19:14:00,572 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.ifile.readahead = [true]
2016-12-02 19:14:00,572 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.ifile.readahead.bytes = [4194304]
2016-12-02 19:14:00,572 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.input.fileinputformat.list-status.num-threads = [1]
2016-12-02 19:14:00,572 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.input.fileinputformat.split.minsize = [0]
2016-12-02 19:14:00,572 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.input.lineinputformat.linespermap = [1]
2016-12-02 19:14:00,572 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.job.acl-modify-job = [ ]
2016-12-02 19:14:00,572 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.job.acl-view-job = [ ]
2016-12-02 19:14:00,572 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.job.classloader = [false]
2016-12-02 19:14:00,572 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.job.committer.setup.cleanup.needed = [true]
2016-12-02 19:14:00,572 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.job.complete.cancel.delegation.tokens = [true]
2016-12-02 19:14:00,572 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.job.counters.max = [130]
2016-12-02 19:14:00,572 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.job.emit-timeline-data = [false]
2016-12-02 19:14:00,572 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.job.end-notification.max.attempts = [5]
2016-12-02 19:14:00,572 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.job.end-notification.max.retry.interval = [5000]
2016-12-02 19:14:00,572 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.job.end-notification.retry.attempts = [0]
2016-12-02 19:14:00,572 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.job.end-notification.retry.interval = [1000]
2016-12-02 19:14:00,572 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.job.hdfs-servers = [hdfs://hscale-dev1-nn:8020]
2016-12-02 19:14:00,572 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.job.jvm.numtasks = [1]
2016-12-02 19:14:00,572 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.job.map.output.collector.class = [org.apache.hadoop.mapred.MapTask$MapOutputBuffer]
2016-12-02 19:14:00,572 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.job.maps = [2]
2016-12-02 19:14:00,572 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.job.max.split.locations = [10]
2016-12-02 19:14:00,572 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.job.maxtaskfailures.per.tracker = [3]
2016-12-02 19:14:00,572 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.job.queuename = [default]
2016-12-02 19:14:00,572 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.job.reduce.shuffle.consumer.plugin.class = [org.apache.hadoop.mapreduce.task.reduce.Shuffle]
2016-12-02 19:14:00,572 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.job.reduce.slowstart.completedmaps = [0.05]
2016-12-02 19:14:00,572 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.job.reducer.preempt.delay.sec = [0]
2016-12-02 19:14:00,572 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.job.reducer.unconditional-preempt.delay.sec = [300]
2016-12-02 19:14:00,572 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.job.reduces = [1]
2016-12-02 19:14:00,572 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.job.running.map.limit = [0]
2016-12-02 19:14:00,572 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.job.running.reduce.limit = [0]
2016-12-02 19:14:00,572 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.job.speculative.minimum-allowed-tasks = [10]
2016-12-02 19:14:00,572 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.job.speculative.retry-after-no-speculate = [1000]
2016-12-02 19:14:00,572 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.job.speculative.retry-after-speculate = [15000]
2016-12-02 19:14:00,572 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.job.speculative.slowtaskthreshold = [1.0]
2016-12-02 19:14:00,572 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.job.speculative.speculative-cap-running-tasks = [0.1]
2016-12-02 19:14:00,572 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.job.speculative.speculative-cap-total-tasks = [0.01]
2016-12-02 19:14:00,572 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.job.split.metainfo.maxsize = [10000000]
2016-12-02 19:14:00,572 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.job.token.tracking.ids.enabled = [false]
2016-12-02 19:14:00,572 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.job.ubertask.enable = [false]
2016-12-02 19:14:00,572 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.job.ubertask.maxmaps = [9]
2016-12-02 19:14:00,573 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.job.ubertask.maxreduces = [1]
2016-12-02 19:14:00,573 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.job.userlog.retain.hours = [24]
2016-12-02 19:14:00,573 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.jobhistory.address = [hscale-dev1-nn:10020]
2016-12-02 19:14:00,573 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.jobhistory.admin.acl = [*]
2016-12-02 19:14:00,573 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.jobhistory.admin.address = [0.0.0.0:10033]
2016-12-02 19:14:00,573 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.jobhistory.bind-host = [0.0.0.0]
2016-12-02 19:14:00,573 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.jobhistory.cleaner.enable = [true]
2016-12-02 19:14:00,573 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.jobhistory.cleaner.interval-ms = [86400000]
2016-12-02 19:14:00,573 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.jobhistory.client.thread-count = [10]
2016-12-02 19:14:00,573 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.jobhistory.datestring.cache.size = [200000]
2016-12-02 19:14:00,573 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.jobhistory.done-dir = [/mr-history/done]
2016-12-02 19:14:00,573 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.jobhistory.http.policy = [HTTP_ONLY]
2016-12-02 19:14:00,573 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.jobhistory.intermediate-done-dir = [/mr-history/tmp]
2016-12-02 19:14:00,573 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.jobhistory.joblist.cache.size = [20000]
2016-12-02 19:14:00,573 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.jobhistory.keytab = [/etc/security/keytabs/jhs.service.keytab]
2016-12-02 19:14:00,573 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.jobhistory.loadedjobs.cache.size = [5]
2016-12-02 19:14:00,573 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.jobhistory.max-age-ms = [604800000]
2016-12-02 19:14:00,573 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.jobhistory.minicluster.fixed.ports = [false]
2016-12-02 19:14:00,573 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.jobhistory.move.interval-ms = [180000]
2016-12-02 19:14:00,573 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.jobhistory.move.thread-count = [3]
2016-12-02 19:14:00,573 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.jobhistory.principal = [jhs/_HOST@HSCALE.COM]
2016-12-02 19:14:00,573 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.jobhistory.recovery.enable = [true]
2016-12-02 19:14:00,573 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.jobhistory.recovery.store.class = [org.apache.hadoop.mapreduce.v2.hs.HistoryServerLeveldbStateStoreService]
2016-12-02 19:14:00,573 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.jobhistory.recovery.store.fs.uri = [/tmp/hadoop-hbase/mapred/history/recoverystore]
2016-12-02 19:14:00,573 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.jobhistory.recovery.store.leveldb.path = [/hadoop/mapreduce/jhs]
2016-12-02 19:14:00,573 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.jobhistory.webapp.address = [hscale-dev1-nn:19888]
2016-12-02 19:14:00,573 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.jobhistory.webapp.rest-csrf.custom-header = [X-XSRF-Header]
2016-12-02 19:14:00,573 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.jobhistory.webapp.rest-csrf.enabled = [false]
2016-12-02 19:14:00,573 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.jobhistory.webapp.rest-csrf.methods-to-ignore = [GET,OPTIONS,HEAD]
2016-12-02 19:14:00,573 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.jobhistory.webapp.spnego-keytab-file = [/etc/security/keytabs/spnego.service.keytab]
2016-12-02 19:14:00,573 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.jobhistory.webapp.spnego-principal = [HTTP/_HOST@HSCALE.COM]
2016-12-02 19:14:00,573 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.jobtracker.address = [local]
2016-12-02 19:14:00,573 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.jobtracker.expire.trackers.interval = [600000]
2016-12-02 19:14:00,573 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.jobtracker.handler.count = [10]
2016-12-02 19:14:00,573 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.jobtracker.heartbeats.in.second = [100]
2016-12-02 19:14:00,573 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.jobtracker.http.address = [0.0.0.0:50030]
2016-12-02 19:14:00,573 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.jobtracker.instrumentation = [org.apache.hadoop.mapred.JobTrackerMetricsInst]
2016-12-02 19:14:00,573 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.jobtracker.jobhistory.block.size = [3145728]
2016-12-02 19:14:00,573 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.jobtracker.jobhistory.lru.cache.size = [5]
2016-12-02 19:14:00,573 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.jobtracker.jobhistory.task.numberprogresssplits = [12]
2016-12-02 19:14:00,573 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.jobtracker.maxtasks.perjob = [-1]
2016-12-02 19:14:00,573 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.jobtracker.persist.jobstatus.active = [true]
2016-12-02 19:14:00,573 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.jobtracker.persist.jobstatus.dir = [/jobtracker/jobsInfo]
2016-12-02 19:14:00,573 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.jobtracker.persist.jobstatus.hours = [1]
2016-12-02 19:14:00,573 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.jobtracker.restart.recover = [false]
2016-12-02 19:14:00,573 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.jobtracker.retiredjobs.cache.size = [1000]
2016-12-02 19:14:00,573 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.jobtracker.staging.root.dir = [/tmp/hadoop-hbase/mapred/staging]
2016-12-02 19:14:00,573 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.jobtracker.system.dir = [/tmp/hadoop-hbase/mapred/system]
2016-12-02 19:14:00,573 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.jobtracker.taskcache.levels = [2]
2016-12-02 19:14:00,573 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.jobtracker.taskscheduler = [org.apache.hadoop.mapred.JobQueueTaskScheduler]
2016-12-02 19:14:00,573 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.jobtracker.tasktracker.maxblacklists = [4]
2016-12-02 19:14:00,573 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.jobtracker.webinterface.trusted = [false]
2016-12-02 19:14:00,573 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.local.clientfactory.class.name = [org.apache.hadoop.mapred.LocalClientFactory]
2016-12-02 19:14:00,573 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.map.cpu.vcores = [1]
2016-12-02 19:14:00,573 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.map.java.opts = [-Xmx1228m]
2016-12-02 19:14:00,573 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.map.log.level = [INFO]
2016-12-02 19:14:00,573 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.map.maxattempts = [4]
2016-12-02 19:14:00,573 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.map.memory.mb = [1536]
2016-12-02 19:14:00,573 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.map.output.compress = [false]
2016-12-02 19:14:00,573 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.map.output.compress.codec = [org.apache.hadoop.io.compress.DefaultCodec]
2016-12-02 19:14:00,573 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.map.skip.maxrecords = [0]
2016-12-02 19:14:00,573 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.map.skip.proc.count.autoincr = [true]
2016-12-02 19:14:00,573 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.map.sort.spill.percent = [0.7]
2016-12-02 19:14:00,573 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.map.speculative = [false]
2016-12-02 19:14:00,573 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.output.fileoutputformat.compress = [false]
2016-12-02 19:14:00,574 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.output.fileoutputformat.compress.codec = [org.apache.hadoop.io.compress.DefaultCodec]
2016-12-02 19:14:00,574 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.output.fileoutputformat.compress.type = [BLOCK]
2016-12-02 19:14:00,574 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.reduce.cpu.vcores = [1]
2016-12-02 19:14:00,574 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.reduce.input.buffer.percent = [0.0]
2016-12-02 19:14:00,574 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.reduce.java.opts = [-Xmx1228m]
2016-12-02 19:14:00,574 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.reduce.log.level = [INFO]
2016-12-02 19:14:00,574 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.reduce.markreset.buffer.percent = [0.0]
2016-12-02 19:14:00,574 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.reduce.maxattempts = [4]
2016-12-02 19:14:00,574 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.reduce.memory.mb = [1536]
2016-12-02 19:14:00,574 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.reduce.merge.inmem.threshold = [1000]
2016-12-02 19:14:00,574 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.reduce.shuffle.connect.timeout = [180000]
2016-12-02 19:14:00,574 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.reduce.shuffle.fetch.retry.enabled = [1]
2016-12-02 19:14:00,574 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.reduce.shuffle.fetch.retry.interval-ms = [1000]
2016-12-02 19:14:00,574 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.reduce.shuffle.fetch.retry.timeout-ms = [30000]
2016-12-02 19:14:00,574 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.reduce.shuffle.input.buffer.percent = [0.7]
2016-12-02 19:14:00,574 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.reduce.shuffle.memory.limit.percent = [0.25]
2016-12-02 19:14:00,574 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.reduce.shuffle.merge.percent = [0.66]
2016-12-02 19:14:00,574 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.reduce.shuffle.parallelcopies = [30]
2016-12-02 19:14:00,574 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.reduce.shuffle.read.timeout = [180000]
2016-12-02 19:14:00,574 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.reduce.shuffle.retry-delay.max.ms = [60000]
2016-12-02 19:14:00,574 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.reduce.skip.maxgroups = [0]
2016-12-02 19:14:00,574 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.reduce.skip.proc.count.autoincr = [true]
2016-12-02 19:14:00,574 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.reduce.speculative = [false]
2016-12-02 19:14:00,574 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.shuffle.connection-keep-alive.enable = [false]
2016-12-02 19:14:00,574 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.shuffle.connection-keep-alive.timeout = [5]
2016-12-02 19:14:00,574 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.shuffle.max.connections = [0]
2016-12-02 19:14:00,574 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.shuffle.max.threads = [0]
2016-12-02 19:14:00,574 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.shuffle.port = [13562]
2016-12-02 19:14:00,574 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.shuffle.ssl.enabled = [false]
2016-12-02 19:14:00,574 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.shuffle.ssl.file.buffer.size = [65536]
2016-12-02 19:14:00,574 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.shuffle.transfer.buffer.size = [131072]
2016-12-02 19:14:00,574 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.task.combine.progress.records = [10000]
2016-12-02 19:14:00,574 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.task.files.preserve.failedtasks = [false]
2016-12-02 19:14:00,574 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.task.io.sort.factor = [100]
2016-12-02 19:14:00,574 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.task.io.sort.mb = [859]
2016-12-02 19:14:00,574 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.task.merge.progress.records = [10000]
2016-12-02 19:14:00,574 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.task.profile = [false]
2016-12-02 19:14:00,574 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.task.profile.map.params = [-agentlib:hprof=cpu=samples,heap=sites,force=n,thread=y,verbose=n,file=%s]
2016-12-02 19:14:00,574 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.task.profile.maps = [0-2]
2016-12-02 19:14:00,574 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.task.profile.params = [-agentlib:hprof=cpu=samples,heap=sites,force=n,thread=y,verbose=n,file=%s]
2016-12-02 19:14:00,574 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.task.profile.reduce.params = [-agentlib:hprof=cpu=samples,heap=sites,force=n,thread=y,verbose=n,file=%s]
2016-12-02 19:14:00,574 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.task.profile.reduces = [0-2]
2016-12-02 19:14:00,574 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.task.skip.start.attempts = [2]
2016-12-02 19:14:00,574 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.task.timeout = [300000]
2016-12-02 19:14:00,574 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.task.userlog.limit.kb = [0]
2016-12-02 19:14:00,574 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.tasktracker.dns.interface = [default]
2016-12-02 19:14:00,574 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.tasktracker.dns.nameserver = [default]
2016-12-02 19:14:00,574 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.tasktracker.healthchecker.interval = [60000]
2016-12-02 19:14:00,574 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.tasktracker.healthchecker.script.timeout = [600000]
2016-12-02 19:14:00,574 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.tasktracker.http.address = [0.0.0.0:50060]
2016-12-02 19:14:00,574 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.tasktracker.http.threads = [40]
2016-12-02 19:14:00,574 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.tasktracker.indexcache.mb = [10]
2016-12-02 19:14:00,574 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.tasktracker.instrumentation = [org.apache.hadoop.mapred.TaskTrackerMetricsInst]
2016-12-02 19:14:00,574 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.tasktracker.local.dir.minspacekill = [0]
2016-12-02 19:14:00,574 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.tasktracker.local.dir.minspacestart = [0]
2016-12-02 19:14:00,574 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.tasktracker.map.tasks.maximum = [2]
2016-12-02 19:14:00,574 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.tasktracker.outofband.heartbeat = [false]
2016-12-02 19:14:00,574 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.tasktracker.reduce.tasks.maximum = [2]
2016-12-02 19:14:00,574 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.tasktracker.report.address = [127.0.0.1:0]
2016-12-02 19:14:00,574 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.tasktracker.taskcontroller = [org.apache.hadoop.mapred.DefaultTaskController]
2016-12-02 19:14:00,574 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.tasktracker.taskmemorymanager.monitoringinterval = [5000]
2016-12-02 19:14:00,574 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.tasktracker.tasks.sleeptimebeforesigkill = [5000]
2016-12-02 19:14:00,574 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  maxBufferEntries = [1000]
2016-12-02 19:14:00,574 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  maxBufferHeapSize = [3145728]
2016-12-02 19:14:00,574 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  maxDdlWait = [60000]
2016-12-02 19:14:00,574 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  maxDependentWrites = [60000]
2016-12-02 19:14:00,574 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  maxIndependentWrites = [60000]
2016-12-02 19:14:00,574 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  maxRetries = [5]
2016-12-02 19:14:00,574 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  maxWriterThreads = [5]
2016-12-02 19:14:00,574 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  namespace = [splice]
2016-12-02 19:14:00,574 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  nestedLoopJoinBatchSize = [10]
2016-12-02 19:14:00,574 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  net.topology.impl = [org.apache.hadoop.net.NetworkTopology]
2016-12-02 19:14:00,575 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  net.topology.node.switch.mapping.impl = [org.apache.hadoop.net.ScriptBasedMapping]
2016-12-02 19:14:00,575 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  net.topology.script.file.name = [/etc/hadoop/conf/topology_script.py]
2016-12-02 19:14:00,575 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  net.topology.script.number.args = [100]
2016-12-02 19:14:00,575 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  networkBindAddress = [0.0.0.0]
2016-12-02 19:14:00,575 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  networkBindPort = [1527]
2016-12-02 19:14:00,575 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  nfs.allow.insecure.ports = [true]
2016-12-02 19:14:00,575 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  nfs.dump.dir = [/tmp/.hdfs-nfs]
2016-12-02 19:14:00,575 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  nfs.exports.allowed.hosts = [* rw]
2016-12-02 19:14:00,575 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  nfs.file.dump.dir = [/tmp/.hdfs-nfs]
2016-12-02 19:14:00,575 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  nfs.mountd.port = [4242]
2016-12-02 19:14:00,575 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  nfs.rtmax = [1048576]
2016-12-02 19:14:00,575 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  nfs.server.port = [2049]
2016-12-02 19:14:00,575 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  nfs.wtmax = [1048576]
2016-12-02 19:14:00,575 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  olapClientTickTime = [1000]
2016-12-02 19:14:00,575 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  olapClientWaitTime = [900000]
2016-12-02 19:14:00,575 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  olapServerBindPort = [60014]
2016-12-02 19:14:00,575 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  olapServerThreads = [16]
2016-12-02 19:14:00,575 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  olapServerTickLimit = [120]
2016-12-02 19:14:00,575 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  optimizerExtraQualifierMultiplier = [0.9]
2016-12-02 19:14:00,575 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  optimizerPlanMaximumTimeout = [9223372036854775807]
2016-12-02 19:14:00,575 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  optimizerPlanMinimumTimeout = [0]
2016-12-02 19:14:00,575 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  partitionCacheExpiration = [60000]
2016-12-02 19:14:00,575 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  partitionserverJmxPort = [10102]
2016-12-02 19:14:00,575 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  partitionserverPort = [16020]
2016-12-02 19:14:00,575 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  phoenix.connection.autoCommit = [true]
2016-12-02 19:14:00,575 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  phoenix.functions.allowUserDefinedFunctions = [ ]
2016-12-02 19:14:00,575 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  phoenix.query.timeoutMs = [60000]
2016-12-02 19:14:00,575 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  pipelineKryoPoolSize = [1024]
2016-12-02 19:14:00,575 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  readResolverQueueSize = [-1]
2016-12-02 19:14:00,575 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  readResolverThreads = [4]
2016-12-02 19:14:00,575 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  regionLoadUpdateInterval = [5]
2016-12-02 19:14:00,575 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  regionMaxFileSize = [10737418240]
2016-12-02 19:14:00,575 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  regionServerHandlerCount = [100]
2016-12-02 19:14:00,575 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  reservedSlotsTimeout = [60]
2016-12-02 19:14:00,575 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  rpc.metrics.quantile.enable = [false]
2016-12-02 19:14:00,575 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  s3.blocksize = [67108864]
2016-12-02 19:14:00,575 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  s3.bytes-per-checksum = [512]
2016-12-02 19:14:00,575 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  s3.client-write-packet-size = [65536]
2016-12-02 19:14:00,575 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  s3.replication = [3]
2016-12-02 19:14:00,575 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  s3.stream-buffer-size = [4096]
2016-12-02 19:14:00,575 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  s3native.blocksize = [67108864]
2016-12-02 19:14:00,575 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  s3native.bytes-per-checksum = [512]
2016-12-02 19:14:00,575 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  s3native.client-write-packet-size = [65536]
2016-12-02 19:14:00,575 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  s3native.replication = [3]
2016-12-02 19:14:00,575 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  s3native.stream-buffer-size = [4096]
2016-12-02 19:14:00,575 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  sequenceBlockSize = [1000]
2016-12-02 19:14:00,575 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  sparkIoCompressionCodec = [lz4]
2016-12-02 19:14:00,575 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  sparkResultStreamingBatchSize = [1024]
2016-12-02 19:14:00,575 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  sparkResultStreamingBatches = [10]
2016-12-02 19:14:00,575 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  splice.authentication = [NATIVE]
2016-12-02 19:14:00,575 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  splice.authentication.native.algorithm = [SHA-512]
2016-12-02 19:14:00,575 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  splice.authentication.native.create.credentials.database = [true]
2016-12-02 19:14:00,575 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  splice.client.numConnections = [1]
2016-12-02 19:14:00,575 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  splice.client.write.maxDependentWrites = [60000]
2016-12-02 19:14:00,575 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  splice.client.write.maxIndependentWrites = [60000]
2016-12-02 19:14:00,575 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  splice.compression = [snappy]
2016-12-02 19:14:00,575 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  splice.marshal.kryoPoolSize = [1100]
2016-12-02 19:14:00,575 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  splice.olap_server.clientWaitTime = [900000]
2016-12-02 19:14:00,575 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  splice.ring.bufferSize = [131072]
2016-12-02 19:14:00,575 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  splice.splitBlockSize = [67108864]
2016-12-02 19:14:00,575 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  splice.timestamp_server.clientWaitTime = [120000]
2016-12-02 19:14:00,575 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  splice.txn.activeTxns.cacheSize = [10240]
2016-12-02 19:14:00,575 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  splice.txn.completedTxns.concurrency = [128]
2016-12-02 19:14:00,575 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  splice.txn.concurrencyLevel = [4096]
2016-12-02 19:14:00,575 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  spliceRootPath = [/splice]
2016-12-02 19:14:00,575 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  splitBlockSize = [67108864]
2016-12-02 19:14:00,575 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  startupLockWaitPeriod = [1000]
2016-12-02 19:14:00,575 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  storageFactoryHome = [hdfs://hscale-dev1-nn:8020/apps/hbase/data]
2016-12-02 19:14:00,576 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  tableSplitSleepInterval = [500]
2016-12-02 19:14:00,576 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  tfile.fs.input.buffer.size = [262144]
2016-12-02 19:14:00,576 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  tfile.fs.output.buffer.size = [262144]
2016-12-02 19:14:00,576 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  tfile.io.chunk.size = [1048576]
2016-12-02 19:14:00,576 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  threadKeepaliveTime = [60]
2016-12-02 19:14:00,576 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  timestampBlockSize = [8192]
2016-12-02 19:14:00,576 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  timestampClientWaitTime = [120000]
2016-12-02 19:14:00,576 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  timestampServerBindPort = [60012]
2016-12-02 19:14:00,576 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  topkSize = [10]
2016-12-02 19:14:00,576 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  transactionKeepAliveInterval = [15000]
2016-12-02 19:14:00,576 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  transactionKeepAliveThreads = [4]
2016-12-02 19:14:00,576 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  transactionLockStripes = [100]
2016-12-02 19:14:00,576 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  transactionTimeout = [150000]
2016-12-02 19:14:00,576 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  upgradeForced = [false]
2016-12-02 19:14:00,576 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  upgradeForcedFrom = [null]
2016-12-02 19:14:00,576 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  writeMaxFlushesPerRegion = [5]
2016-12-02 19:14:00,576 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.acl.enable = [false]
2016-12-02 19:14:00,576 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.admin.acl = [yarn]
2016-12-02 19:14:00,576 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.am.liveness-monitor.expiry-interval-ms = [600000]
2016-12-02 19:14:00,576 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.app.mapreduce.am.admin-command-opts = [-Dhdp.version=2.4.2.0-258]
2016-12-02 19:14:00,576 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.app.mapreduce.am.command-opts = [-Xmx410m]
2016-12-02 19:14:00,576 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.app.mapreduce.am.container.log.backups = [0]
2016-12-02 19:14:00,576 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.app.mapreduce.am.container.log.limit.kb = [0]
2016-12-02 19:14:00,576 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.app.mapreduce.am.containerlauncher.threadpool-initial-size = [10]
2016-12-02 19:14:00,576 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.app.mapreduce.am.hard-kill-timeout-ms = [10000]
2016-12-02 19:14:00,576 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.app.mapreduce.am.job.committer.cancel-timeout = [60000]
2016-12-02 19:14:00,576 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.app.mapreduce.am.job.committer.commit-window = [10000]
2016-12-02 19:14:00,576 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.app.mapreduce.am.job.task.listener.thread-count = [30]
2016-12-02 19:14:00,576 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.app.mapreduce.am.log.level = [INFO]
2016-12-02 19:14:00,576 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.app.mapreduce.am.resource.cpu-vcores = [1]
2016-12-02 19:14:00,576 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.app.mapreduce.am.resource.mb = [512]
2016-12-02 19:14:00,576 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.app.mapreduce.am.scheduler.heartbeat.interval-ms = [1000]
2016-12-02 19:14:00,576 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.app.mapreduce.am.staging-dir = [/user]
2016-12-02 19:14:00,576 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.app.mapreduce.client-am.ipc.max-retries = [3]
2016-12-02 19:14:00,576 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.app.mapreduce.client-am.ipc.max-retries-on-timeouts = [3]
2016-12-02 19:14:00,576 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.app.mapreduce.client.job.max-retries = [0]
2016-12-02 19:14:00,576 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.app.mapreduce.client.job.retry-interval = [2000]
2016-12-02 19:14:00,576 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.app.mapreduce.client.max-retries = [3]
2016-12-02 19:14:00,576 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.app.mapreduce.shuffle.log.backups = [0]
2016-12-02 19:14:00,576 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.app.mapreduce.shuffle.log.limit.kb = [0]
2016-12-02 19:14:00,576 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.app.mapreduce.shuffle.log.separate = [true]
2016-12-02 19:14:00,576 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.app.mapreduce.task.container.log.backups = [0]
2016-12-02 19:14:00,576 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.application.classpath = [$HADOOP_CONF_DIR, /usr/hdp/current/hadoop-client/*, /usr/hdp/current/hadoop-client/lib/*, /usr/hdp/current/hadoop-hdfs-client/*, /usr/hdp/current/hadoop-hdfs-client/lib/*, /usr/hdp/current/hadoop-yarn-client/*, /usr/hdp/current/hadoop-yarn-client/lib/*, /usr/hdp/current/hadoop-mapreduce-client/*, /usr/hdp/current/hadoop-mapreduce-client/lib/*, /usr/hdp/current/hbase-regionserver/*, /usr/hdp/current/hbase-regionserver/lib/*, /opt/splice/default/lib/*]
2016-12-02 19:14:00,576 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.client.application-client-protocol.poll-interval-ms = [200]
2016-12-02 19:14:00,576 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.client.failover-proxy-provider = [org.apache.hadoop.yarn.client.ConfiguredRMFailoverProxyProvider]
2016-12-02 19:14:00,576 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.client.failover-retries = [0]
2016-12-02 19:14:00,576 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.client.failover-retries-on-socket-timeouts = [0]
2016-12-02 19:14:00,576 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.client.max-cached-nodemanagers-proxies = [0]
2016-12-02 19:14:00,576 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.client.nodemanager-client-async.thread-pool-max-size = [500]
2016-12-02 19:14:00,576 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.client.nodemanager-connect.max-wait-ms = [60000]
2016-12-02 19:14:00,576 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.client.nodemanager-connect.retry-interval-ms = [10000]
2016-12-02 19:14:00,576 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.dispatcher.drain-events.timeout = [300000]
2016-12-02 19:14:00,576 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.fail-fast = [false]
2016-12-02 19:14:00,576 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.http.policy = [HTTP_ONLY]
2016-12-02 19:14:00,576 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.ipc.rpc.class = [org.apache.hadoop.yarn.ipc.HadoopYarnProtoRPC]
2016-12-02 19:14:00,576 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.log-aggregation-enable = [true]
2016-12-02 19:14:00,576 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.log-aggregation.retain-check-interval-seconds = [-1]
2016-12-02 19:14:00,576 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.log-aggregation.retain-seconds = [2592000]
2016-12-02 19:14:00,576 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.log.server.url = [http://hscale-dev1-nn:19888/jobhistory/logs]
2016-12-02 19:14:00,576 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.nm.liveness-monitor.expiry-interval-ms = [600000]
2016-12-02 19:14:00,576 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.node-labels.enabled = [false]
2016-12-02 19:14:00,576 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.node-labels.fs-store.impl.class = [org.apache.hadoop.yarn.nodelabels.FileSystemNodeLabelsStore]
2016-12-02 19:14:00,576 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.node-labels.fs-store.retry-policy-spec = [2000, 500]
2016-12-02 19:14:00,576 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.node-labels.fs-store.root-dir = [/system/yarn/node-labels]
2016-12-02 19:14:00,576 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.nodemanager.address = [0.0.0.0:45454]
2016-12-02 19:14:00,576 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.nodemanager.admin-env = [MALLOC_ARENA_MAX=$MALLOC_ARENA_MAX]
2016-12-02 19:14:00,576 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.nodemanager.aux-services = [mapreduce_shuffle,spark_shuffle]
2016-12-02 19:14:00,576 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.nodemanager.aux-services.mapreduce_shuffle.class = [org.apache.hadoop.mapred.ShuffleHandler]
2016-12-02 19:14:00,576 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.nodemanager.aux-services.spark_shuffle.class = [org.apache.spark.network.yarn.YarnShuffleService]
2016-12-02 19:14:00,576 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.nodemanager.bind-host = [0.0.0.0]
2016-12-02 19:14:00,576 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.nodemanager.container-executor.class = [org.apache.hadoop.yarn.server.nodemanager.DefaultContainerExecutor]
2016-12-02 19:14:00,576 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.nodemanager.container-manager.thread-count = [20]
2016-12-02 19:14:00,576 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.nodemanager.container-metrics.unregister-delay-ms = [10000]
2016-12-02 19:14:00,576 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.nodemanager.container-monitor.interval-ms = [3000]
2016-12-02 19:14:00,576 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.nodemanager.container-monitor.procfs-tree.smaps-based-rss.enabled = [false]
2016-12-02 19:14:00,576 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.nodemanager.delete.debug-delay-sec = [86400]
2016-12-02 19:14:00,576 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.nodemanager.delete.thread-count = [4]
2016-12-02 19:14:00,576 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.nodemanager.disk-health-checker.interval-ms = [120000]
2016-12-02 19:14:00,576 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.nodemanager.disk-health-checker.max-disk-utilization-per-disk-percentage = [90]
2016-12-02 19:14:00,576 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.nodemanager.disk-health-checker.min-free-space-per-disk-mb = [1000]
2016-12-02 19:14:00,576 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.nodemanager.disk-health-checker.min-healthy-disks = [0.25]
2016-12-02 19:14:00,577 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.nodemanager.docker-container-executor.exec-name = [/usr/bin/docker]
2016-12-02 19:14:00,577 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.nodemanager.env-whitelist = [JAVA_HOME,HADOOP_COMMON_HOME,HADOOP_HDFS_HOME,HADOOP_CONF_DIR,HADOOP_YARN_HOME]
2016-12-02 19:14:00,577 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.nodemanager.health-checker.interval-ms = [135000]
2016-12-02 19:14:00,577 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.nodemanager.health-checker.script.timeout-ms = [60000]
2016-12-02 19:14:00,577 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.nodemanager.hostname = [0.0.0.0]
2016-12-02 19:14:00,577 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.nodemanager.keytab = [/etc/krb5.keytab]
2016-12-02 19:14:00,577 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.nodemanager.linux-container-executor.cgroups.hierarchy = [hadoop-yarn]
2016-12-02 19:14:00,577 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.nodemanager.linux-container-executor.cgroups.mount = [false]
2016-12-02 19:14:00,577 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.nodemanager.linux-container-executor.cgroups.strict-resource-usage = [false]
2016-12-02 19:14:00,577 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.nodemanager.linux-container-executor.group = [hadoop]
2016-12-02 19:14:00,577 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.nodemanager.linux-container-executor.nonsecure-mode.limit-users = [true]
2016-12-02 19:14:00,577 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.nodemanager.linux-container-executor.nonsecure-mode.local-user = [nobody]
2016-12-02 19:14:00,577 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.nodemanager.linux-container-executor.nonsecure-mode.user-pattern = [^[_.A-Za-z0-9][-@_.A-Za-z0-9]{0,255}?[$]?$]
2016-12-02 19:14:00,577 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.nodemanager.linux-container-executor.resources-handler.class = [org.apache.hadoop.yarn.server.nodemanager.util.DefaultLCEResourcesHandler]
2016-12-02 19:14:00,577 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.nodemanager.local-cache.max-files-per-directory = [8192]
2016-12-02 19:14:00,577 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.nodemanager.local-dirs = [/hadoop/yarn/local]
2016-12-02 19:14:00,577 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.nodemanager.localizer.address = [0.0.0.0:8040]
2016-12-02 19:14:00,577 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.nodemanager.localizer.cache.cleanup.interval-ms = [600000]
2016-12-02 19:14:00,577 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.nodemanager.localizer.cache.target-size-mb = [10240]
2016-12-02 19:14:00,577 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.nodemanager.localizer.client.thread-count = [5]
2016-12-02 19:14:00,577 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.nodemanager.localizer.fetch.thread-count = [4]
2016-12-02 19:14:00,577 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.nodemanager.log-aggregation.compression-type = [gz]
2016-12-02 19:14:00,577 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.nodemanager.log-aggregation.debug-enabled = [false]
2016-12-02 19:14:00,577 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.nodemanager.log-aggregation.num-log-files-per-app = [30]
2016-12-02 19:14:00,577 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.nodemanager.log-aggregation.roll-monitoring-interval-seconds = [-1]
2016-12-02 19:14:00,577 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.nodemanager.log-container-debug-info.enabled = [true]
2016-12-02 19:14:00,577 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.nodemanager.log-dirs = [/hadoop/yarn/log]
2016-12-02 19:14:00,577 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.nodemanager.log.retain-second = [604800]
2016-12-02 19:14:00,577 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.nodemanager.log.retain-seconds = [10800]
2016-12-02 19:14:00,577 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.nodemanager.pmem-check-enabled = [true]
2016-12-02 19:14:00,577 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.nodemanager.process-kill-wait.ms = [2000]
2016-12-02 19:14:00,577 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.nodemanager.recovery.dir = [/var/log/hadoop-yarn/nodemanager/recovery-state]
2016-12-02 19:14:00,577 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.nodemanager.recovery.enabled = [true]
2016-12-02 19:14:00,577 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.nodemanager.remote-app-log-dir = [/app-logs]
2016-12-02 19:14:00,577 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.nodemanager.remote-app-log-dir-suffix = [logs]
2016-12-02 19:14:00,577 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.nodemanager.resource.cpu-vcores = [8]
2016-12-02 19:14:00,577 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.nodemanager.resource.memory-mb = [5120]
2016-12-02 19:14:00,577 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.nodemanager.resource.percentage-physical-cpu-limit = [80]
2016-12-02 19:14:00,577 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.nodemanager.resourcemanager.minimum.version = [NONE]
2016-12-02 19:14:00,577 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.nodemanager.sleep-delay-before-sigkill.ms = [250]
2016-12-02 19:14:00,577 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.nodemanager.vmem-check-enabled = [false]
2016-12-02 19:14:00,577 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.nodemanager.vmem-pmem-ratio = [2.1]
2016-12-02 19:14:00,577 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.nodemanager.webapp.address = [0.0.0.0:8042]
2016-12-02 19:14:00,577 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.nodemanager.webapp.cross-origin.enabled = [false]
2016-12-02 19:14:00,577 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.nodemanager.webapp.rest-csrf.custom-header = [X-XSRF-Header]
2016-12-02 19:14:00,577 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.nodemanager.webapp.rest-csrf.enabled = [false]
2016-12-02 19:14:00,577 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.nodemanager.webapp.rest-csrf.methods-to-ignore = [GET,OPTIONS,HEAD]
2016-12-02 19:14:00,577 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.nodemanager.windows-container.cpu-limit.enabled = [false]
2016-12-02 19:14:00,577 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.nodemanager.windows-container.memory-limit.enabled = [false]
2016-12-02 19:14:00,577 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.resourcemanager.address = [hscale-dev1-nn:8050]
2016-12-02 19:14:00,577 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.resourcemanager.admin.address = [hscale-dev1-nn:8141]
2016-12-02 19:14:00,577 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.resourcemanager.admin.client.thread-count = [1]
2016-12-02 19:14:00,577 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.resourcemanager.am-rm-tokens.master-key-rolling-interval-secs = [86400]
2016-12-02 19:14:00,577 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.resourcemanager.am.max-attempts = [2]
2016-12-02 19:14:00,577 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.resourcemanager.amlauncher.thread-count = [50]
2016-12-02 19:14:00,577 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.resourcemanager.bind-host = [0.0.0.0]
2016-12-02 19:14:00,577 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.resourcemanager.client.thread-count = [50]
2016-12-02 19:14:00,577 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.resourcemanager.configuration.provider-class = [org.apache.hadoop.yarn.LocalConfigurationProvider]
2016-12-02 19:14:00,577 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.resourcemanager.connect.max-wait.ms = [900000]
2016-12-02 19:14:00,577 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.resourcemanager.connect.retry-interval.ms = [30000]
2016-12-02 19:14:00,577 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.resourcemanager.container-tokens.master-key-rolling-interval-secs = [86400]
2016-12-02 19:14:00,577 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.resourcemanager.container.liveness-monitor.interval-ms = [600000]
2016-12-02 19:14:00,577 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.resourcemanager.delayed.delegation-token.removal-interval-ms = [30000]
2016-12-02 19:14:00,577 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.resourcemanager.fail-fast = [false]
2016-12-02 19:14:00,577 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.resourcemanager.fs.state-store.num-retries = [0]
2016-12-02 19:14:00,577 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.resourcemanager.fs.state-store.retry-interval-ms = [1000]
2016-12-02 19:14:00,577 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.resourcemanager.fs.state-store.retry-policy-spec = [2000, 500]
2016-12-02 19:14:00,577 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.resourcemanager.fs.state-store.uri = [ ]
2016-12-02 19:14:00,577 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.resourcemanager.ha.automatic-failover.embedded = [true]
2016-12-02 19:14:00,577 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.resourcemanager.ha.automatic-failover.enabled = [true]
2016-12-02 19:14:00,577 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.resourcemanager.ha.automatic-failover.zk-base-path = [/yarn-leader-election]
2016-12-02 19:14:00,577 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.resourcemanager.ha.enabled = [false]
2016-12-02 19:14:00,577 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.resourcemanager.hostname = [hscale-dev1-nn]
2016-12-02 19:14:00,577 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.resourcemanager.keytab = [/etc/krb5.keytab]
2016-12-02 19:14:00,577 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.resourcemanager.leveldb-state-store.path = [/tmp/hadoop-hbase/yarn/system/rmstore]
2016-12-02 19:14:00,577 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.resourcemanager.max-completed-applications = [10000]
2016-12-02 19:14:00,577 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.resourcemanager.max-log-aggregation-diagnostics-in-memory = [10]
2016-12-02 19:14:00,577 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.resourcemanager.nodemanager-connect-retries = [10]
2016-12-02 19:14:00,577 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.resourcemanager.nodemanager.minimum.version = [NONE]
2016-12-02 19:14:00,577 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.resourcemanager.nodemanagers.heartbeat-interval-ms = [1000]
2016-12-02 19:14:00,577 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.resourcemanager.nodes.exclude-path = [/etc/hadoop/conf/yarn.exclude]
2016-12-02 19:14:00,577 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.resourcemanager.proxy-user-privileges.enabled = [false]
2016-12-02 19:14:00,577 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.resourcemanager.recovery.enabled = [true]
2016-12-02 19:14:00,577 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.resourcemanager.resource-tracker.address = [hscale-dev1-nn:8025]
2016-12-02 19:14:00,577 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.resourcemanager.resource-tracker.client.thread-count = [50]
2016-12-02 19:14:00,577 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.resourcemanager.scheduler.address = [hscale-dev1-nn:8030]
2016-12-02 19:14:00,577 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.resourcemanager.scheduler.class = [org.apache.hadoop.yarn.server.resourcemanager.scheduler.capacity.CapacityScheduler]
2016-12-02 19:14:00,578 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.resourcemanager.scheduler.client.thread-count = [50]
2016-12-02 19:14:00,578 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.resourcemanager.scheduler.monitor.enable = [false]
2016-12-02 19:14:00,578 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.resourcemanager.scheduler.monitor.policies = [org.apache.hadoop.yarn.server.resourcemanager.monitor.capacity.ProportionalCapacityPreemptionPolicy]
2016-12-02 19:14:00,578 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.resourcemanager.state-store.max-completed-applications = [10000]
2016-12-02 19:14:00,578 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.resourcemanager.store.class = [org.apache.hadoop.yarn.server.resourcemanager.recovery.ZKRMStateStore]
2016-12-02 19:14:00,578 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.resourcemanager.system-metrics-publisher.dispatcher.pool-size = [10]
2016-12-02 19:14:00,578 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.resourcemanager.system-metrics-publisher.enabled = [true]
2016-12-02 19:14:00,578 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.resourcemanager.webapp.address = [hscale-dev1-nn:8088]
2016-12-02 19:14:00,578 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.resourcemanager.webapp.cross-origin.enabled = [true]
2016-12-02 19:14:00,578 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.resourcemanager.webapp.delegation-token-auth-filter.enabled = [false]
2016-12-02 19:14:00,578 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.resourcemanager.webapp.https.address = [hscale-dev1-nn:8090]
2016-12-02 19:14:00,578 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.resourcemanager.webapp.rest-csrf.custom-header = [X-XSRF-Header]
2016-12-02 19:14:00,578 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.resourcemanager.webapp.rest-csrf.enabled = [false]
2016-12-02 19:14:00,578 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.resourcemanager.webapp.rest-csrf.methods-to-ignore = [GET,OPTIONS,HEAD]
2016-12-02 19:14:00,578 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.resourcemanager.work-preserving-recovery.enabled = [true]
2016-12-02 19:14:00,578 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.resourcemanager.work-preserving-recovery.scheduling-wait-ms = [10000]
2016-12-02 19:14:00,578 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.resourcemanager.zk-acl = [world:anyone:rwcda]
2016-12-02 19:14:00,578 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.resourcemanager.zk-address = [hscale-dev1-dn1:2181,hscale-dev1-dn3:2181,hscale-dev1-dn2:2181,hscale-dev1-dn4:2181]
2016-12-02 19:14:00,578 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.resourcemanager.zk-num-retries = [1000]
2016-12-02 19:14:00,578 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.resourcemanager.zk-retry-interval-ms = [1000]
2016-12-02 19:14:00,578 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.resourcemanager.zk-state-store.parent-path = [/rmstore]
2016-12-02 19:14:00,578 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.resourcemanager.zk-timeout-ms = [10000]
2016-12-02 19:14:00,578 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.scheduler.maximum-allocation-mb = [5120]
2016-12-02 19:14:00,578 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.scheduler.maximum-allocation-vcores = [8]
2016-12-02 19:14:00,578 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.scheduler.minimum-allocation-mb = [512]
2016-12-02 19:14:00,578 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.scheduler.minimum-allocation-vcores = [1]
2016-12-02 19:14:00,578 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.sharedcache.admin.address = [0.0.0.0:8047]
2016-12-02 19:14:00,578 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.sharedcache.admin.thread-count = [1]
2016-12-02 19:14:00,578 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.sharedcache.app-checker.class = [org.apache.hadoop.yarn.server.sharedcachemanager.RemoteAppChecker]
2016-12-02 19:14:00,578 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.sharedcache.checksum.algo.impl = [org.apache.hadoop.yarn.sharedcache.ChecksumSHA256Impl]
2016-12-02 19:14:00,578 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.sharedcache.cleaner.initial-delay-mins = [10]
2016-12-02 19:14:00,578 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.sharedcache.cleaner.period-mins = [1440]
2016-12-02 19:14:00,578 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.sharedcache.cleaner.resource-sleep-ms = [0]
2016-12-02 19:14:00,578 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.sharedcache.client-server.address = [0.0.0.0:8045]
2016-12-02 19:14:00,578 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.sharedcache.client-server.thread-count = [50]
2016-12-02 19:14:00,578 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.sharedcache.enabled = [false]
2016-12-02 19:14:00,578 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.sharedcache.nested-level = [3]
2016-12-02 19:14:00,578 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.sharedcache.nm.uploader.replication.factor = [10]
2016-12-02 19:14:00,578 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.sharedcache.nm.uploader.thread-count = [20]
2016-12-02 19:14:00,578 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.sharedcache.root-dir = [/sharedcache]
2016-12-02 19:14:00,578 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.sharedcache.store.class = [org.apache.hadoop.yarn.server.sharedcachemanager.store.InMemorySCMStore]
2016-12-02 19:14:00,578 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.sharedcache.store.in-memory.check-period-mins = [720]
2016-12-02 19:14:00,578 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.sharedcache.store.in-memory.initial-delay-mins = [10]
2016-12-02 19:14:00,578 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.sharedcache.store.in-memory.staleness-period-mins = [10080]
2016-12-02 19:14:00,578 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.sharedcache.uploader.server.address = [0.0.0.0:8046]
2016-12-02 19:14:00,578 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.sharedcache.uploader.server.thread-count = [50]
2016-12-02 19:14:00,578 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.sharedcache.webapp.address = [0.0.0.0:8788]
2016-12-02 19:14:00,578 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.timeline-service.address = [hscale-dev1-nn:10200]
2016-12-02 19:14:00,578 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.timeline-service.bind-host = [0.0.0.0]
2016-12-02 19:14:00,578 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.timeline-service.client.best-effort = [false]
2016-12-02 19:14:00,578 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.timeline-service.client.max-retries = [30]
2016-12-02 19:14:00,578 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.timeline-service.client.retry-interval-ms = [1000]
2016-12-02 19:14:00,578 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.timeline-service.enabled = [true]
2016-12-02 19:14:00,578 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.timeline-service.entity-group-fs-store.active-dir = [/ats/active/]
2016-12-02 19:14:00,578 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.timeline-service.entity-group-fs-store.cleaner-interval-seconds = [3600]
2016-12-02 19:14:00,578 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.timeline-service.entity-group-fs-store.done-dir = [/ats/done/]
2016-12-02 19:14:00,578 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.timeline-service.entity-group-fs-store.retain-seconds = [604800]
2016-12-02 19:14:00,578 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.timeline-service.entity-group-fs-store.scan-interval-seconds = [60]
2016-12-02 19:14:00,578 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.timeline-service.entity-group-fs-store.summary-store = [org.apache.hadoop.yarn.server.timeline.RollingLevelDBTimelineStore]
2016-12-02 19:14:00,578 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.timeline-service.generic-application-history.max-applications = [10000]
2016-12-02 19:14:00,578 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.timeline-service.generic-application-history.save-non-am-container-meta-info = [false]
2016-12-02 19:14:00,578 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.timeline-service.generic-application-history.store-class = [org.apache.hadoop.yarn.server.applicationhistoryservice.NullApplicationHistoryStore]
2016-12-02 19:14:00,578 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.timeline-service.handler-thread-count = [10]
2016-12-02 19:14:00,578 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.timeline-service.hostname = [0.0.0.0]
2016-12-02 19:14:00,578 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.timeline-service.http-authentication.simple.anonymous.allowed = [true]
2016-12-02 19:14:00,578 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.timeline-service.http-authentication.type = [simple]
2016-12-02 19:14:00,578 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.timeline-service.keytab = [/etc/krb5.keytab]
2016-12-02 19:14:00,578 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.timeline-service.leveldb-state-store.path = [/hadoop/yarn/timeline]
2016-12-02 19:14:00,578 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.timeline-service.leveldb-timeline-store.path = [/hadoop/yarn/timeline]
2016-12-02 19:14:00,578 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.timeline-service.leveldb-timeline-store.read-cache-size = [104857600]
2016-12-02 19:14:00,578 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.timeline-service.leveldb-timeline-store.start-time-read-cache-size = [10000]
2016-12-02 19:14:00,578 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.timeline-service.leveldb-timeline-store.start-time-write-cache-size = [10000]
2016-12-02 19:14:00,578 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.timeline-service.leveldb-timeline-store.ttl-interval-ms = [300000]
2016-12-02 19:14:00,578 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.timeline-service.recovery.enabled = [true]
2016-12-02 19:14:00,578 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.timeline-service.state-store-class = [org.apache.hadoop.yarn.server.timeline.recovery.LeveldbTimelineStateStore]
2016-12-02 19:14:00,578 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.timeline-service.store-class = [org.apache.hadoop.yarn.server.timeline.EntityGroupFSTimelineStore]
2016-12-02 19:14:00,578 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.timeline-service.ttl-enable = [true]
2016-12-02 19:14:00,578 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.timeline-service.ttl-ms = [2678400000]
2016-12-02 19:14:00,578 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.timeline-service.version = [1.5]
2016-12-02 19:14:00,578 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.timeline-service.webapp.address = [hscale-dev1-nn:8188]
2016-12-02 19:14:00,578 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.timeline-service.webapp.https.address = [hscale-dev1-nn:8190]
2016-12-02 19:14:00,578 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.timeline-service.webapp.rest-csrf.custom-header = [X-XSRF-Header]
2016-12-02 19:14:00,578 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.timeline-service.webapp.rest-csrf.enabled = [false]
2016-12-02 19:14:00,578 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.timeline-service.webapp.rest-csrf.methods-to-ignore = [GET,OPTIONS,HEAD]
2016-12-02 19:14:00,578 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  zookeeper.session.timeout = [120000]
2016-12-02 19:14:00,578 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  zookeeper.znode.acl.parent = [acl]
2016-12-02 19:14:00,578 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  zookeeper.znode.parent = [/hbase-unsecure]
2016-12-02 19:14:00,578 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  zookeeper.znode.rootserver = [root-region-server]
2016-12-02 19:14:00,579 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] zookeeper.RecoverableZooKeeper: Process identifier=spliceconnection connecting to ZooKeeper ensemble=hscale-dev1-dn1:2181,hscale-dev1-dn3:2181,hscale-dev1-dn2:2181,hscale-dev1-dn4:2181
2016-12-02 19:14:00,579 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] zookeeper.ZooKeeper: Initiating client connection, connectString=hscale-dev1-dn1:2181,hscale-dev1-dn3:2181,hscale-dev1-dn2:2181,hscale-dev1-dn4:2181 sessionTimeout=120000 watcher=spliceconnection0x0, quorum=hscale-dev1-dn1:2181,hscale-dev1-dn3:2181,hscale-dev1-dn2:2181,hscale-dev1-dn4:2181, baseZNode=/hbase-unsecure
2016-12-02 19:14:00,579 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020-SendThread(hscale-dev1-dn3:2181)] zookeeper.ClientCnxn: Opening socket connection to server hscale-dev1-dn3/10.60.70.13:2181. Will not attempt to authenticate using SASL (unknown error)
2016-12-02 19:14:00,580 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020-SendThread(hscale-dev1-dn3:2181)] zookeeper.ClientCnxn: Socket connection established to hscale-dev1-dn3/10.60.70.13:2181, initiating session
2016-12-02 19:14:00,589 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] hbase.ZkTimestampSource: Creating the TimestampClient...
2016-12-02 19:14:00,591 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] zookeeper.RecoverableZooKeeper: Process identifier=hconnection-0x47e92f4e connecting to ZooKeeper ensemble=hscale-dev1-dn1:2181,hscale-dev1-dn3:2181,hscale-dev1-dn2:2181,hscale-dev1-dn4:2181
2016-12-02 19:14:00,591 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] zookeeper.ZooKeeper: Initiating client connection, connectString=hscale-dev1-dn1:2181,hscale-dev1-dn3:2181,hscale-dev1-dn2:2181,hscale-dev1-dn4:2181 sessionTimeout=120000 watcher=hconnection-0x47e92f4e0x0, quorum=hscale-dev1-dn1:2181,hscale-dev1-dn3:2181,hscale-dev1-dn2:2181,hscale-dev1-dn4:2181, baseZNode=/hbase-unsecure
2016-12-02 19:14:00,591 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020-SendThread(hscale-dev1-dn3:2181)] zookeeper.ClientCnxn: Opening socket connection to server hscale-dev1-dn3/10.60.70.13:2181. Will not attempt to authenticate using SASL (unknown error)
2016-12-02 19:14:00,592 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020-SendThread(hscale-dev1-dn3:2181)] zookeeper.ClientCnxn: Socket connection established to hscale-dev1-dn3/10.60.70.13:2181, initiating session
2016-12-02 19:14:00,595 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020-SendThread(hscale-dev1-dn3:2181)] zookeeper.ClientCnxn: Session establishment complete on server hscale-dev1-dn3/10.60.70.13:2181, sessionid = 0x358bf3f651f001d, negotiated timeout = 120000
2016-12-02 19:14:00,607 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020-SendThread(hscale-dev1-dn3:2181)] zookeeper.ClientCnxn: Session establishment complete on server hscale-dev1-dn3/10.60.70.13:2181, sessionid = 0x358bf3f651f001e, negotiated timeout = 120000
2016-12-02 19:14:00,657 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] impl.TimestampClient: TimestampClient on region server successfully registered with JMX
2016-12-02 19:14:01,313 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] coprocessor.CoprocessorHost: System coprocessor com.splicemachine.hbase.RegionServerLifecycleObserver was loaded successfully with priority (536870911).
2016-12-02 19:14:01,314 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] regionserver.HRegionServer: reportForDuty to master=hscale-dev1-nn,16000,1480686229143 with port=16020, startcode=1480686239118
2016-12-02 19:14:01,324 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] hfile.CacheConfig: Allocating LruBlockCache size=768 MB, blockSize=64 KB
2016-12-02 19:14:01,332 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] hfile.CacheConfig: blockCache=LruBlockCache{blockCount=0, currentSize=827112, freeSize=804479256, maxSize=805306368, heapSize=827112, minSize=765041024, minFactor=0.95, multiSize=382520512, multiFactor=0.5, singleSize=191260256, singleFactor=0.25}, cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=true, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false
2016-12-02 19:14:01,488 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] wal.WALFactory: Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.BoundedRegionGroupingProvider
2016-12-02 19:14:01,488 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] wal.RegionGroupingProvider: Instantiating RegionGroupingStrategy of type class org.apache.hadoop.hbase.wal.RegionGroupingProvider$IdentityGroupingStrategy
2016-12-02 19:14:01,488 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] wal.WALFactory: Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.DefaultWALProvider
2016-12-02 19:14:01,551 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] wal.FSHLog: WAL configuration: blocksize=128 MB, rollsize=121.60 MB, prefix=hscale-dev1-dn1%2C16020%2C1480686239118.null0, suffix=, logDir=hdfs://hscale-dev1-nn:8020/apps/hbase/data/WALs/hscale-dev1-dn1,16020,1480686239118, archiveDir=hdfs://hscale-dev1-nn:8020/apps/hbase/data/oldWALs
2016-12-02 19:14:01,747 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] wal.FSHLog: Slow sync cost: 135 ms, current pipeline: []
2016-12-02 19:14:01,748 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] wal.FSHLog: New WAL /apps/hbase/data/WALs/hscale-dev1-dn1,16020,1480686239118/hscale-dev1-dn1%2C16020%2C1480686239118.null0.1480686241552
2016-12-02 19:14:01,762 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] wal.WALFactory: Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.DefaultWALProvider
2016-12-02 19:14:01,768 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] wal.FSHLog: WAL configuration: blocksize=128 MB, rollsize=121.60 MB, prefix=hscale-dev1-dn1%2C16020%2C1480686239118.null1, suffix=, logDir=hdfs://hscale-dev1-nn:8020/apps/hbase/data/WALs/hscale-dev1-dn1,16020,1480686239118, archiveDir=hdfs://hscale-dev1-nn:8020/apps/hbase/data/oldWALs
2016-12-02 19:14:01,817 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] wal.FSHLog: Slow sync cost: 33 ms, current pipeline: []
2016-12-02 19:14:01,817 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] wal.FSHLog: New WAL /apps/hbase/data/WALs/hscale-dev1-dn1,16020,1480686239118/hscale-dev1-dn1%2C16020%2C1480686239118.null1.1480686241768
2016-12-02 19:14:01,819 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] wal.WALFactory: Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.DefaultWALProvider
2016-12-02 19:14:01,826 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] wal.FSHLog: WAL configuration: blocksize=128 MB, rollsize=121.60 MB, prefix=hscale-dev1-dn1%2C16020%2C1480686239118.null2, suffix=, logDir=hdfs://hscale-dev1-nn:8020/apps/hbase/data/WALs/hscale-dev1-dn1,16020,1480686239118, archiveDir=hdfs://hscale-dev1-nn:8020/apps/hbase/data/oldWALs
2016-12-02 19:14:01,871 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] wal.FSHLog: Slow sync cost: 32 ms, current pipeline: []
2016-12-02 19:14:01,871 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] wal.FSHLog: New WAL /apps/hbase/data/WALs/hscale-dev1-dn1,16020,1480686239118/hscale-dev1-dn1%2C16020%2C1480686239118.null2.1480686241826
2016-12-02 19:14:01,872 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] wal.WALFactory: Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.DefaultWALProvider
2016-12-02 19:14:01,879 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] wal.FSHLog: WAL configuration: blocksize=128 MB, rollsize=121.60 MB, prefix=hscale-dev1-dn1%2C16020%2C1480686239118.null3, suffix=, logDir=hdfs://hscale-dev1-nn:8020/apps/hbase/data/WALs/hscale-dev1-dn1,16020,1480686239118, archiveDir=hdfs://hscale-dev1-nn:8020/apps/hbase/data/oldWALs
2016-12-02 19:14:01,925 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] wal.FSHLog: Slow sync cost: 32 ms, current pipeline: []
2016-12-02 19:14:01,925 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] wal.FSHLog: New WAL /apps/hbase/data/WALs/hscale-dev1-dn1,16020,1480686239118/hscale-dev1-dn1%2C16020%2C1480686239118.null3.1480686241879
2016-12-02 19:14:01,926 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] wal.WALFactory: Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.DefaultWALProvider
2016-12-02 19:14:01,932 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] wal.FSHLog: WAL configuration: blocksize=128 MB, rollsize=121.60 MB, prefix=hscale-dev1-dn1%2C16020%2C1480686239118.null4, suffix=, logDir=hdfs://hscale-dev1-nn:8020/apps/hbase/data/WALs/hscale-dev1-dn1,16020,1480686239118, archiveDir=hdfs://hscale-dev1-nn:8020/apps/hbase/data/oldWALs
2016-12-02 19:14:01,987 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] wal.FSHLog: Slow sync cost: 40 ms, current pipeline: []
2016-12-02 19:14:01,987 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] wal.FSHLog: New WAL /apps/hbase/data/WALs/hscale-dev1-dn1,16020,1480686239118/hscale-dev1-dn1%2C16020%2C1480686239118.null4.1480686241932
2016-12-02 19:14:01,990 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] wal.WALFactory: Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.DefaultWALProvider
2016-12-02 19:14:02,000 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] wal.FSHLog: WAL configuration: blocksize=128 MB, rollsize=121.60 MB, prefix=hscale-dev1-dn1%2C16020%2C1480686239118.null5, suffix=, logDir=hdfs://hscale-dev1-nn:8020/apps/hbase/data/WALs/hscale-dev1-dn1,16020,1480686239118, archiveDir=hdfs://hscale-dev1-nn:8020/apps/hbase/data/oldWALs
2016-12-02 19:14:02,050 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] wal.FSHLog: Slow sync cost: 32 ms, current pipeline: []
2016-12-02 19:14:02,050 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] wal.FSHLog: New WAL /apps/hbase/data/WALs/hscale-dev1-dn1,16020,1480686239118/hscale-dev1-dn1%2C16020%2C1480686239118.null5.1480686242000
2016-12-02 19:14:02,051 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] wal.WALFactory: Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.DefaultWALProvider
2016-12-02 19:14:02,056 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] wal.FSHLog: WAL configuration: blocksize=128 MB, rollsize=121.60 MB, prefix=hscale-dev1-dn1%2C16020%2C1480686239118.null6, suffix=, logDir=hdfs://hscale-dev1-nn:8020/apps/hbase/data/WALs/hscale-dev1-dn1,16020,1480686239118, archiveDir=hdfs://hscale-dev1-nn:8020/apps/hbase/data/oldWALs
2016-12-02 19:14:02,141 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] wal.FSHLog: Slow sync cost: 65 ms, current pipeline: []
2016-12-02 19:14:02,141 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] wal.FSHLog: New WAL /apps/hbase/data/WALs/hscale-dev1-dn1,16020,1480686239118/hscale-dev1-dn1%2C16020%2C1480686239118.null6.1480686242056
2016-12-02 19:14:02,144 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] wal.WALFactory: Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.DefaultWALProvider
2016-12-02 19:14:02,154 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] wal.FSHLog: WAL configuration: blocksize=128 MB, rollsize=121.60 MB, prefix=hscale-dev1-dn1%2C16020%2C1480686239118.null7, suffix=, logDir=hdfs://hscale-dev1-nn:8020/apps/hbase/data/WALs/hscale-dev1-dn1,16020,1480686239118, archiveDir=hdfs://hscale-dev1-nn:8020/apps/hbase/data/oldWALs
2016-12-02 19:14:02,229 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] wal.FSHLog: Slow sync cost: 57 ms, current pipeline: []
2016-12-02 19:14:02,229 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] wal.FSHLog: New WAL /apps/hbase/data/WALs/hscale-dev1-dn1,16020,1480686239118/hscale-dev1-dn1%2C16020%2C1480686239118.null7.1480686242154
2016-12-02 19:14:02,231 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] wal.WALFactory: Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.DefaultWALProvider
2016-12-02 19:14:02,237 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] wal.FSHLog: WAL configuration: blocksize=128 MB, rollsize=121.60 MB, prefix=hscale-dev1-dn1%2C16020%2C1480686239118.null8, suffix=, logDir=hdfs://hscale-dev1-nn:8020/apps/hbase/data/WALs/hscale-dev1-dn1,16020,1480686239118, archiveDir=hdfs://hscale-dev1-nn:8020/apps/hbase/data/oldWALs
2016-12-02 19:14:02,303 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] wal.FSHLog: Slow sync cost: 40 ms, current pipeline: []
2016-12-02 19:14:02,303 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] wal.FSHLog: New WAL /apps/hbase/data/WALs/hscale-dev1-dn1,16020,1480686239118/hscale-dev1-dn1%2C16020%2C1480686239118.null8.1480686242237
2016-12-02 19:14:02,305 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] wal.WALFactory: Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.DefaultWALProvider
2016-12-02 19:14:02,316 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] wal.FSHLog: WAL configuration: blocksize=128 MB, rollsize=121.60 MB, prefix=hscale-dev1-dn1%2C16020%2C1480686239118.null9, suffix=, logDir=hdfs://hscale-dev1-nn:8020/apps/hbase/data/WALs/hscale-dev1-dn1,16020,1480686239118, archiveDir=hdfs://hscale-dev1-nn:8020/apps/hbase/data/oldWALs
2016-12-02 19:14:02,386 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] wal.FSHLog: Slow sync cost: 52 ms, current pipeline: []
2016-12-02 19:14:02,386 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] wal.FSHLog: New WAL /apps/hbase/data/WALs/hscale-dev1-dn1,16020,1480686239118/hscale-dev1-dn1%2C16020%2C1480686239118.null9.1480686242316
2016-12-02 19:14:02,388 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] wal.WALFactory: Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.DefaultWALProvider
2016-12-02 19:14:02,397 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] wal.FSHLog: WAL configuration: blocksize=128 MB, rollsize=121.60 MB, prefix=hscale-dev1-dn1%2C16020%2C1480686239118.null10, suffix=, logDir=hdfs://hscale-dev1-nn:8020/apps/hbase/data/WALs/hscale-dev1-dn1,16020,1480686239118, archiveDir=hdfs://hscale-dev1-nn:8020/apps/hbase/data/oldWALs
2016-12-02 19:14:02,465 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] wal.FSHLog: Slow sync cost: 48 ms, current pipeline: []
2016-12-02 19:14:02,465 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] wal.FSHLog: New WAL /apps/hbase/data/WALs/hscale-dev1-dn1,16020,1480686239118/hscale-dev1-dn1%2C16020%2C1480686239118.null10.1480686242397
2016-12-02 19:14:02,465 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] wal.WALFactory: Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.DefaultWALProvider
2016-12-02 19:14:02,469 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] wal.FSHLog: WAL configuration: blocksize=128 MB, rollsize=121.60 MB, prefix=hscale-dev1-dn1%2C16020%2C1480686239118.null11, suffix=, logDir=hdfs://hscale-dev1-nn:8020/apps/hbase/data/WALs/hscale-dev1-dn1,16020,1480686239118, archiveDir=hdfs://hscale-dev1-nn:8020/apps/hbase/data/oldWALs
2016-12-02 19:14:02,540 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] wal.FSHLog: Slow sync cost: 45 ms, current pipeline: []
2016-12-02 19:14:02,540 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] wal.FSHLog: New WAL /apps/hbase/data/WALs/hscale-dev1-dn1,16020,1480686239118/hscale-dev1-dn1%2C16020%2C1480686239118.null11.1480686242478
2016-12-02 19:14:02,541 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] wal.WALFactory: Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.DefaultWALProvider
2016-12-02 19:14:02,553 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] wal.FSHLog: WAL configuration: blocksize=128 MB, rollsize=121.60 MB, prefix=hscale-dev1-dn1%2C16020%2C1480686239118.null12, suffix=, logDir=hdfs://hscale-dev1-nn:8020/apps/hbase/data/WALs/hscale-dev1-dn1,16020,1480686239118, archiveDir=hdfs://hscale-dev1-nn:8020/apps/hbase/data/oldWALs
2016-12-02 19:14:02,635 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] wal.FSHLog: Slow sync cost: 65 ms, current pipeline: []
2016-12-02 19:14:02,636 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] wal.FSHLog: New WAL /apps/hbase/data/WALs/hscale-dev1-dn1,16020,1480686239118/hscale-dev1-dn1%2C16020%2C1480686239118.null12.1480686242553
2016-12-02 19:14:02,637 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] wal.WALFactory: Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.DefaultWALProvider
2016-12-02 19:14:02,647 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] wal.FSHLog: WAL configuration: blocksize=128 MB, rollsize=121.60 MB, prefix=hscale-dev1-dn1%2C16020%2C1480686239118.null13, suffix=, logDir=hdfs://hscale-dev1-nn:8020/apps/hbase/data/WALs/hscale-dev1-dn1,16020,1480686239118, archiveDir=hdfs://hscale-dev1-nn:8020/apps/hbase/data/oldWALs
2016-12-02 19:14:02,716 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] wal.FSHLog: Slow sync cost: 49 ms, current pipeline: []
2016-12-02 19:14:02,716 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] wal.FSHLog: New WAL /apps/hbase/data/WALs/hscale-dev1-dn1,16020,1480686239118/hscale-dev1-dn1%2C16020%2C1480686239118.null13.1480686242647
2016-12-02 19:14:02,717 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] wal.WALFactory: Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.DefaultWALProvider
2016-12-02 19:14:02,725 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] wal.FSHLog: WAL configuration: blocksize=128 MB, rollsize=121.60 MB, prefix=hscale-dev1-dn1%2C16020%2C1480686239118.null14, suffix=, logDir=hdfs://hscale-dev1-nn:8020/apps/hbase/data/WALs/hscale-dev1-dn1,16020,1480686239118, archiveDir=hdfs://hscale-dev1-nn:8020/apps/hbase/data/oldWALs
2016-12-02 19:14:02,811 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] wal.FSHLog: Slow sync cost: 74 ms, current pipeline: []
2016-12-02 19:14:02,811 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] wal.FSHLog: New WAL /apps/hbase/data/WALs/hscale-dev1-dn1,16020,1480686239118/hscale-dev1-dn1%2C16020%2C1480686239118.null14.1480686242725
2016-12-02 19:14:02,811 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] wal.WALFactory: Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.DefaultWALProvider
2016-12-02 19:14:02,818 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] wal.FSHLog: WAL configuration: blocksize=128 MB, rollsize=121.60 MB, prefix=hscale-dev1-dn1%2C16020%2C1480686239118.null15, suffix=, logDir=hdfs://hscale-dev1-nn:8020/apps/hbase/data/WALs/hscale-dev1-dn1,16020,1480686239118, archiveDir=hdfs://hscale-dev1-nn:8020/apps/hbase/data/oldWALs
2016-12-02 19:14:02,903 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] wal.FSHLog: Slow sync cost: 57 ms, current pipeline: []
2016-12-02 19:14:02,903 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] wal.FSHLog: New WAL /apps/hbase/data/WALs/hscale-dev1-dn1,16020,1480686239118/hscale-dev1-dn1%2C16020%2C1480686239118.null15.1480686242818
2016-12-02 19:14:02,904 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] wal.BoundedRegionGroupingProvider: Configured to run with 16 delegate WAL providers.
2016-12-02 19:14:02,907 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] regionserver.MetricsRegionServerWrapperImpl: Computing regionserver metrics every 5000 milliseconds
2016-12-02 19:14:02,917 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] regionserver.ReplicationSourceManager: Current list of replicators: [hscale-dev1-dn1,16020,1480686239118, hscale-dev1-dn3,16020,1480686232696] other RSs: [hscale-dev1-dn1,16020,1480686239118, hscale-dev1-dn2,16020,1480686240889, hscale-dev1-dn3,16020,1480686232696]
2016-12-02 19:14:02,956 INFO  [SplitLogWorker-hscale-dev1-dn1:16020] regionserver.SplitLogWorker: SplitLogWorker hscale-dev1-dn1,16020,1480686239118 starting
2016-12-02 19:14:02,957 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] regionserver.HeapMemoryManager: Starting HeapMemoryTuner chore.
2016-12-02 19:14:02,958 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] regionserver.HRegionServer: Serving as hscale-dev1-dn1,16020,1480686239118, RpcServer on hscale-dev1-dn1/10.60.70.11:16020, sessionid=0x358bf3f651f001c
2016-12-02 19:14:02,966 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] quotas.RegionServerQuotaManager: Quota support disabled
2016-12-02 19:14:03,441 ERROR [pool-10-thread-1] lifecycle.DatabaseLifecycleManager: Error during during startup of service com.splicemachine.derby.lifecycle.MonitoredLifecycleService@2bca434:
org.apache.hadoop.hbase.client.RetriesExhaustedException: Failed after attempts=5, exceptions:
Fri Dec 02 19:14:01 IST 2016, RpcRetryingCaller{globalStartTime=1480686241331, pause=100, retries=5}, org.apache.hadoop.hbase.ipc.RemoteWithExtrasException(org.apache.hadoop.hbase.PleaseHoldException): org.apache.hadoop.hbase.PleaseHoldException: Please Hold - Starting
    at com.splicemachine.hbase.SpliceMasterObserver.preCreateTable(SpliceMasterObserver.java:115)
    at org.apache.hadoop.hbase.master.MasterCoprocessorHost$11.call(MasterCoprocessorHost.java:216)
    at org.apache.hadoop.hbase.master.MasterCoprocessorHost.execOperation(MasterCoprocessorHost.java:1140)
    at org.apache.hadoop.hbase.master.MasterCoprocessorHost.preCreateTable(MasterCoprocessorHost.java:212)
    at org.apache.hadoop.hbase.master.HMaster.createTable(HMaster.java:1533)
    at org.apache.hadoop.hbase.master.MasterRpcServices.createTable(MasterRpcServices.java:454)
    at org.apache.hadoop.hbase.protobuf.generated.MasterProtos$MasterService$2.callBlockingMethod(MasterProtos.java:55401)
    at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2114)
    at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:101)
    at org.apache.hadoop.hbase.ipc.RpcExecutor.consumerLoop(RpcExecutor.java:130)
    at org.apache.hadoop.hbase.ipc.RpcExecutor$1.run(RpcExecutor.java:107)
    at java.lang.Thread.run(Thread.java:745)

Fri Dec 02 19:14:01 IST 2016, RpcRetryingCaller{globalStartTime=1480686241331, pause=100, retries=5}, org.apache.hadoop.hbase.ipc.RemoteWithExtrasException(org.apache.hadoop.hbase.PleaseHoldException): org.apache.hadoop.hbase.PleaseHoldException: Please Hold - Starting
    at com.splicemachine.hbase.SpliceMasterObserver.preCreateTable(SpliceMasterObserver.java:115)
    at org.apache.hadoop.hbase.master.MasterCoprocessorHost$11.call(MasterCoprocessorHost.java:216)
    at org.apache.hadoop.hbase.master.MasterCoprocessorHost.execOperation(MasterCoprocessorHost.java:1140)
    at org.apache.hadoop.hbase.master.MasterCoprocessorHost.preCreateTable(MasterCoprocessorHost.java:212)
    at org.apache.hadoop.hbase.master.HMaster.createTable(HMaster.java:1533)
    at org.apache.hadoop.hbase.master.MasterRpcServices.createTable(MasterRpcServices.java:454)
    at org.apache.hadoop.hbase.protobuf.generated.MasterProtos$MasterService$2.callBlockingMethod(MasterProtos.java:55401)
    at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2114)
    at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:101)
    at org.apache.hadoop.hbase.ipc.RpcExecutor.consumerLoop(RpcExecutor.java:130)
    at org.apache.hadoop.hbase.ipc.RpcExecutor$1.run(RpcExecutor.java:107)
    at java.lang.Thread.run(Thread.java:745)

Fri Dec 02 19:14:01 IST 2016, RpcRetryingCaller{globalStartTime=1480686241331, pause=100, retries=5}, org.apache.hadoop.hbase.ipc.RemoteWithExtrasException(org.apache.hadoop.hbase.PleaseHoldException): org.apache.hadoop.hbase.PleaseHoldException: Please Hold - Starting
    at com.splicemachine.hbase.SpliceMasterObserver.preCreateTable(SpliceMasterObserver.java:115)
    at org.apache.hadoop.hbase.master.MasterCoprocessorHost$11.call(MasterCoprocessorHost.java:216)
    at org.apache.hadoop.hbase.master.MasterCoprocessorHost.execOperation(MasterCoprocessorHost.java:1140)
    at org.apache.hadoop.hbase.master.MasterCoprocessorHost.preCreateTable(MasterCoprocessorHost.java:212)
    at org.apache.hadoop.hbase.master.HMaster.createTable(HMaster.java:1533)
    at org.apache.hadoop.hbase.master.MasterRpcServices.createTable(MasterRpcServices.java:454)
    at org.apache.hadoop.hbase.protobuf.generated.MasterProtos$MasterService$2.callBlockingMethod(MasterProtos.java:55401)
    at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2114)
    at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:101)
    at org.apache.hadoop.hbase.ipc.RpcExecutor.consumerLoop(RpcExecutor.java:130)
    at org.apache.hadoop.hbase.ipc.RpcExecutor$1.run(RpcExecutor.java:107)
    at java.lang.Thread.run(Thread.java:745)

Fri Dec 02 19:14:02 IST 2016, RpcRetryingCaller{globalStartTime=1480686241331, pause=100, retries=5}, org.apache.hadoop.hbase.ipc.RemoteWithExtrasException(org.apache.hadoop.hbase.PleaseHoldException): org.apache.hadoop.hbase.PleaseHoldException: Please Hold - Starting
    at com.splicemachine.hbase.SpliceMasterObserver.preCreateTable(SpliceMasterObserver.java:115)
    at org.apache.hadoop.hbase.master.MasterCoprocessorHost$11.call(MasterCoprocessorHost.java:216)
    at org.apache.hadoop.hbase.master.MasterCoprocessorHost.execOperation(MasterCoprocessorHost.java:1140)
    at org.apache.hadoop.hbase.master.MasterCoprocessorHost.preCreateTable(MasterCoprocessorHost.java:212)
    at org.apache.hadoop.hbase.master.HMaster.createTable(HMaster.java:1533)
    at org.apache.hadoop.hbase.master.MasterRpcServices.createTable(MasterRpcServices.java:454)
    at org.apache.hadoop.hbase.protobuf.generated.MasterProtos$MasterService$2.callBlockingMethod(MasterProtos.java:55401)
    at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2114)
    at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:101)
    at org.apache.hadoop.hbase.ipc.RpcExecutor.consumerLoop(RpcExecutor.java:130)
    at org.apache.hadoop.hbase.ipc.RpcExecutor$1.run(RpcExecutor.java:107)
    at java.lang.Thread.run(Thread.java:745)

Fri Dec 02 19:14:03 IST 2016, RpcRetryingCaller{globalStartTime=1480686241331, pause=100, retries=5}, org.apache.hadoop.hbase.ipc.RemoteWithExtrasException(org.apache.hadoop.hbase.PleaseHoldException): org.apache.hadoop.hbase.PleaseHoldException: Please Hold - Starting
    at com.splicemachine.hbase.SpliceMasterObserver.preCreateTable(SpliceMasterObserver.java:115)
    at org.apache.hadoop.hbase.master.MasterCoprocessorHost$11.call(MasterCoprocessorHost.java:216)
    at org.apache.hadoop.hbase.master.MasterCoprocessorHost.execOperation(MasterCoprocessorHost.java:1140)
    at org.apache.hadoop.hbase.master.MasterCoprocessorHost.preCreateTable(MasterCoprocessorHost.java:212)
    at org.apache.hadoop.hbase.master.HMaster.createTable(HMaster.java:1533)
    at org.apache.hadoop.hbase.master.MasterRpcServices.createTable(MasterRpcServices.java:454)
    at org.apache.hadoop.hbase.protobuf.generated.MasterProtos$MasterService$2.callBlockingMethod(MasterProtos.java:55401)
    at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2114)
    at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:101)
    at org.apache.hadoop.hbase.ipc.RpcExecutor.consumerLoop(RpcExecutor.java:130)
    at org.apache.hadoop.hbase.ipc.RpcExecutor$1.run(RpcExecutor.java:107)
    at java.lang.Thread.run(Thread.java:745)


    at org.apache.hadoop.hbase.client.RpcRetryingCaller.callWithRetries(RpcRetryingCaller.java:147)
    at org.apache.hadoop.hbase.client.HBaseAdmin.executeCallable(HBaseAdmin.java:4083)
    at org.apache.hadoop.hbase.client.HBaseAdmin.createTableAsyncV2(HBaseAdmin.java:723)
    at org.apache.hadoop.hbase.client.HBaseAdmin.createTable(HBaseAdmin.java:644)
    at org.apache.hadoop.hbase.client.HBaseAdmin.createTable(HBaseAdmin.java:577)
    at com.splicemachine.lifecycle.RegionServerLifecycle.distributedStart(RegionServerLifecycle.java:66)
    at com.splicemachine.derby.lifecycle.EngineLifecycleService.start(EngineLifecycleService.java:81)
    at com.splicemachine.lifecycle.DatabaseLifecycleManager$Startup.bootServices(DatabaseLifecycleManager.java:229)
    at com.splicemachine.lifecycle.DatabaseLifecycleManager$Startup.run(DatabaseLifecycleManager.java:220)
    at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
    at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
    at java.lang.Thread.run(Thread.java:745)
Caused by: org.apache.hadoop.hbase.ipc.RemoteWithExtrasException(org.apache.hadoop.hbase.PleaseHoldException): org.apache.hadoop.hbase.PleaseHoldException: Please Hold - Starting
    at com.splicemachine.hbase.SpliceMasterObserver.preCreateTable(SpliceMasterObserver.java:115)
    at org.apache.hadoop.hbase.master.MasterCoprocessorHost$11.call(MasterCoprocessorHost.java:216)
    at org.apache.hadoop.hbase.master.MasterCoprocessorHost.execOperation(MasterCoprocessorHost.java:1140)
    at org.apache.hadoop.hbase.master.MasterCoprocessorHost.preCreateTable(MasterCoprocessorHost.java:212)
    at org.apache.hadoop.hbase.master.HMaster.createTable(HMaster.java:1533)
    at org.apache.hadoop.hbase.master.MasterRpcServices.createTable(MasterRpcServices.java:454)
    at org.apache.hadoop.hbase.protobuf.generated.MasterProtos$MasterService$2.callBlockingMethod(MasterProtos.java:55401)
    at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2114)
    at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:101)
    at org.apache.hadoop.hbase.ipc.RpcExecutor.consumerLoop(RpcExecutor.java:130)
    at org.apache.hadoop.hbase.ipc.RpcExecutor$1.run(RpcExecutor.java:107)
    at java.lang.Thread.run(Thread.java:745)

    at org.apache.hadoop.hbase.ipc.RpcClientImpl.call(RpcClientImpl.java:1226)
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient.callBlockingMethod(AbstractRpcClient.java:213)
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient$BlockingRpcChannelImplementation.callBlockingMethod(AbstractRpcClient.java:287)
    at org.apache.hadoop.hbase.protobuf.generated.MasterProtos$MasterService$BlockingStub.createTable(MasterProtos.java:58320)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation$4.createTable(ConnectionManager.java:1821)
    at org.apache.hadoop.hbase.client.HBaseAdmin$5.call(HBaseAdmin.java:728)
    at org.apache.hadoop.hbase.client.HBaseAdmin$5.call(HBaseAdmin.java:724)
    at org.apache.hadoop.hbase.client.RpcRetryingCaller.callWithRetries(RpcRetryingCaller.java:126)
    ... 11 more
2016-12-02 19:14:03,446 ERROR [pool-10-thread-1] lifecycle.DatabaseLifecycleManager: Error during shutdown of service com.splicemachine.derby.lifecycle.NetworkLifecycleService@3ba69bc4:
java.lang.NullPointerException
    at com.splicemachine.derby.lifecycle.NetworkLifecycleService.shutdown(NetworkLifecycleService.java:65)
    at com.splicemachine.lifecycle.DatabaseLifecycleManager$Shutdown.run(DatabaseLifecycleManager.java:268)
    at com.splicemachine.lifecycle.DatabaseLifecycleManager$Startup.bootServices(DatabaseLifecycleManager.java:233)
    at com.splicemachine.lifecycle.DatabaseLifecycleManager$Startup.run(DatabaseLifecycleManager.java:220)
    at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
    at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
    at java.lang.Thread.run(Thread.java:745)
2016-12-02 19:14:03,446 INFO  [pool-10-thread-1] impl.TimestampClient: shutting down TimestampClient state=SHUTDOWN
2016-12-02 19:14:03,862 INFO  [PriorityRpcServer.handler=0,queue=0,port=16020] regionserver.RSRpcServices: Open splice:SPLICE_TXN,\x09,1480686242157.ace1654b29a8315feb0793d463b78246.
2016-12-02 19:14:04,107 INFO  [PriorityRpcServer.handler=0,queue=0,port=16020] regionserver.RSRpcServices: Open splice:SPLICE_TXN,,1480686242157.befbb055ce59b866459081d27215d354.
2016-12-02 19:14:04,109 INFO  [PriorityRpcServer.handler=0,queue=0,port=16020] regionserver.RSRpcServices: Open splice:SPLICE_TXN,\x07,1480686242157.c3cc35c8ce6fe046e78dbb491c804d1d.
2016-12-02 19:14:04,112 INFO  [PriorityRpcServer.handler=0,queue=0,port=16020] regionserver.RSRpcServices: Open splice:SPLICE_TXN,\x0A,1480686242157.91b7a8c3c77f908c0cff6940ccfd6d7d.
2016-12-02 19:14:04,115 INFO  [PriorityRpcServer.handler=0,queue=0,port=16020] regionserver.RSRpcServices: Open splice:SPLICE_TXN,\x0F,1480686242157.c161d85dba1212e9f283596f80de56c8.
2016-12-02 19:14:04,421 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-2] coprocessor.CoprocessorHost: System coprocessor org.apache.hadoop.hbase.security.access.SecureBulkLoadEndpoint was loaded successfully with priority (536870911).
2016-12-02 19:14:04,421 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-0] coprocessor.CoprocessorHost: System coprocessor org.apache.hadoop.hbase.security.access.SecureBulkLoadEndpoint was loaded successfully with priority (536870911).
2016-12-02 19:14:04,421 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-1] coprocessor.CoprocessorHost: System coprocessor org.apache.hadoop.hbase.security.access.SecureBulkLoadEndpoint was loaded successfully with priority (536870911).
2016-12-02 19:14:04,453 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-2] coprocessor.CoprocessorHost: System coprocessor com.splicemachine.hbase.MemstoreAwareObserver was loaded successfully with priority (536870912).
2016-12-02 19:14:04,453 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-0] coprocessor.CoprocessorHost: System coprocessor com.splicemachine.hbase.MemstoreAwareObserver was loaded successfully with priority (536870912).
2016-12-02 19:14:04,453 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-1] coprocessor.CoprocessorHost: System coprocessor com.splicemachine.hbase.MemstoreAwareObserver was loaded successfully with priority (536870912).
2016-12-02 19:14:04,473 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-1] coprocessor.CoprocessorHost: System coprocessor com.splicemachine.derby.hbase.SpliceIndexObserver was loaded successfully with priority (536870913).
2016-12-02 19:14:04,474 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-0] coprocessor.CoprocessorHost: System coprocessor com.splicemachine.derby.hbase.SpliceIndexObserver was loaded successfully with priority (536870913).
2016-12-02 19:14:04,474 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-2] coprocessor.CoprocessorHost: System coprocessor com.splicemachine.derby.hbase.SpliceIndexObserver was loaded successfully with priority (536870913).
2016-12-02 19:14:04,493 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-1] coprocessor.CoprocessorHost: System coprocessor com.splicemachine.derby.hbase.SpliceIndexEndpoint was loaded successfully with priority (536870914).
2016-12-02 19:14:04,493 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-0] coprocessor.CoprocessorHost: System coprocessor com.splicemachine.derby.hbase.SpliceIndexEndpoint was loaded successfully with priority (536870914).
2016-12-02 19:14:04,493 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-2] coprocessor.CoprocessorHost: System coprocessor com.splicemachine.derby.hbase.SpliceIndexEndpoint was loaded successfully with priority (536870914).
2016-12-02 19:14:04,501 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-0] coprocessor.CoprocessorHost: System coprocessor com.splicemachine.hbase.RegionSizeEndpoint was loaded successfully with priority (536870915).
2016-12-02 19:14:04,501 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-1] coprocessor.CoprocessorHost: System coprocessor com.splicemachine.hbase.RegionSizeEndpoint was loaded successfully with priority (536870915).
2016-12-02 19:14:04,501 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-2] coprocessor.CoprocessorHost: System coprocessor com.splicemachine.hbase.RegionSizeEndpoint was loaded successfully with priority (536870915).
2016-12-02 19:14:04,543 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-0] coprocessor.CoprocessorHost: System coprocessor com.splicemachine.si.data.hbase.coprocessor.TxnLifecycleEndpoint was loaded successfully with priority (536870916).
2016-12-02 19:14:04,543 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-2] coprocessor.CoprocessorHost: System coprocessor com.splicemachine.si.data.hbase.coprocessor.TxnLifecycleEndpoint was loaded successfully with priority (536870916).
2016-12-02 19:14:04,543 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-1] coprocessor.CoprocessorHost: System coprocessor com.splicemachine.si.data.hbase.coprocessor.TxnLifecycleEndpoint was loaded successfully with priority (536870916).
2016-12-02 19:14:04,565 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-2] coprocessor.CoprocessorHost: System coprocessor com.splicemachine.si.data.hbase.coprocessor.SIObserver was loaded successfully with priority (536870917).
2016-12-02 19:14:04,565 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-0] coprocessor.CoprocessorHost: System coprocessor com.splicemachine.si.data.hbase.coprocessor.SIObserver was loaded successfully with priority (536870917).
2016-12-02 19:14:04,565 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-1] coprocessor.CoprocessorHost: System coprocessor com.splicemachine.si.data.hbase.coprocessor.SIObserver was loaded successfully with priority (536870917).
2016-12-02 19:14:04,583 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-0] coprocessor.CoprocessorHost: System coprocessor com.splicemachine.hbase.BackupEndpointObserver was loaded successfully with priority (536870918).
2016-12-02 19:14:04,583 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-1] coprocessor.CoprocessorHost: System coprocessor com.splicemachine.hbase.BackupEndpointObserver was loaded successfully with priority (536870918).
2016-12-02 19:14:04,583 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-2] coprocessor.CoprocessorHost: System coprocessor com.splicemachine.hbase.BackupEndpointObserver was loaded successfully with priority (536870918).
2016-12-02 19:14:04,696 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-2] compress.CodecPool: Got brand-new compressor [.snappy]
2016-12-02 19:14:04,696 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-1] compress.CodecPool: Got brand-new compressor [.snappy]
2016-12-02 19:14:04,696 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-0] compress.CodecPool: Got brand-new compressor [.snappy]
2016-12-02 19:14:04,810 INFO  [StoreOpener-c3cc35c8ce6fe046e78dbb491c804d1d-1] hfile.CacheConfig: blockCache=LruBlockCache{blockCount=0, currentSize=827112, freeSize=804479256, maxSize=805306368, heapSize=827112, minSize=765041024, minFactor=0.95, multiSize=382520512, multiFactor=0.5, singleSize=191260256, singleFactor=0.25}, cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=true, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false
2016-12-02 19:14:04,810 INFO  [StoreOpener-ace1654b29a8315feb0793d463b78246-1] hfile.CacheConfig: blockCache=LruBlockCache{blockCount=0, currentSize=827112, freeSize=804479256, maxSize=805306368, heapSize=827112, minSize=765041024, minFactor=0.95, multiSize=382520512, multiFactor=0.5, singleSize=191260256, singleFactor=0.25}, cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=true, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false
2016-12-02 19:14:04,810 INFO  [StoreOpener-befbb055ce59b866459081d27215d354-1] hfile.CacheConfig: blockCache=LruBlockCache{blockCount=0, currentSize=827112, freeSize=804479256, maxSize=805306368, heapSize=827112, minSize=765041024, minFactor=0.95, multiSize=382520512, multiFactor=0.5, singleSize=191260256, singleFactor=0.25}, cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=true, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false
2016-12-02 19:14:04,851 INFO  [StoreOpener-befbb055ce59b866459081d27215d354-1] compactions.CompactionConfiguration: size [16777216, 260046848); files [5, 10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000
2016-12-02 19:14:04,851 INFO  [StoreOpener-c3cc35c8ce6fe046e78dbb491c804d1d-1] compactions.CompactionConfiguration: size [16777216, 260046848); files [5, 10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000
2016-12-02 19:14:04,851 INFO  [StoreOpener-ace1654b29a8315feb0793d463b78246-1] compactions.CompactionConfiguration: size [16777216, 260046848); files [5, 10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000
2016-12-02 19:14:04,901 INFO  [StoreOpener-befbb055ce59b866459081d27215d354-1] hfile.CacheConfig: blockCache=LruBlockCache{blockCount=0, currentSize=827112, freeSize=804479256, maxSize=805306368, heapSize=827112, minSize=765041024, minFactor=0.95, multiSize=382520512, multiFactor=0.5, singleSize=191260256, singleFactor=0.25}, cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=true, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false
2016-12-02 19:14:04,901 INFO  [StoreOpener-ace1654b29a8315feb0793d463b78246-1] hfile.CacheConfig: blockCache=LruBlockCache{blockCount=0, currentSize=827112, freeSize=804479256, maxSize=805306368, heapSize=827112, minSize=765041024, minFactor=0.95, multiSize=382520512, multiFactor=0.5, singleSize=191260256, singleFactor=0.25}, cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=true, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false
2016-12-02 19:14:04,901 INFO  [StoreOpener-befbb055ce59b866459081d27215d354-1] compactions.CompactionConfiguration: size [16777216, 260046848); files [5, 10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000
2016-12-02 19:14:04,901 INFO  [StoreOpener-c3cc35c8ce6fe046e78dbb491c804d1d-1] hfile.CacheConfig: blockCache=LruBlockCache{blockCount=0, currentSize=827112, freeSize=804479256, maxSize=805306368, heapSize=827112, minSize=765041024, minFactor=0.95, multiSize=382520512, multiFactor=0.5, singleSize=191260256, singleFactor=0.25}, cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=true, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false
2016-12-02 19:14:04,902 INFO  [StoreOpener-c3cc35c8ce6fe046e78dbb491c804d1d-1] compactions.CompactionConfiguration: size [16777216, 260046848); files [5, 10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000
2016-12-02 19:14:04,901 INFO  [StoreOpener-ace1654b29a8315feb0793d463b78246-1] compactions.CompactionConfiguration: size [16777216, 260046848); files [5, 10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000
2016-12-02 19:14:05,001 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-2] regionserver.HRegion: Onlined c3cc35c8ce6fe046e78dbb491c804d1d; next sequenceid=2
2016-12-02 19:14:05,001 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-0] regionserver.HRegion: Onlined ace1654b29a8315feb0793d463b78246; next sequenceid=2
2016-12-02 19:14:05,001 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-1] regionserver.HRegion: Onlined befbb055ce59b866459081d27215d354; next sequenceid=2
2016-12-02 19:14:05,059 INFO  [PostOpenDeployTasks:befbb055ce59b866459081d27215d354] regionserver.HRegionServer: Post open deploy tasks for splice:SPLICE_TXN,,1480686242157.befbb055ce59b866459081d27215d354.
2016-12-02 19:14:05,060 INFO  [PostOpenDeployTasks:c3cc35c8ce6fe046e78dbb491c804d1d] regionserver.HRegionServer: Post open deploy tasks for splice:SPLICE_TXN,\x07,1480686242157.c3cc35c8ce6fe046e78dbb491c804d1d.
2016-12-02 19:14:05,060 INFO  [PostOpenDeployTasks:ace1654b29a8315feb0793d463b78246] regionserver.HRegionServer: Post open deploy tasks for splice:SPLICE_TXN,\x09,1480686242157.ace1654b29a8315feb0793d463b78246.
2016-12-02 19:14:05,168 INFO  [PostOpenDeployTasks:c3cc35c8ce6fe046e78dbb491c804d1d] hbase.MetaTableAccessor: Updated row splice:SPLICE_TXN,\x07,1480686242157.c3cc35c8ce6fe046e78dbb491c804d1d. with server=hscale-dev1-dn1,16020,1480686239118
2016-12-02 19:14:05,169 INFO  [PostOpenDeployTasks:ace1654b29a8315feb0793d463b78246] hbase.MetaTableAccessor: Updated row splice:SPLICE_TXN,\x09,1480686242157.ace1654b29a8315feb0793d463b78246. with server=hscale-dev1-dn1,16020,1480686239118
2016-12-02 19:14:05,169 INFO  [PostOpenDeployTasks:befbb055ce59b866459081d27215d354] hbase.MetaTableAccessor: Updated row splice:SPLICE_TXN,,1480686242157.befbb055ce59b866459081d27215d354. with server=hscale-dev1-dn1,16020,1480686239118
2016-12-02 19:14:05,247 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-2] coprocessor.CoprocessorHost: System coprocessor org.apache.hadoop.hbase.security.access.SecureBulkLoadEndpoint was loaded successfully with priority (536870911).
2016-12-02 19:14:05,248 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-2] coprocessor.CoprocessorHost: System coprocessor com.splicemachine.hbase.MemstoreAwareObserver was loaded successfully with priority (536870912).
2016-12-02 19:14:05,248 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-2] coprocessor.CoprocessorHost: System coprocessor com.splicemachine.derby.hbase.SpliceIndexObserver was loaded successfully with priority (536870913).
2016-12-02 19:14:05,248 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-2] coprocessor.CoprocessorHost: System coprocessor com.splicemachine.derby.hbase.SpliceIndexEndpoint was loaded successfully with priority (536870914).
2016-12-02 19:14:05,248 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-2] coprocessor.CoprocessorHost: System coprocessor com.splicemachine.hbase.RegionSizeEndpoint was loaded successfully with priority (536870915).
2016-12-02 19:14:05,249 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-2] coprocessor.CoprocessorHost: System coprocessor com.splicemachine.si.data.hbase.coprocessor.TxnLifecycleEndpoint was loaded successfully with priority (536870916).
2016-12-02 19:14:05,249 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-2] coprocessor.CoprocessorHost: System coprocessor com.splicemachine.si.data.hbase.coprocessor.SIObserver was loaded successfully with priority (536870917).
2016-12-02 19:14:05,250 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-2] coprocessor.CoprocessorHost: System coprocessor com.splicemachine.hbase.BackupEndpointObserver was loaded successfully with priority (536870918).
2016-12-02 19:14:05,258 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-0] coprocessor.CoprocessorHost: System coprocessor org.apache.hadoop.hbase.security.access.SecureBulkLoadEndpoint was loaded successfully with priority (536870911).
2016-12-02 19:14:05,258 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-0] coprocessor.CoprocessorHost: System coprocessor com.splicemachine.hbase.MemstoreAwareObserver was loaded successfully with priority (536870912).
2016-12-02 19:14:05,258 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-0] coprocessor.CoprocessorHost: System coprocessor com.splicemachine.derby.hbase.SpliceIndexObserver was loaded successfully with priority (536870913).
2016-12-02 19:14:05,258 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-0] coprocessor.CoprocessorHost: System coprocessor com.splicemachine.derby.hbase.SpliceIndexEndpoint was loaded successfully with priority (536870914).
2016-12-02 19:14:05,259 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-0] coprocessor.CoprocessorHost: System coprocessor com.splicemachine.hbase.RegionSizeEndpoint was loaded successfully with priority (536870915).
2016-12-02 19:14:05,259 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-0] coprocessor.CoprocessorHost: System coprocessor com.splicemachine.si.data.hbase.coprocessor.TxnLifecycleEndpoint was loaded successfully with priority (536870916).
2016-12-02 19:14:05,259 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-0] coprocessor.CoprocessorHost: System coprocessor com.splicemachine.si.data.hbase.coprocessor.SIObserver was loaded successfully with priority (536870917).
2016-12-02 19:14:05,259 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-0] coprocessor.CoprocessorHost: System coprocessor com.splicemachine.hbase.BackupEndpointObserver was loaded successfully with priority (536870918).
2016-12-02 19:14:05,269 INFO  [StoreOpener-91b7a8c3c77f908c0cff6940ccfd6d7d-1] hfile.CacheConfig: blockCache=LruBlockCache{blockCount=0, currentSize=827112, freeSize=804479256, maxSize=805306368, heapSize=827112, minSize=765041024, minFactor=0.95, multiSize=382520512, multiFactor=0.5, singleSize=191260256, singleFactor=0.25}, cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=true, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false
2016-12-02 19:14:05,269 INFO  [StoreOpener-91b7a8c3c77f908c0cff6940ccfd6d7d-1] compactions.CompactionConfiguration: size [16777216, 260046848); files [5, 10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000
2016-12-02 19:14:05,283 INFO  [StoreOpener-c161d85dba1212e9f283596f80de56c8-1] hfile.CacheConfig: blockCache=LruBlockCache{blockCount=0, currentSize=827112, freeSize=804479256, maxSize=805306368, heapSize=827112, minSize=765041024, minFactor=0.95, multiSize=382520512, multiFactor=0.5, singleSize=191260256, singleFactor=0.25}, cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=true, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false
2016-12-02 19:14:05,283 INFO  [StoreOpener-c161d85dba1212e9f283596f80de56c8-1] compactions.CompactionConfiguration: size [16777216, 260046848); files [5, 10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000
2016-12-02 19:14:05,295 INFO  [StoreOpener-91b7a8c3c77f908c0cff6940ccfd6d7d-1] hfile.CacheConfig: blockCache=LruBlockCache{blockCount=0, currentSize=827112, freeSize=804479256, maxSize=805306368, heapSize=827112, minSize=765041024, minFactor=0.95, multiSize=382520512, multiFactor=0.5, singleSize=191260256, singleFactor=0.25}, cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=true, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false
2016-12-02 19:14:05,295 INFO  [StoreOpener-91b7a8c3c77f908c0cff6940ccfd6d7d-1] compactions.CompactionConfiguration: size [16777216, 260046848); files [5, 10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000
2016-12-02 19:14:05,308 INFO  [StoreOpener-c161d85dba1212e9f283596f80de56c8-1] hfile.CacheConfig: blockCache=LruBlockCache{blockCount=0, currentSize=827112, freeSize=804479256, maxSize=805306368, heapSize=827112, minSize=765041024, minFactor=0.95, multiSize=382520512, multiFactor=0.5, singleSize=191260256, singleFactor=0.25}, cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=true, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false
2016-12-02 19:14:05,309 INFO  [StoreOpener-c161d85dba1212e9f283596f80de56c8-1] compactions.CompactionConfiguration: size [16777216, 260046848); files [5, 10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000
2016-12-02 19:14:05,347 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-2] regionserver.HRegion: Onlined 91b7a8c3c77f908c0cff6940ccfd6d7d; next sequenceid=2
2016-12-02 19:14:05,361 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-0] regionserver.HRegion: Onlined c161d85dba1212e9f283596f80de56c8; next sequenceid=2
2016-12-02 19:14:05,388 INFO  [PostOpenDeployTasks:c161d85dba1212e9f283596f80de56c8] regionserver.HRegionServer: Post open deploy tasks for splice:SPLICE_TXN,\x0F,1480686242157.c161d85dba1212e9f283596f80de56c8.
2016-12-02 19:14:05,388 INFO  [PostOpenDeployTasks:91b7a8c3c77f908c0cff6940ccfd6d7d] regionserver.HRegionServer: Post open deploy tasks for splice:SPLICE_TXN,\x0A,1480686242157.91b7a8c3c77f908c0cff6940ccfd6d7d.
2016-12-02 19:14:05,395 INFO  [PostOpenDeployTasks:91b7a8c3c77f908c0cff6940ccfd6d7d] hbase.MetaTableAccessor: Updated row splice:SPLICE_TXN,\x0A,1480686242157.91b7a8c3c77f908c0cff6940ccfd6d7d. with server=hscale-dev1-dn1,16020,1480686239118
2016-12-02 19:14:05,396 INFO  [PostOpenDeployTasks:c161d85dba1212e9f283596f80de56c8] hbase.MetaTableAccessor: Updated row splice:SPLICE_TXN,\x0F,1480686242157.c161d85dba1212e9f283596f80de56c8. with server=hscale-dev1-dn1,16020,1480686239118
2016-12-02 19:14:07,679 INFO  [PriorityRpcServer.handler=2,queue=0,port=16020] regionserver.RSRpcServices: Open splice:TENTATIVE_DDL,,1480686246654.7ba2eae2b5a4d7d5a163e2c50aa33a6b.
2016-12-02 19:14:07,725 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-1] coprocessor.CoprocessorHost: System coprocessor org.apache.hadoop.hbase.security.access.SecureBulkLoadEndpoint was loaded successfully with priority (536870911).
2016-12-02 19:14:07,726 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-1] coprocessor.CoprocessorHost: System coprocessor com.splicemachine.hbase.MemstoreAwareObserver was loaded successfully with priority (536870912).
2016-12-02 19:14:07,726 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-1] coprocessor.CoprocessorHost: System coprocessor com.splicemachine.derby.hbase.SpliceIndexObserver was loaded successfully with priority (536870913).
2016-12-02 19:14:07,726 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-1] coprocessor.CoprocessorHost: System coprocessor com.splicemachine.derby.hbase.SpliceIndexEndpoint was loaded successfully with priority (536870914).
2016-12-02 19:14:07,726 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-1] coprocessor.CoprocessorHost: System coprocessor com.splicemachine.hbase.RegionSizeEndpoint was loaded successfully with priority (536870915).
2016-12-02 19:14:07,727 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-1] coprocessor.CoprocessorHost: System coprocessor com.splicemachine.si.data.hbase.coprocessor.TxnLifecycleEndpoint was loaded successfully with priority (536870916).
2016-12-02 19:14:07,735 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-1] coprocessor.CoprocessorHost: System coprocessor com.splicemachine.si.data.hbase.coprocessor.SIObserver was loaded successfully with priority (536870917).
2016-12-02 19:14:07,735 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-1] coprocessor.CoprocessorHost: System coprocessor com.splicemachine.hbase.BackupEndpointObserver was loaded successfully with priority (536870918).
2016-12-02 19:14:07,748 INFO  [StoreOpener-7ba2eae2b5a4d7d5a163e2c50aa33a6b-1] hfile.CacheConfig: blockCache=LruBlockCache{blockCount=0, currentSize=827112, freeSize=804479256, maxSize=805306368, heapSize=827112, minSize=765041024, minFactor=0.95, multiSize=382520512, multiFactor=0.5, singleSize=191260256, singleFactor=0.25}, cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=true, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false
2016-12-02 19:14:07,748 INFO  [StoreOpener-7ba2eae2b5a4d7d5a163e2c50aa33a6b-1] compactions.CompactionConfiguration: size [16777216, 260046848); files [5, 10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000
2016-12-02 19:14:07,783 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-1] regionserver.HRegion: Onlined 7ba2eae2b5a4d7d5a163e2c50aa33a6b; next sequenceid=2
2016-12-02 19:14:07,791 INFO  [PostOpenDeployTasks:7ba2eae2b5a4d7d5a163e2c50aa33a6b] regionserver.HRegionServer: Post open deploy tasks for splice:TENTATIVE_DDL,,1480686246654.7ba2eae2b5a4d7d5a163e2c50aa33a6b.
2016-12-02 19:14:07,804 INFO  [PostOpenDeployTasks:7ba2eae2b5a4d7d5a163e2c50aa33a6b] hbase.MetaTableAccessor: Updated row splice:TENTATIVE_DDL,,1480686246654.7ba2eae2b5a4d7d5a163e2c50aa33a6b. with server=hscale-dev1-dn1,16020,1480686239118
2016-12-02 19:14:12,171 INFO  [PriorityRpcServer.handler=4,queue=0,port=16020] regionserver.RSRpcServices: Open splice:SPLICE_SEQUENCES,,1480686251218.969f9fff45a516829701ab973a11c9ce.
2016-12-02 19:14:12,251 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-0] coprocessor.CoprocessorHost: System coprocessor org.apache.hadoop.hbase.security.access.SecureBulkLoadEndpoint was loaded successfully with priority (536870911).
2016-12-02 19:14:12,251 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-0] coprocessor.CoprocessorHost: System coprocessor com.splicemachine.hbase.MemstoreAwareObserver was loaded successfully with priority (536870912).
2016-12-02 19:14:12,251 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-0] coprocessor.CoprocessorHost: System coprocessor com.splicemachine.derby.hbase.SpliceIndexObserver was loaded successfully with priority (536870913).
2016-12-02 19:14:12,251 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-0] coprocessor.CoprocessorHost: System coprocessor com.splicemachine.derby.hbase.SpliceIndexEndpoint was loaded successfully with priority (536870914).
2016-12-02 19:14:12,252 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-0] coprocessor.CoprocessorHost: System coprocessor com.splicemachine.hbase.RegionSizeEndpoint was loaded successfully with priority (536870915).
2016-12-02 19:14:12,252 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-0] coprocessor.CoprocessorHost: System coprocessor com.splicemachine.si.data.hbase.coprocessor.TxnLifecycleEndpoint was loaded successfully with priority (536870916).
2016-12-02 19:14:12,252 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-0] coprocessor.CoprocessorHost: System coprocessor com.splicemachine.si.data.hbase.coprocessor.SIObserver was loaded successfully with priority (536870917).
2016-12-02 19:14:12,252 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-0] coprocessor.CoprocessorHost: System coprocessor com.splicemachine.hbase.BackupEndpointObserver was loaded successfully with priority (536870918).
2016-12-02 19:14:12,270 INFO  [StoreOpener-969f9fff45a516829701ab973a11c9ce-1] hfile.CacheConfig: blockCache=LruBlockCache{blockCount=0, currentSize=827112, freeSize=804479256, maxSize=805306368, heapSize=827112, minSize=765041024, minFactor=0.95, multiSize=382520512, multiFactor=0.5, singleSize=191260256, singleFactor=0.25}, cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=true, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false
2016-12-02 19:14:12,271 INFO  [StoreOpener-969f9fff45a516829701ab973a11c9ce-1] compactions.CompactionConfiguration: size [16777216, 260046848); files [5, 10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000
2016-12-02 19:14:12,322 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-0] regionserver.HRegion: Onlined 969f9fff45a516829701ab973a11c9ce; next sequenceid=2
2016-12-02 19:14:12,330 INFO  [PostOpenDeployTasks:969f9fff45a516829701ab973a11c9ce] regionserver.HRegionServer: Post open deploy tasks for splice:SPLICE_SEQUENCES,,1480686251218.969f9fff45a516829701ab973a11c9ce.
2016-12-02 19:14:12,343 INFO  [PostOpenDeployTasks:969f9fff45a516829701ab973a11c9ce] hbase.MetaTableAccessor: Updated row splice:SPLICE_SEQUENCES,,1480686251218.969f9fff45a516829701ab973a11c9ce. with server=hscale-dev1-dn1,16020,1480686239118
2016-12-02 19:14:15,444 INFO  [PriorityRpcServer.handler=1,queue=1,port=16020] regionserver.RSRpcServices: Open splice:16,,1480686254027.75c25e3e9e9f5281c8face7c909f0b04.
2016-12-02 19:14:15,515 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-2] coprocessor.CoprocessorHost: System coprocessor org.apache.hadoop.hbase.security.access.SecureBulkLoadEndpoint was loaded successfully with priority (536870911).
2016-12-02 19:14:15,515 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-2] coprocessor.CoprocessorHost: System coprocessor com.splicemachine.hbase.MemstoreAwareObserver was loaded successfully with priority (536870912).
2016-12-02 19:14:15,518 ERROR [RS_OPEN_REGION-hscale-dev1-dn1:16020-2] coprocessor.CoprocessorHost: The coprocessor com.splicemachine.derby.hbase.SpliceIndexObserver threw java.io.IOException: java.lang.IllegalStateException: Unable to register service, startup failed
java.io.IOException: java.lang.IllegalStateException: Unable to register service, startup failed
    at com.splicemachine.derby.hbase.SpliceIndexObserver.start(SpliceIndexObserver.java:141)
    at org.apache.hadoop.hbase.coprocessor.CoprocessorHost$Environment.startup(CoprocessorHost.java:411)
    at org.apache.hadoop.hbase.coprocessor.CoprocessorHost.loadInstance(CoprocessorHost.java:253)
    at org.apache.hadoop.hbase.coprocessor.CoprocessorHost.loadSystemCoprocessors(CoprocessorHost.java:156)
    at org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost.<init>(RegionCoprocessorHost.java:231)
    at org.apache.hadoop.hbase.regionserver.HRegion.<init>(HRegion.java:723)
    at org.apache.hadoop.hbase.regionserver.HRegion.<init>(HRegion.java:631)
    at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
    at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62)
    at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
    at java.lang.reflect.Constructor.newInstance(Constructor.java:422)
    at org.apache.hadoop.hbase.regionserver.HRegion.newHRegion(HRegion.java:5978)
    at org.apache.hadoop.hbase.regionserver.HRegion.openHRegion(HRegion.java:6288)
    at org.apache.hadoop.hbase.regionserver.HRegion.openHRegion(HRegion.java:6260)
    at org.apache.hadoop.hbase.regionserver.HRegion.openHRegion(HRegion.java:6216)
    at org.apache.hadoop.hbase.regionserver.HRegion.openHRegion(HRegion.java:6167)
    at org.apache.hadoop.hbase.regionserver.handler.OpenRegionHandler.openRegion(OpenRegionHandler.java:362)
    at org.apache.hadoop.hbase.regionserver.handler.OpenRegionHandler.process(OpenRegionHandler.java:129)
    at org.apache.hadoop.hbase.executor.EventHandler.run(EventHandler.java:128)
    at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
    at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
    at java.lang.Thread.run(Thread.java:745)
Caused by: java.lang.IllegalStateException: Unable to register service, startup failed
    at com.splicemachine.lifecycle.DatabaseLifecycleManager.registerGeneralService(DatabaseLifecycleManager.java:183)
    at com.splicemachine.derby.hbase.SpliceIndexObserver.start(SpliceIndexObserver.java:139)
    ... 21 more
2016-12-02 19:14:15,519 FATAL [RS_OPEN_REGION-hscale-dev1-dn1:16020-2] regionserver.HRegionServer: ABORTING region server hscale-dev1-dn1,16020,1480686239118: The coprocessor com.splicemachine.derby.hbase.SpliceIndexObserver threw java.io.IOException: java.lang.IllegalStateException: Unable to register service, startup failed
java.io.IOException: java.lang.IllegalStateException: Unable to register service, startup failed
    at com.splicemachine.derby.hbase.SpliceIndexObserver.start(SpliceIndexObserver.java:141)
    at org.apache.hadoop.hbase.coprocessor.CoprocessorHost$Environment.startup(CoprocessorHost.java:411)
    at org.apache.hadoop.hbase.coprocessor.CoprocessorHost.loadInstance(CoprocessorHost.java:253)
    at org.apache.hadoop.hbase.coprocessor.CoprocessorHost.loadSystemCoprocessors(CoprocessorHost.java:156)
    at org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost.<init>(RegionCoprocessorHost.java:231)
    at org.apache.hadoop.hbase.regionserver.HRegion.<init>(HRegion.java:723)
    at org.apache.hadoop.hbase.regionserver.HRegion.<init>(HRegion.java:631)
    at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
    at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62)
    at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
    at java.lang.reflect.Constructor.newInstance(Constructor.java:422)
    at org.apache.hadoop.hbase.regionserver.HRegion.newHRegion(HRegion.java:5978)
    at org.apache.hadoop.hbase.regionserver.HRegion.openHRegion(HRegion.java:6288)
    at org.apache.hadoop.hbase.regionserver.HRegion.openHRegion(HRegion.java:6260)
    at org.apache.hadoop.hbase.regionserver.HRegion.openHRegion(HRegion.java:6216)
    at org.apache.hadoop.hbase.regionserver.HRegion.openHRegion(HRegion.java:6167)
    at org.apache.hadoop.hbase.regionserver.handler.OpenRegionHandler.openRegion(OpenRegionHandler.java:362)
    at org.apache.hadoop.hbase.regionserver.handler.OpenRegionHandler.process(OpenRegionHandler.java:129)
    at org.apache.hadoop.hbase.executor.EventHandler.run(EventHandler.java:128)
    at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
    at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
    at java.lang.Thread.run(Thread.java:745)
Caused by: java.lang.IllegalStateException: Unable to register service, startup failed
    at com.splicemachine.lifecycle.DatabaseLifecycleManager.registerGeneralService(DatabaseLifecycleManager.java:183)
    at com.splicemachine.derby.hbase.SpliceIndexObserver.start(SpliceIndexObserver.java:139)
    ... 21 more
2016-12-02 19:14:15,520 FATAL [RS_OPEN_REGION-hscale-dev1-dn1:16020-2] regionserver.HRegionServer: RegionServer abort: loaded coprocessors are: [com.splicemachine.hbase.RegionSizeEndpoint, com.splicemachine.hbase.MemstoreAwareObserver, com.splicemachine.hbase.BackupEndpointObserver, com.splicemachine.hbase.RegionServerLifecycleObserver, com.splicemachine.derby.hbase.SpliceIndexObserver, com.splicemachine.si.data.hbase.coprocessor.TxnLifecycleEndpoint, com.splicemachine.derby.hbase.SpliceIndexEndpoint, com.splicemachine.si.data.hbase.coprocessor.SIObserver, org.apache.hadoop.hbase.security.access.SecureBulkLoadEndpoint]
2016-12-02 19:14:15,562 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-2] regionserver.HRegionServer: Dump of metrics as JSON on abort: {
  "beans" : [ {
    "name" : "java.lang:type=Memory",
    "modelerType" : "sun.management.MemoryImpl",
    "ObjectPendingFinalizationCount" : 0,
    "HeapMemoryUsage" : {
      "committed" : 3221225472,
      "init" : 3221225472,
      "max" : 3221225472,
      "used" : 287309824
    },
    "NonHeapMemoryUsage" : {
      "committed" : 64643072,
      "init" : 2555904,
      "max" : -1,
      "used" : 62405432
    },
    "Verbose" : true,
    "ObjectName" : "java.lang:type=Memory"
  } ],
  "beans" : [ {
    "name" : "Hadoop:service=HBase,name=RegionServer,sub=IPC",
    "modelerType" : "RegionServer,sub=IPC",
    "tag.Context" : "regionserver",
    "tag.Hostname" : "hscale-dev1-dn1"
  } ],
  "beans" : [ {
    "name" : "Hadoop:service=HBase,name=RegionServer,sub=Replication",
    "modelerType" : "RegionServer,sub=Replication",
    "tag.Context" : "regionserver",
    "tag.Hostname" : "hscale-dev1-dn1"
  } ],
  "beans" : [ {
    "name" : "Hadoop:service=HBase,name=RegionServer,sub=Server",
    "modelerType" : "RegionServer,sub=Server",
    "tag.Context" : "regionserver",
    "tag.Hostname" : "hscale-dev1-dn1"
  } ]
}
2016-12-02 19:14:15,571 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-2] regionserver.HRegionServer: STOPPED: The coprocessor com.splicemachine.derby.hbase.SpliceIndexObserver threw java.io.IOException: java.lang.IllegalStateException: Unable to register service, startup failed
2016-12-02 19:14:15,571 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] regionserver.SplitLogWorker: Sending interrupt to stop the worker thread
2016-12-02 19:14:15,572 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] regionserver.HRegionServer: Stopping infoServer
2016-12-02 19:14:15,572 INFO  [SplitLogWorker-hscale-dev1-dn1:16020] regionserver.SplitLogWorker: SplitLogWorker interrupted. Exiting.
2016-12-02 19:14:15,572 ERROR [RS_OPEN_REGION-hscale-dev1-dn1:16020-2] coprocessor.CoprocessorHost: The coprocessor com.splicemachine.derby.hbase.SpliceIndexEndpoint threw java.io.IOException: java.lang.IllegalStateException: Unable to register service, startup failed
java.io.IOException: java.lang.IllegalStateException: Unable to register service, startup failed
    at com.splicemachine.derby.hbase.SpliceIndexEndpoint.start(SpliceIndexEndpoint.java:125)
    at org.apache.hadoop.hbase.coprocessor.CoprocessorHost$Environment.startup(CoprocessorHost.java:411)
    at org.apache.hadoop.hbase.coprocessor.CoprocessorHost.loadInstance(CoprocessorHost.java:253)
    at org.apache.hadoop.hbase.coprocessor.CoprocessorHost.loadSystemCoprocessors(CoprocessorHost.java:156)
    at org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost.<init>(RegionCoprocessorHost.java:231)
    at org.apache.hadoop.hbase.regionserver.HRegion.<init>(HRegion.java:723)
    at org.apache.hadoop.hbase.regionserver.HRegion.<init>(HRegion.java:631)
    at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
    at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62)
    at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
    at java.lang.reflect.Constructor.newInstance(Constructor.java:422)
    at org.apache.hadoop.hbase.regionserver.HRegion.newHRegion(HRegion.java:5978)
    at org.apache.hadoop.hbase.regionserver.HRegion.openHRegion(HRegion.java:6288)
    at org.apache.hadoop.hbase.regionserver.HRegion.openHRegion(HRegion.java:6260)
    at org.apache.hadoop.hbase.regionserver.HRegion.openHRegion(HRegion.java:6216)
    at org.apache.hadoop.hbase.regionserver.HRegion.openHRegion(HRegion.java:6167)
    at org.apache.hadoop.hbase.regionserver.handler.OpenRegionHandler.openRegion(OpenRegionHandler.java:362)
    at org.apache.hadoop.hbase.regionserver.handler.OpenRegionHandler.process(OpenRegionHandler.java:129)
    at org.apache.hadoop.hbase.executor.EventHandler.run(EventHandler.java:128)
    at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
    at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
    at java.lang.Thread.run(Thread.java:745)
Caused by: java.lang.IllegalStateException: Unable to register service, startup failed
    at com.splicemachine.lifecycle.DatabaseLifecycleManager.registerGeneralService(DatabaseLifecycleManager.java:183)
    at com.splicemachine.derby.hbase.SpliceIndexEndpoint.start(SpliceIndexEndpoint.java:123)
    ... 21 more
2016-12-02 19:14:15,572 INFO  [SplitLogWorker-hscale-dev1-dn1:16020] regionserver.SplitLogWorker: SplitLogWorker hscale-dev1-dn1,16020,1480686239118 exiting
2016-12-02 19:14:15,572 FATAL [RS_OPEN_REGION-hscale-dev1-dn1:16020-2] regionserver.HRegionServer: ABORTING region server hscale-dev1-dn1,16020,1480686239118: The coprocessor com.splicemachine.derby.hbase.SpliceIndexEndpoint threw java.io.IOException: java.lang.IllegalStateException: Unable to register service, startup failed
java.io.IOException: java.lang.IllegalStateException: Unable to register service, startup failed
    at com.splicemachine.derby.hbase.SpliceIndexEndpoint.start(SpliceIndexEndpoint.java:125)
    at org.apache.hadoop.hbase.coprocessor.CoprocessorHost$Environment.startup(CoprocessorHost.java:411)
    at org.apache.hadoop.hbase.coprocessor.CoprocessorHost.loadInstance(CoprocessorHost.java:253)
    at org.apache.hadoop.hbase.coprocessor.CoprocessorHost.loadSystemCoprocessors(CoprocessorHost.java:156)
    at org.apache.hadoop.hbase.regionserver.RegionCoprocessorHost.<init>(RegionCoprocessorHost.java:231)
    at org.apache.hadoop.hbase.regionserver.HRegion.<init>(HRegion.java:723)
    at org.apache.hadoop.hbase.regionserver.HRegion.<init>(HRegion.java:631)
    at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
    at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62)
    at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
    at java.lang.reflect.Constructor.newInstance(Constructor.java:422)
    at org.apache.hadoop.hbase.regionserver.HRegion.newHRegion(HRegion.java:5978)
    at org.apache.hadoop.hbase.regionserver.HRegion.openHRegion(HRegion.java:6288)
    at org.apache.hadoop.hbase.regionserver.HRegion.openHRegion(HRegion.java:6260)
    at org.apache.hadoop.hbase.regionserver.HRegion.openHRegion(HRegion.java:6216)
    at org.apache.hadoop.hbase.regionserver.HRegion.openHRegion(HRegion.java:6167)
    at org.apache.hadoop.hbase.regionserver.handler.OpenRegionHandler.openRegion(OpenRegionHandler.java:362)
    at org.apache.hadoop.hbase.regionserver.handler.OpenRegionHandler.process(OpenRegionHandler.java:129)
    at org.apache.hadoop.hbase.executor.EventHandler.run(EventHandler.java:128)
    at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
    at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
    at java.lang.Thread.run(Thread.java:745)
Caused by: java.lang.IllegalStateException: Unable to register service, startup failed
    at com.splicemachine.lifecycle.DatabaseLifecycleManager.registerGeneralService(DatabaseLifecycleManager.java:183)
    at com.splicemachine.derby.hbase.SpliceIndexEndpoint.start(SpliceIndexEndpoint.java:123)
    ... 21 more
2016-12-02 19:14:15,573 FATAL [RS_OPEN_REGION-hscale-dev1-dn1:16020-2] regionserver.HRegionServer: RegionServer abort: loaded coprocessors are: [com.splicemachine.hbase.RegionSizeEndpoint, com.splicemachine.hbase.MemstoreAwareObserver, com.splicemachine.hbase.BackupEndpointObserver, com.splicemachine.hbase.RegionServerLifecycleObserver, com.splicemachine.derby.hbase.SpliceIndexObserver, com.splicemachine.si.data.hbase.coprocessor.TxnLifecycleEndpoint, com.splicemachine.derby.hbase.SpliceIndexEndpoint, com.splicemachine.si.data.hbase.coprocessor.SIObserver, org.apache.hadoop.hbase.security.access.SecureBulkLoadEndpoint]
2016-12-02 19:14:15,573 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] mortbay.log: Stopped SelectChannelConnector@0.0.0.0:16030
2016-12-02 19:14:15,573 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] regionserver.HeapMemoryManager: Stoping HeapMemoryTuner chore.
2016-12-02 19:14:15,573 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] flush.RegionServerFlushTableProcedureManager: Stopping region server flush procedure manager abruptly.
2016-12-02 19:14:15,573 INFO  [MemStoreFlusher.0] regionserver.MemStoreFlusher: MemStoreFlusher.0 exiting
2016-12-02 19:14:15,575 INFO  [MemStoreFlusher.1] regionserver.MemStoreFlusher: MemStoreFlusher.1 exiting
2016-12-02 19:14:15,574 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-2] regionserver.HRegionServer: Dump of metrics as JSON on abort: {
  "beans" : [ {
    "name" : "java.lang:type=Memory",
    "modelerType" : "sun.management.MemoryImpl",
    "ObjectPendingFinalizationCount" : 0,
    "HeapMemoryUsage" : {
      "committed" : 3221225472,
      "init" : 3221225472,
      "max" : 3221225472,
      "used" : 288358400
    },
    "NonHeapMemoryUsage" : {
      "committed" : 64905216,
      "init" : 2555904,
      "max" : -1,
      "used" : 62501936
    },
    "Verbose" : true,
    "ObjectName" : "java.lang:type=Memory"
  } ],
  "beans" : [ {
    "name" : "Hadoop:service=HBase,name=RegionServer,sub=IPC",
    "modelerType" : "RegionServer,sub=IPC",
    "tag.Context" : "regionserver",
    "tag.Hostname" : "hscale-dev1-dn1"
  } ],
  "beans" : [ {
    "name" : "Hadoop:service=HBase,name=RegionServer,sub=Replication",
    "modelerType" : "RegionServer,sub=Replication",
    "tag.Context" : "regionserver",
    "tag.Hostname" : "hscale-dev1-dn1"
  } ],
  "beans" : [ {
    "name" : "Hadoop:service=HBase,name=RegionServer,sub=Server",
    "modelerType" : "RegionServer,sub=Server",
    "tag.Context" : "regionserver",
    "tag.Hostname" : "hscale-dev1-dn1"
  } ]
}
2016-12-02 19:14:15,573 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] snapshot.RegionServerSnapshotManager: Stopping RegionServerSnapshotManager abruptly.
2016-12-02 19:14:15,579 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-2] coprocessor.CoprocessorHost: System coprocessor com.splicemachine.hbase.RegionSizeEndpoint was loaded successfully with priority (536870913).
2016-12-02 19:14:15,581 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-2] coprocessor.CoprocessorHost: System coprocessor com.splicemachine.si.data.hbase.coprocessor.TxnLifecycleEndpoint was loaded successfully with priority (536870914).
2016-12-02 19:14:15,581 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-2] coprocessor.CoprocessorHost: System coprocessor com.splicemachine.si.data.hbase.coprocessor.SIObserver was loaded successfully with priority (536870915).
2016-12-02 19:14:15,581 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-2] coprocessor.CoprocessorHost: System coprocessor com.splicemachine.hbase.BackupEndpointObserver was loaded successfully with priority (536870916).
2016-12-02 19:14:15,585 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] regionserver.HRegionServer: aborting server hscale-dev1-dn1,16020,1480686239118
2016-12-02 19:14:15,586 INFO  [StoreCloserThread-splice:SPLICE_TXN,,1480686242157.befbb055ce59b866459081d27215d354.-1] regionserver.HStore: Closed P
2016-12-02 19:14:15,586 INFO  [StoreCloserThread-splice:SPLICE_TXN,\x0A,1480686242157.91b7a8c3c77f908c0cff6940ccfd6d7d.-1] regionserver.HStore: Closed P
2016-12-02 19:14:15,593 INFO  [StoreCloserThread-splice:SPLICE_TXN,,1480686242157.befbb055ce59b866459081d27215d354.-1] regionserver.HStore: Closed V
2016-12-02 19:14:15,593 INFO  [StoreCloserThread-splice:SPLICE_TXN,\x0A,1480686242157.91b7a8c3c77f908c0cff6940ccfd6d7d.-1] regionserver.HStore: Closed V
2016-12-02 19:14:15,593 INFO  [StoreCloserThread-splice:SPLICE_TXN,\x09,1480686242157.ace1654b29a8315feb0793d463b78246.-1] regionserver.HStore: Closed P
2016-12-02 19:14:15,593 INFO  [StoreCloserThread-splice:SPLICE_TXN,\x09,1480686242157.ace1654b29a8315feb0793d463b78246.-1] regionserver.HStore: Closed V
2016-12-02 19:14:15,593 INFO  [RS_CLOSE_REGION-hscale-dev1-dn1:16020-2] coprocessor.TxnLifecycleEndpoint: Shutting down TxnLifecycleEndpoint
2016-12-02 19:14:15,593 INFO  [StoreOpener-75c25e3e9e9f5281c8face7c909f0b04-1] hfile.CacheConfig: blockCache=LruBlockCache{blockCount=0, currentSize=827112, freeSize=804479256, maxSize=805306368, heapSize=827112, minSize=765041024, minFactor=0.95, multiSize=382520512, multiFactor=0.5, singleSize=191260256, singleFactor=0.25}, cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=true, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false
2016-12-02 19:14:15,593 INFO  [StoreOpener-75c25e3e9e9f5281c8face7c909f0b04-1] compactions.CompactionConfiguration: size [16777216, 260046848); files [5, 10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000
2016-12-02 19:14:15,594 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] client.ConnectionManager$HConnectionImplementation: Closing zookeeper sessionid=0x158bf3f64e00029
2016-12-02 19:14:15,594 INFO  [RS_CLOSE_REGION-hscale-dev1-dn1:16020-2] regionserver.HRegion: Closed splice:SPLICE_TXN,\x0A,1480686242157.91b7a8c3c77f908c0cff6940ccfd6d7d.
2016-12-02 19:14:15,593 INFO  [RS_CLOSE_REGION-hscale-dev1-dn1:16020-0] coprocessor.TxnLifecycleEndpoint: Shutting down TxnLifecycleEndpoint
2016-12-02 19:14:15,597 INFO  [RS_CLOSE_REGION-hscale-dev1-dn1:16020-0] regionserver.HRegion: Closed splice:SPLICE_TXN,,1480686242157.befbb055ce59b866459081d27215d354.
2016-12-02 19:14:15,593 INFO  [RS_CLOSE_REGION-hscale-dev1-dn1:16020-1] coprocessor.TxnLifecycleEndpoint: Shutting down TxnLifecycleEndpoint
2016-12-02 19:14:15,598 INFO  [RS_CLOSE_REGION-hscale-dev1-dn1:16020-1] regionserver.HRegion: Closed splice:SPLICE_TXN,\x09,1480686242157.ace1654b29a8315feb0793d463b78246.
2016-12-02 19:14:15,598 INFO  [StoreCloserThread-splice:TENTATIVE_DDL,,1480686246654.7ba2eae2b5a4d7d5a163e2c50aa33a6b.-1] regionserver.HStore: Closed V
2016-12-02 19:14:15,598 INFO  [RS_CLOSE_REGION-hscale-dev1-dn1:16020-2] coprocessor.TxnLifecycleEndpoint: Shutting down TxnLifecycleEndpoint
2016-12-02 19:14:15,598 INFO  [RS_CLOSE_REGION-hscale-dev1-dn1:16020-2] regionserver.HRegion: Closed splice:TENTATIVE_DDL,,1480686246654.7ba2eae2b5a4d7d5a163e2c50aa33a6b.
2016-12-02 19:14:15,600 INFO  [StoreCloserThread-splice:SPLICE_TXN,\x0F,1480686242157.c161d85dba1212e9f283596f80de56c8.-1] regionserver.HStore: Closed P
2016-12-02 19:14:15,600 INFO  [StoreCloserThread-splice:SPLICE_TXN,\x0F,1480686242157.c161d85dba1212e9f283596f80de56c8.-1] regionserver.HStore: Closed V
2016-12-02 19:14:15,600 INFO  [StoreCloserThread-splice:SPLICE_SEQUENCES,,1480686251218.969f9fff45a516829701ab973a11c9ce.-1] regionserver.HStore: Closed V
2016-12-02 19:14:15,600 INFO  [RS_CLOSE_REGION-hscale-dev1-dn1:16020-0] coprocessor.TxnLifecycleEndpoint: Shutting down TxnLifecycleEndpoint
2016-12-02 19:14:15,601 INFO  [RS_CLOSE_REGION-hscale-dev1-dn1:16020-0] regionserver.HRegion: Closed splice:SPLICE_TXN,\x0F,1480686242157.c161d85dba1212e9f283596f80de56c8.
2016-12-02 19:14:15,601 INFO  [StoreCloserThread-splice:SPLICE_TXN,\x07,1480686242157.c3cc35c8ce6fe046e78dbb491c804d1d.-1] regionserver.HStore: Closed P
2016-12-02 19:14:15,601 INFO  [StoreCloserThread-splice:SPLICE_TXN,\x07,1480686242157.c3cc35c8ce6fe046e78dbb491c804d1d.-1] regionserver.HStore: Closed V
2016-12-02 19:14:15,601 INFO  [RS_CLOSE_REGION-hscale-dev1-dn1:16020-1] coprocessor.TxnLifecycleEndpoint: Shutting down TxnLifecycleEndpoint
2016-12-02 19:14:15,601 INFO  [RS_CLOSE_REGION-hscale-dev1-dn1:16020-1] regionserver.HRegion: Closed splice:SPLICE_TXN,\x07,1480686242157.c3cc35c8ce6fe046e78dbb491c804d1d.
2016-12-02 19:14:15,602 INFO  [RS_CLOSE_REGION-hscale-dev1-dn1:16020-2] coprocessor.TxnLifecycleEndpoint: Shutting down TxnLifecycleEndpoint
2016-12-02 19:14:15,602 INFO  [RS_CLOSE_REGION-hscale-dev1-dn1:16020-2] regionserver.HRegion: Closed splice:SPLICE_SEQUENCES,,1480686251218.969f9fff45a516829701ab973a11c9ce.
2016-12-02 19:14:15,616 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] zookeeper.ZooKeeper: Session: 0x158bf3f64e00029 closed
2016-12-02 19:14:15,616 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020-EventThread] zookeeper.ClientCnxn: EventThread shut down
2016-12-02 19:14:15,627 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] regionserver.HRegionServer: stopping server hscale-dev1-dn1,16020,1480686239118; all regions closed.
2016-12-02 19:14:15,689 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-2] regionserver.HRegion: Onlined 75c25e3e9e9f5281c8face7c909f0b04; next sequenceid=5
2016-12-02 19:14:15,698 INFO  [StoreCloserThread-splice:16,,1480686254027.75c25e3e9e9f5281c8face7c909f0b04.-1] regionserver.HStore: Closed V
2016-12-02 19:14:15,801 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-2] coprocessor.TxnLifecycleEndpoint: Shutting down TxnLifecycleEndpoint
2016-12-02 19:14:15,802 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-2] regionserver.HRegion: Closed splice:16,,1480686254027.75c25e3e9e9f5281c8face7c909f0b04.
2016-12-02 19:14:15,802 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-2] coordination.ZkOpenRegionCoordination: Opening of region {ENCODED => 75c25e3e9e9f5281c8face7c909f0b04, NAME => 'splice:16,,1480686254027.75c25e3e9e9f5281c8face7c909f0b04.', STARTKEY => '', ENDKEY => ''} failed, transitioning from OPENING to FAILED_OPEN in ZK, expecting version 4
2016-12-02 19:14:16,039 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] regionserver.Leases: regionserver/hscale-dev1-dn1/10.60.70.11:16020 closing leases
2016-12-02 19:14:16,039 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] regionserver.Leases: regionserver/hscale-dev1-dn1/10.60.70.11:16020 closed leases
2016-12-02 19:14:16,039 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] hbase.ChoreService: Chore service for: hscale-dev1-dn1,16020,1480686239118 had [[ScheduledChore: Name: MovedRegionsCleaner for region hscale-dev1-dn1,16020,1480686239118 Period: 120000 Unit: MILLISECONDS], [ScheduledChore: Name: hscale-dev1-dn1,16020,1480686239118-MemstoreFlusherChore Period: 10000 Unit: MILLISECONDS]] on shutdown
2016-12-02 19:14:22,911 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020.logRoller] regionserver.LogRoller: LogRoller exiting.
2016-12-02 19:14:22,912 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] regionserver.CompactSplitThread: Waiting for Split Thread to finish...
2016-12-02 19:14:22,912 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] regionserver.CompactSplitThread: Waiting for Merge Thread to finish...
2016-12-02 19:14:22,912 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] regionserver.CompactSplitThread: Waiting for Large Compaction Thread to finish...
2016-12-02 19:14:22,912 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] regionserver.CompactSplitThread: Waiting for Small Compaction Thread to finish...
2016-12-02 19:14:22,912 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020.leaseChecker] regionserver.Leases: regionserver/hscale-dev1-dn1/10.60.70.11:16020.leaseChecker closing leases
2016-12-02 19:14:22,912 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020.leaseChecker] regionserver.Leases: regionserver/hscale-dev1-dn1/10.60.70.11:16020.leaseChecker closed leases
2016-12-02 19:14:23,041 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] ipc.RpcServer: Stopping server on 16020
2016-12-02 19:14:23,042 INFO  [RpcServer.listener,port=16020] ipc.RpcServer: RpcServer.listener,port=16020: stopping
2016-12-02 19:14:23,043 INFO  [RpcServer.responder] ipc.RpcServer: RpcServer.responder: stopped
2016-12-02 19:14:23,043 INFO  [RpcServer.responder] ipc.RpcServer: RpcServer.responder: stopping
2016-12-02 19:14:23,096 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] zookeeper.ZooKeeper: Session: 0x358bf3f651f001c closed
2016-12-02 19:14:23,096 INFO  [main-EventThread] zookeeper.ClientCnxn: EventThread shut down
2016-12-02 19:14:23,097 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] regionserver.HRegionServer: stopping server hscale-dev1-dn1,16020,1480686239118; zookeeper connection closed.
2016-12-02 19:14:23,097 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] regionserver.HRegionServer: regionserver/hscale-dev1-dn1/10.60.70.11:16020 exiting
2016-12-02 19:14:23,097 ERROR [main] regionserver.HRegionServerCommandLine: Region server exiting
java.lang.RuntimeException: HRegionServer Aborted
    at org.apache.hadoop.hbase.regionserver.HRegionServerCommandLine.start(HRegionServerCommandLine.java:68)
    at org.apache.hadoop.hbase.regionserver.HRegionServerCommandLine.run(HRegionServerCommandLine.java:87)
    at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:76)
    at org.apache.hadoop.hbase.util.ServerCommandLine.doMain(ServerCommandLine.java:126)
    at org.apache.hadoop.hbase.regionserver.HRegionServer.main(HRegionServer.java:2651)
2016-12-02 19:14:23,100 INFO  [Thread-8] regionserver.ShutdownHook: Shutdown hook starting; hbase.shutdown.hook=true; fsShutdownHook=org.apache.hadoop.fs.FileSystem$Cache$ClientFinalizer@77b325b3
2016-12-02 19:14:23,100 INFO  [Thread-8] regionserver.ShutdownHook: Starting fs shutdown hook thread.
2016-12-02 19:14:23,101 INFO  [Thread-8] regionserver.ShutdownHook: Shutdown hook finished.
Fri Dec  2 19:20:15 IST 2016 Starting regionserver on hscale-dev1-dn1
core file size          (blocks, -c) 0
data seg size           (kbytes, -d) unlimited
scheduling priority             (-e) 0
file size               (blocks, -f) unlimited
pending signals                 (-i) 62057
max locked memory       (kbytes, -l) 64
max memory size         (kbytes, -m) unlimited
open files                      (-n) 32000
pipe size            (512 bytes, -p) 8
POSIX message queues     (bytes, -q) 819200
real-time priority              (-r) 0
stack size              (kbytes, -s) 10240
cpu time               (seconds, -t) unlimited
max user processes              (-u) 16000
virtual memory          (kbytes, -v) unlimited
file locks                      (-x) unlimited
2016-12-02 19:20:16,694 INFO  [main] util.VersionInfo: HBase 1.1.2.2.4.2.0-258
2016-12-02 19:20:16,695 INFO  [main] util.VersionInfo: Source code repository file:///grid/0/jenkins/workspace/HDP-build-centos6/bigtop/build/hbase/rpm/BUILD/hbase-1.1.2.2.4.2.0 revision=Unknown
2016-12-02 19:20:16,695 INFO  [main] util.VersionInfo: Compiled by jenkins on Mon Apr 25 06:36:21 UTC 2016
2016-12-02 19:20:16,695 INFO  [main] util.VersionInfo: From source with checksum 4f661ee4f9f148ce7bfcad5b0d667c27
2016-12-02 19:20:16,972 INFO  [main] util.ServerCommandLine: env:PATH=/usr/sbin:/sbin:/usr/lib/ambari-server/*:/usr/lib64/qt-3.3/bin:/usr/local/sbin:/usr/local/bin:/sbin:/bin:/usr/sbin:/usr/bin:/usr/jdk64/jdk1.8.0_60/bin:/usr/jdk64/jdk1.8.0_60/jre/bin:/root/bin:/var/lib/ambari-agent
2016-12-02 19:20:16,972 INFO  [main] util.ServerCommandLine: env:HISTCONTROL=ignoredups
2016-12-02 19:20:16,972 INFO  [main] util.ServerCommandLine: env:HBASE_PID_DIR=/var/run/hbase
2016-12-02 19:20:16,973 INFO  [main] util.ServerCommandLine: env:HBASE_REGIONSERVER_OPTS= -Xmn512m -Xms3072m -Xmx3072m  -XX:+HeapDumpOnOutOfMemoryError -XX:MaxDirectMemorySize=2g -XX:+AlwaysPreTouch -XX:+UseG1GC -XX:MaxNewSize=4g -XX:InitiatingHeapOccupancyPercent=60 -XX:ParallelGCThreads=24 -XX:+ParallelRefProcEnabled -XX:MaxGCPauseMillis=5000 -Dcom.sun.management.jmxremote.authenticate=false -Dcom.sun.management.jmxremote.ssl=false -Dcom.sun.management.jmxremote.port=10102
2016-12-02 19:20:16,973 INFO  [main] util.ServerCommandLine: env:HBASE_CONF_DIR=/usr/hdp/current/hbase-regionserver/conf
2016-12-02 19:20:16,973 INFO  [main] util.ServerCommandLine: env:JRE_HOME=/usr/jdk64/jdk1.8.0_60/jre
2016-12-02 19:20:16,973 INFO  [main] util.ServerCommandLine: env:MAIL=/var/spool/mail/hbase
2016-12-02 19:20:16,973 INFO  [main] util.ServerCommandLine: env:LD_LIBRARY_PATH=::/usr/hdp/2.4.2.0-258/hadoop/lib/native/Linux-amd64-64:/usr/hdp/2.4.2.0-258/hadoop/lib/native
2016-12-02 19:20:16,973 INFO  [main] util.ServerCommandLine: env:LOGNAME=hbase
2016-12-02 19:20:16,973 INFO  [main] util.ServerCommandLine: env:HBASE_REST_OPTS=
2016-12-02 19:20:16,973 INFO  [main] util.ServerCommandLine: env:PWD=/home/hbase
2016-12-02 19:20:16,973 INFO  [main] util.ServerCommandLine: env:HBASE_ROOT_LOGGER=INFO,RFA
2016-12-02 19:20:16,973 INFO  [main] util.ServerCommandLine: env:LESSOPEN=||/usr/bin/lesspipe.sh %s
2016-12-02 19:20:16,973 INFO  [main] util.ServerCommandLine: env:SHELL=/bin/bash
2016-12-02 19:20:16,973 INFO  [main] util.ServerCommandLine: env:HBASE_ENV_INIT=true
2016-12-02 19:20:16,973 INFO  [main] util.ServerCommandLine: env:QTINC=/usr/lib64/qt-3.3/include
2016-12-02 19:20:16,973 INFO  [main] util.ServerCommandLine: env:HBASE_MASTER_OPTS= -Xms4096m -Xmx4096m  -XX:+HeapDumpOnOutOfMemoryError -XX:MaxDirectMemorySize=2g -XX:+AlwaysPreTouch -XX:+UseParNewGC -XX:+UseConcMarkSweepGC -XX:CMSInitiatingOccupancyFraction=70 -XX:+CMSParallelRemarkEnabled -Dcom.sun.management.jmxremote.authenticate=false -Dcom.sun.management.jmxremote.ssl=false -Dcom.sun.management.jmxremote.port=10101 -Dsplice.spark.enabled=true -Dsplice.spark.app.name=SpliceMachine -Dsplice.spark.master=yarn-client -Dsplice.spark.logConf=true -Dsplice.spark.broadcast.factory=org.apache.spark.broadcast.HttpBroadcastFactory -Dsplice.spark.driver.maxResultSize=1g -Dsplice.spark.driver.memory=1g -Dsplice.spark.dynamicAllocation.enabled=true -Dsplice.spark.dynamicAllocation.executorIdleTimeout=600 -Dsplice.spark.dynamicAllocation.minExecutors=0 -Dsplice.spark.io.compression.lz4.blockSize=32k -Dsplice.spark.kryo.referenceTracking=false -Dsplice.spark.kryo.registrator=com.splicemachine.derby.impl.SpliceSparkKryoRegistrator -Dsplice.spark.kryoserializer.buffer.max=512m -Dsplice.spark.kryoserializer.buffer=4m -Dsplice.spark.locality.wait=100 -Dsplice.spark.scheduler.mode=FAIR -Dsplice.spark.serializer=org.apache.spark.serializer.KryoSerializer -Dsplice.spark.shuffle.compress=false -Dsplice.spark.shuffle.file.buffer=128k -Dsplice.spark.shuffle.memoryFraction=0.7 -Dsplice.spark.shuffle.service.enabled=true -Dsplice.spark.storage.memoryFraction=0.1 -Dsplice.spark.yarn.am.extraLibraryPath=/usr/hdp/current/hadoop-client/lib/native -Dsplice.spark.yarn.am.waitTime=10s -Dsplice.spark.yarn.executor.memoryOverhead=2048 -Dsplice.spark.driver.extraJavaOptions=-Dlog4j.configuration=file:/etc/spark/conf/log4j.properties -Dsplice.spark.driver.extraLibraryPath=/usr/hdp/current/hadoop-client/lib/native -Dsplice.spark.driver.extraClassPath=/usr/hdp/current/hbase-regionserver/conf:/usr/hdp/current/hbase-regionserver/lib/htrace-core-3.1.0-incubating.jar -Dsplice.spark.executor.extraJavaOptions=-Dlog4j.configuration=file:/etc/spark/conf/log4j.properties -Dsplice.spark.executor.extraLibraryPath=/usr/hdp/current/hadoop-client/lib/native -Dsplice.spark.executor.extraClassPath=/usr/hdp/current/hbase-regionserver/conf:/usr/hdp/current/hbase-regionserver/lib/htrace-core-3.1.0-incubating.jar -Dsplice.spark.ui.retainedJobs=100 -Dsplice.spark.ui.retainedStages=100 -Dsplice.spark.worker.ui.retainedExecutors=100 -Dsplice.spark.worker.ui.retainedDrivers=100 -Dsplice.spark.streaming.ui.retainedBatches=100 -Dsplice.spark.executor.cores=4 -Dsplice.spark.executor.memory=8g -Dspark.compaction.reserved.slots=4 -Dsplice.spark.eventLog.enabled=true -Dsplice.spark.eventLog.dir=hdfs:///user/splice/history -Dsplice.spark.local.dir=/diska/tmp,/diskb/tmp,/diskc/tmp,/diskd/tmp
2016-12-02 19:20:16,973 INFO  [main] util.ServerCommandLine: env:HBASE_MANAGES_ZK=false
2016-12-02 19:20:16,973 INFO  [main] util.ServerCommandLine: env:HBASE_REGIONSERVERS=/usr/hdp/current/hbase-regionserver/conf/regionservers
2016-12-02 19:20:16,973 INFO  [main] util.ServerCommandLine: env:HADOOP_HOME=/usr/hdp/2.4.2.0-258/hadoop
2016-12-02 19:20:16,973 INFO  [main] util.ServerCommandLine: env:HBASE_NICENESS=0
2016-12-02 19:20:16,973 INFO  [main] util.ServerCommandLine: env:HBASE_OPTS=-Dhdp.version=2.4.2.0-258  -XX:ErrorFile=/var/log/hbase/hs_err_pid%p.log -Djava.io.tmpdir=/tmp -verbose:gc -XX:+PrintGCDetails -XX:+PrintGCDateStamps -Xloggc:/var/log/hbase/gc.log-201612021920  -Xmn512m -Xms3072m -Xmx3072m  -XX:+HeapDumpOnOutOfMemoryError -XX:MaxDirectMemorySize=2g -XX:+AlwaysPreTouch -XX:+UseG1GC -XX:MaxNewSize=4g -XX:InitiatingHeapOccupancyPercent=60 -XX:ParallelGCThreads=24 -XX:+ParallelRefProcEnabled -XX:MaxGCPauseMillis=5000 -Dcom.sun.management.jmxremote.authenticate=false -Dcom.sun.management.jmxremote.ssl=false -Dcom.sun.management.jmxremote.port=10102  -Dhbase.log.dir=/var/log/hbase -Dhbase.log.file=hbase-hbase-regionserver-hscale-dev1-dn1.log -Dhbase.home.dir=/usr/hdp/current/hbase-regionserver/bin/.. -Dhbase.id.str=hbase -Dhbase.root.logger=INFO,RFA -Djava.library.path=:/usr/hdp/2.4.2.0-258/hadoop/lib/native/Linux-amd64-64:/usr/hdp/2.4.2.0-258/hadoop/lib/native -Dhbase.security.logger=INFO,RFAS
2016-12-02 19:20:16,973 INFO  [main] util.ServerCommandLine: env:HBASE_START_FILE=/var/run/hbase/hbase-hbase-regionserver.autorestart
2016-12-02 19:20:16,973 INFO  [main] util.ServerCommandLine: env:HBASE_SECURITY_LOGGER=INFO,RFAS
2016-12-02 19:20:16,973 INFO  [main] util.ServerCommandLine: env:SHLVL=3
2016-12-02 19:20:16,973 INFO  [main] util.ServerCommandLine: env:ZOOKEEPER_HOME=/usr/hdp/2.4.2.0-258/zookeeper
2016-12-02 19:20:16,973 INFO  [main] util.ServerCommandLine: env:HBASE_LOGFILE=hbase-hbase-regionserver-hscale-dev1-dn1.log
2016-12-02 19:20:16,973 INFO  [main] util.ServerCommandLine: env:HISTSIZE=1000
2016-12-02 19:20:16,973 INFO  [main] util.ServerCommandLine: env:JAVA_HOME=/usr/jdk64/jdk1.8.0_60
2016-12-02 19:20:16,973 INFO  [main] util.ServerCommandLine: env:HDP_VERSION=2.4.2.0-258
2016-12-02 19:20:16,973 INFO  [main] util.ServerCommandLine: env:XFILESEARCHPATH=/usr/dt/app-defaults/%L/Dt
2016-12-02 19:20:16,974 INFO  [main] util.ServerCommandLine: env:LANG=en_US.UTF-8
2016-12-02 19:20:16,974 INFO  [main] util.ServerCommandLine: env:G_BROKEN_FILENAMES=1
2016-12-02 19:20:16,974 INFO  [main] util.ServerCommandLine: env:HBASE_CLASSPATH=/usr/hdp/2.4.2.0-258/hadoop/conf:/usr/hdp/2.4.2.0-258/hadoop/*:/usr/hdp/2.4.2.0-258/hadoop/lib/*:/usr/hdp/2.4.2.0-258/zookeeper/*:/usr/hdp/2.4.2.0-258/zookeeper/lib/*::/opt/splice/default/lib/concurrentlinkedhashmap-lru-1.4.2.jar:/opt/splice/default/lib/db-client-2.0.1.28.jar:/opt/splice/default/lib/db-drda-2.0.1.28.jar:/opt/splice/default/lib/db-engine-2.0.1.28.jar:/opt/splice/default/lib/db-shared-2.0.1.28.jar:/opt/splice/default/lib/db-tools-i18n-2.0.1.28.jar:/opt/splice/default/lib/db-tools-ij-2.0.1.28.jar:/opt/splice/default/lib/disruptor-3.2.1.jar:/opt/splice/default/lib/gson-2.2.2.jar:/opt/splice/default/lib/hbase_pipeline-hdp2.4.2-2.0.1.28.jar:/opt/splice/default/lib/hbase_sql-hdp2.4.2-2.0.1.28.jar:/opt/splice/default/lib/hbase_storage-hdp2.4.2-2.0.1.28.jar:/opt/splice/default/lib/hppc-0.5.2.jar:/opt/splice/default/lib/kryo-2.21.jar:/opt/splice/default/lib/kryo-serializers-0.26.jar:/opt/splice/default/lib/lucene-core-4.3.1.jar:/opt/splice/default/lib/opencsv-2.3.jar:/opt/splice/default/lib/pipeline_api-2.0.1.28.jar:/opt/splice/default/lib/protobuf-java-2.5.0.jar:/opt/splice/default/lib/spark-assembly-hadoop2.7.1.2.4.2.0-258-1.6.2.jar:/opt/splice/default/lib/splice_access_api-2.0.1.28.jar:/opt/splice/default/lib/splice_auth-2.0.1.28.jar:/opt/splice/default/lib/splice_backup-hdp2.4.2-2.0.1.28.jar:/opt/splice/default/lib/splice_colperms-2.0.1.28.jar:/opt/splice/default/lib/splice_ee-hdp2.4.2-2.0.1.28.jar:/opt/splice/default/lib/splice_encoding-2.0.1.28.jar:/opt/splice/default/lib/splice_encryption-2.0.1.28.jar:/opt/splice/default/lib/splice_machine-2.0.1.28.jar:/opt/splice/default/lib/splice_protocol-2.0.1.28.jar:/opt/splice/default/lib/splice_si_api-2.0.1.28.jar:/opt/splice/default/lib/splice_timestamp_api-2.0.1.28.jar:/opt/splice/default/lib/stats-2.0.1.28.jar:/opt/splice/default/lib/super-csv-2.4.0.jar:/opt/splice/default/lib/utilities-2.0.1.28.jar
2016-12-02 19:20:16,974 INFO  [main] util.ServerCommandLine: env:CVS_RSH=ssh
2016-12-02 19:20:16,974 INFO  [main] util.ServerCommandLine: env:HBASE_IDENT_STRING=hbase
2016-12-02 19:20:16,974 INFO  [main] util.ServerCommandLine: env:HBASE_ZNODE_FILE=/var/run/hbase/hbase-hbase-regionserver.znode
2016-12-02 19:20:16,974 INFO  [main] util.ServerCommandLine: env:HBASE_LOG_PREFIX=hbase-hbase-regionserver-hscale-dev1-dn1
2016-12-02 19:20:16,974 INFO  [main] util.ServerCommandLine: env:HBASE_LOG_DIR=/var/log/hbase
2016-12-02 19:20:16,974 INFO  [main] util.ServerCommandLine: env:USER=hbase
2016-12-02 19:20:16,974 INFO  [main] util.ServerCommandLine: env:CLASSPATH=/usr/hdp/current/hbase-regionserver/conf:/usr/jdk64/jdk1.8.0_60/lib/tools.jar:/usr/hdp/current/hbase-regionserver/bin/..:/usr/hdp/current/hbase-regionserver/bin/../lib/activation-1.1.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/aopalliance-1.0.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/apacheds-i18n-2.0.0-M15.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/apacheds-kerberos-codec-2.0.0-M15.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/api-asn1-api-1.0.0-M20.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/api-util-1.0.0-M20.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/asm-3.1.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/avro-1.7.4.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/commons-beanutils-1.7.0.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/commons-beanutils-core-1.8.0.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/commons-cli-1.2.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/commons-codec-1.9.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/commons-collections-3.2.2.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/commons-compress-1.4.1.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/commons-configuration-1.6.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/commons-daemon-1.0.13.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/commons-digester-1.8.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/commons-el-1.0.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/commons-httpclient-3.1.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/commons-io-2.4.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/commons-lang-2.6.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/commons-logging-1.2.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/commons-math-2.2.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/commons-math3-3.1.1.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/commons-net-3.1.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/curator-client-2.7.1.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/curator-framework-2.7.1.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/curator-recipes-2.7.1.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/disruptor-3.3.0.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/findbugs-annotations-1.3.9-1.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/gson-2.2.4.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/guava-12.0.1.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/guice-3.0.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/guice-servlet-3.0.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/hbase-annotations-1.1.2.2.4.2.0-258.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/hbase-annotations-1.1.2.2.4.2.0-258-tests.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/hbase-annotations.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/hbase-client-1.1.2.2.4.2.0-258.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/hbase-client.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/hbase-common-1.1.2.2.4.2.0-258.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/hbase-common-1.1.2.2.4.2.0-258-tests.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/hbase-common.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/hbase-examples-1.1.2.2.4.2.0-258.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/hbase-examples.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/hbase-hadoop2-compat-1.1.2.2.4.2.0-258.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/hbase-hadoop2-compat.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/hbase-hadoop-compat-1.1.2.2.4.2.0-258.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/hbase-hadoop-compat.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/hbase-it-1.1.2.2.4.2.0-258.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/hbase-it-1.1.2.2.4.2.0-258-tests.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/hbase-it.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/hbase-prefix-tree-1.1.2.2.4.2.0-258.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/hbase-prefix-tree.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/hbase-procedure-1.1.2.2.4.2.0-258.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/hbase-procedure.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/hbase-protocol-1.1.2.2.4.2.0-258.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/hbase-protocol.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/hbase-resource-bundle-1.1.2.2.4.2.0-258.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/hbase-resource-bundle.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/hbase-rest-1.1.2.2.4.2.0-258.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/hbase-rest.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/hbase-server-1.1.2.2.4.2.0-258.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/hbase-server-1.1.2.2.4.2.0-258-tests.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/hbase-server.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/hbase-shell-1.1.2.2.4.2.0-258.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/hbase-shell.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/hbase-thrift-1.1.2.2.4.2.0-258.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/hbase-thrift.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/htrace-core-3.1.0-incubating.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/httpclient-4.2.5.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/httpcore-4.2.5.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/jackson-core-2.2.3.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/jackson-core-asl-1.9.13.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/jackson-jaxrs-1.9.13.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/jackson-mapper-asl-1.9.13.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/jackson-xc-1.9.13.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/jamon-runtime-2.3.1.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/jasper-compiler-5.5.23.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/jasper-runtime-5.5.23.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/javax.inject-1.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/java-xmlbuilder-0.4.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/jaxb-api-2.2.2.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/jaxb-impl-2.2.3-1.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/jcodings-1.0.8.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/jersey-client-1.9.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/jersey-core-1.9.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/jersey-guice-1.9.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/jersey-json-1.9.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/jersey-server-1.9.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/jets3t-0.9.0.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/jettison-1.3.3.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/jetty-6.1.26.hwx.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/jetty-sslengine-6.1.26.hwx.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/jetty-util-6.1.26.hwx.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/joni-2.1.2.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/jruby-complete-1.6.8.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/jsch-0.1.42.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/jsp-2.1-6.1.14.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/jsp-api-2.1-6.1.14.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/jsr305-1.3.9.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/junit-4.11.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/leveldbjni-all-1.8.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/libthrift-0.9.0.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/log4j-1.2.17.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/metrics-core-2.2.0.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/microsoft-windowsazure-storage-sdk-0.6.0.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/netty-3.2.4.Final.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/netty-all-4.0.23.Final.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/ojdbc6.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/okhttp-2.4.0.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/okio-1.4.0.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/paranamer-2.3.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/phoenix-4.8.0-HBase-1.1-server.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/phoenix-server.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/protobuf-java-2.5.0.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/ranger-hbase-plugin-shim-0.5.0.2.4.2.0-258.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/ranger-plugin-classloader-0.5.0.2.4.2.0-258.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/servlet-api-2.5-6.1.14.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/servlet-api-2.5.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/slf4j-api-1.7.7.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/snappy-java-1.0.4.1.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/spymemcached-2.11.6.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/xercesImpl-2.9.1.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/xml-apis-1.3.04.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/xmlenc-0.52.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/xz-1.0.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/zookeeper.jar:/usr/hdp/2.4.2.0-258/hadoop/conf:/usr/hdp/2.4.2.0-258/hadoop/lib/*:/usr/hdp/2.4.2.0-258/hadoop/.//*:/usr/hdp/2.4.2.0-258/hadoop-hdfs/./:/usr/hdp/2.4.2.0-258/hadoop-hdfs/lib/*:/usr/hdp/2.4.2.0-258/hadoop-hdfs/.//*:/usr/hdp/2.4.2.0-258/hadoop-yarn/lib/*:/usr/hdp/2.4.2.0-258/hadoop-yarn/.//*:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/lib/*:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//*::mysql-connector-java-5.1.17.jar:mysql-connector-java-5.1.17.jar_bkp:mysql-connector-java.jar:/usr/hdp/2.4.2.0-258/tez/*:/usr/hdp/2.4.2.0-258/tez/lib/*:/usr/hdp/2.4.2.0-258/tez/conf:/usr/hdp/2.4.2.0-258/hadoop/conf:/usr/hdp/2.4.2.0-258/hadoop/*:/usr/hdp/2.4.2.0-258/hadoop/lib/*:/usr/hdp/2.4.2.0-258/zookeeper/*:/usr/hdp/2.4.2.0-258/zookeeper/lib/*::/opt/splice/default/lib/concurrentlinkedhashmap-lru-1.4.2.jar:/opt/splice/default/lib/db-client-2.0.1.28.jar:/opt/splice/default/lib/db-drda-2.0.1.28.jar:/opt/splice/default/lib/db-engine-2.0.1.28.jar:/opt/splice/default/lib/db-shared-2.0.1.28.jar:/opt/splice/default/lib/db-tools-i18n-2.0.1.28.jar:/opt/splice/default/lib/db-tools-ij-2.0.1.28.jar:/opt/splice/default/lib/disruptor-3.2.1.jar:/opt/splice/default/lib/gson-2.2.2.jar:/opt/splice/default/lib/hbase_pipeline-hdp2.4.2-2.0.1.28.jar:/opt/splice/default/lib/hbase_sql-hdp2.4.2-2.0.1.28.jar:/opt/splice/default/lib/hbase_storage-hdp2.4.2-2.0.1.28.jar:/opt/splice/default/lib/hppc-0.5.2.jar:/opt/splice/default/lib/kryo-2.21.jar:/opt/splice/default/lib/kryo-serializers-0.26.jar:/opt/splice/default/lib/lucene-core-4.3.1.jar:/opt/splice/default/lib/opencsv-2.3.jar:/opt/splice/default/lib/pipeline_api-2.0.1.28.jar:/opt/splice/default/lib/protobuf-java-2.5.0.jar:/opt/splice/default/lib/spark-assembly-hadoop2.7.1.2.4.2.0-258-1.6.2.jar:/opt/splice/default/lib/splice_access_api-2.0.1.28.jar:/opt/splice/default/lib/splice_auth-2.0.1.28.jar:/opt/splice/default/lib/splice_backup-hdp2.4.2-2.0.1.28.jar:/opt/splice/default/lib/splice_colperms-2.0.1.28.jar:/opt/splice/default/lib/splice_ee-hdp2.4.2-2.0.1.28.jar:/opt/splice/default/lib/splice_encoding-2.0.1.28.jar:/opt/splice/default/lib/splice_encryption-2.0.1.28.jar:/opt/splice/default/lib/splice_machine-2.0.1.28.jar:/opt/splice/default/lib/splice_protocol-2.0.1.28.jar:/opt/splice/default/lib/splice_si_api-2.0.1.28.jar:/opt/splice/default/lib/splice_timestamp_api-2.0.1.28.jar:/opt/splice/default/lib/stats-2.0.1.28.jar:/opt/splice/default/lib/super-csv-2.4.0.jar:/opt/splice/default/lib/utilities-2.0.1.28.jar
2016-12-02 19:20:16,974 INFO  [main] util.ServerCommandLine: env:SERVER_GC_OPTS=-verbose:gc -XX:+PrintGCDetails -XX:+PrintGCDateStamps -Xloggc:/var/log/hbase/gc.log-201612021920
2016-12-02 19:20:16,974 INFO  [main] util.ServerCommandLine: env:HADOOP_CONF=/usr/hdp/2.4.2.0-258/hadoop/conf
2016-12-02 19:20:16,974 INFO  [main] util.ServerCommandLine: env:HOSTNAME=hscale-dev1-dn1
2016-12-02 19:20:16,974 INFO  [main] util.ServerCommandLine: env:QTDIR=/usr/lib64/qt-3.3
2016-12-02 19:20:16,974 INFO  [main] util.ServerCommandLine: env:NLSPATH=/usr/dt/lib/nls/msg/%L/%N.cat
2016-12-02 19:20:16,975 INFO  [main] util.ServerCommandLine: env:HBASE_THRIFT_OPTS=
2016-12-02 19:20:16,975 INFO  [main] util.ServerCommandLine: env:HBASE_HOME=/usr/hdp/current/hbase-regionserver/bin/..
2016-12-02 19:20:16,975 INFO  [main] util.ServerCommandLine: env:QTLIB=/usr/lib64/qt-3.3/lib
2016-12-02 19:20:16,975 INFO  [main] util.ServerCommandLine: env:HOME=/home/hbase
2016-12-02 19:20:16,975 INFO  [main] util.ServerCommandLine: env:MALLOC_ARENA_MAX=4
2016-12-02 19:20:16,975 INFO  [main] util.ServerCommandLine: vmName=Java HotSpot(TM) 64-Bit Server VM, vmVendor=Oracle Corporation, vmVersion=25.60-b23
2016-12-02 19:20:16,975 INFO  [main] util.ServerCommandLine: vmInputArguments=[-Dproc_regionserver, -XX:OnOutOfMemoryError=kill -9 %p, -Dhdp.version=2.4.2.0-258, -XX:ErrorFile=/var/log/hbase/hs_err_pid%p.log, -Djava.io.tmpdir=/tmp, -verbose:gc, -XX:+PrintGCDetails, -XX:+PrintGCDateStamps, -Xloggc:/var/log/hbase/gc.log-201612021920, -Xmn512m, -Xms3072m, -Xmx3072m, -XX:+HeapDumpOnOutOfMemoryError, -XX:MaxDirectMemorySize=2g, -XX:+AlwaysPreTouch, -XX:+UseG1GC, -XX:MaxNewSize=4g, -XX:InitiatingHeapOccupancyPercent=60, -XX:ParallelGCThreads=24, -XX:+ParallelRefProcEnabled, -XX:MaxGCPauseMillis=5000, -Dcom.sun.management.jmxremote.authenticate=false, -Dcom.sun.management.jmxremote.ssl=false, -Dcom.sun.management.jmxremote.port=10102, -Dhbase.log.dir=/var/log/hbase, -Dhbase.log.file=hbase-hbase-regionserver-hscale-dev1-dn1.log, -Dhbase.home.dir=/usr/hdp/current/hbase-regionserver/bin/.., -Dhbase.id.str=hbase, -Dhbase.root.logger=INFO,RFA, -Djava.library.path=:/usr/hdp/2.4.2.0-258/hadoop/lib/native/Linux-amd64-64:/usr/hdp/2.4.2.0-258/hadoop/lib/native, -Dhbase.security.logger=INFO,RFAS]
2016-12-02 19:20:17,172 INFO  [main] regionserver.RSRpcServices: regionserver/hscale-dev1-dn1/10.60.70.11:16020 server-side HConnection retries=50
2016-12-02 19:20:17,249 INFO  [main] ipc.SimpleRpcScheduler: Using deadline as user call queue, count=10
2016-12-02 19:20:17,262 INFO  [main] ipc.RpcServer: regionserver/hscale-dev1-dn1/10.60.70.11:16020: started 10 reader(s).
2016-12-02 19:20:17,295 INFO  [main] impl.MetricsConfig: loaded properties from hadoop-metrics2-hbase.properties
2016-12-02 19:20:17,311 INFO  [main] timeline.HadoopTimelineMetricsSink: Initializing Timeline metrics sink.
2016-12-02 19:20:17,311 INFO  [main] timeline.HadoopTimelineMetricsSink: Identified hostname = hscale-dev1-dn1, serviceName = hbase
2016-12-02 19:20:17,314 INFO  [main] timeline.HadoopTimelineMetricsSink: Collector Uri: http://hscale-dev1-nn:6188/ws/v1/timeline/metrics
2016-12-02 19:20:17,319 INFO  [main] impl.MetricsSinkAdapter: Sink timeline started
2016-12-02 19:20:17,328 INFO  [main] impl.MetricsSystemImpl: Scheduled snapshot period at 60 second(s).
2016-12-02 19:20:17,328 INFO  [main] impl.MetricsSystemImpl: HBase metrics system started
2016-12-02 19:20:18,021 INFO  [main] fs.HFileSystem: Added intercepting call to namenode#getBlockLocations so can do block reordering using class class org.apache.hadoop.hbase.fs.HFileSystem$ReorderWALBlocks
2016-12-02 19:20:18,168 INFO  [main] zookeeper.RecoverableZooKeeper: Process identifier=regionserver:16020 connecting to ZooKeeper ensemble=hscale-dev1-dn1:2181,hscale-dev1-dn3:2181,hscale-dev1-dn2:2181,hscale-dev1-dn4:2181
2016-12-02 19:20:18,172 INFO  [main] zookeeper.ZooKeeper: Client environment:zookeeper.version=3.4.6-258--1, built on 04/25/2016 05:22 GMT
2016-12-02 19:20:18,172 INFO  [main] zookeeper.ZooKeeper: Client environment:host.name=hscale-dev1-dn1
2016-12-02 19:20:18,172 INFO  [main] zookeeper.ZooKeeper: Client environment:java.version=1.8.0_60
2016-12-02 19:20:18,172 INFO  [main] zookeeper.ZooKeeper: Client environment:java.vendor=Oracle Corporation
2016-12-02 19:20:18,172 INFO  [main] zookeeper.ZooKeeper: Client environment:java.home=/usr/jdk64/jdk1.8.0_60/jre
2016-12-02 19:20:18,172 INFO  [main] zookeeper.ZooKeeper: Client environment:java.class.path=/usr/hdp/current/hbase-regionserver/conf:/usr/jdk64/jdk1.8.0_60/lib/tools.jar:/usr/hdp/current/hbase-regionserver/bin/..:/usr/hdp/current/hbase-regionserver/bin/../lib/activation-1.1.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/aopalliance-1.0.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/apacheds-i18n-2.0.0-M15.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/apacheds-kerberos-codec-2.0.0-M15.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/api-asn1-api-1.0.0-M20.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/api-util-1.0.0-M20.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/asm-3.1.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/avro-1.7.4.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/commons-beanutils-1.7.0.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/commons-beanutils-core-1.8.0.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/commons-cli-1.2.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/commons-codec-1.9.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/commons-collections-3.2.2.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/commons-compress-1.4.1.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/commons-configuration-1.6.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/commons-daemon-1.0.13.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/commons-digester-1.8.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/commons-el-1.0.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/commons-httpclient-3.1.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/commons-io-2.4.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/commons-lang-2.6.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/commons-logging-1.2.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/commons-math-2.2.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/commons-math3-3.1.1.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/commons-net-3.1.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/curator-client-2.7.1.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/curator-framework-2.7.1.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/curator-recipes-2.7.1.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/disruptor-3.3.0.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/findbugs-annotations-1.3.9-1.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/gson-2.2.4.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/guava-12.0.1.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/guice-3.0.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/guice-servlet-3.0.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/hbase-annotations-1.1.2.2.4.2.0-258.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/hbase-annotations-1.1.2.2.4.2.0-258-tests.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/hbase-annotations.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/hbase-client-1.1.2.2.4.2.0-258.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/hbase-client.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/hbase-common-1.1.2.2.4.2.0-258.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/hbase-common-1.1.2.2.4.2.0-258-tests.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/hbase-common.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/hbase-examples-1.1.2.2.4.2.0-258.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/hbase-examples.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/hbase-hadoop2-compat-1.1.2.2.4.2.0-258.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/hbase-hadoop2-compat.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/hbase-hadoop-compat-1.1.2.2.4.2.0-258.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/hbase-hadoop-compat.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/hbase-it-1.1.2.2.4.2.0-258.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/hbase-it-1.1.2.2.4.2.0-258-tests.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/hbase-it.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/hbase-prefix-tree-1.1.2.2.4.2.0-258.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/hbase-prefix-tree.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/hbase-procedure-1.1.2.2.4.2.0-258.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/hbase-procedure.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/hbase-protocol-1.1.2.2.4.2.0-258.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/hbase-protocol.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/hbase-resource-bundle-1.1.2.2.4.2.0-258.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/hbase-resource-bundle.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/hbase-rest-1.1.2.2.4.2.0-258.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/hbase-rest.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/hbase-server-1.1.2.2.4.2.0-258.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/hbase-server-1.1.2.2.4.2.0-258-tests.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/hbase-server.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/hbase-shell-1.1.2.2.4.2.0-258.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/hbase-shell.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/hbase-thrift-1.1.2.2.4.2.0-258.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/hbase-thrift.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/htrace-core-3.1.0-incubating.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/httpclient-4.2.5.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/httpcore-4.2.5.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/jackson-core-2.2.3.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/jackson-core-asl-1.9.13.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/jackson-jaxrs-1.9.13.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/jackson-mapper-asl-1.9.13.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/jackson-xc-1.9.13.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/jamon-runtime-2.3.1.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/jasper-compiler-5.5.23.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/jasper-runtime-5.5.23.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/javax.inject-1.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/java-xmlbuilder-0.4.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/jaxb-api-2.2.2.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/jaxb-impl-2.2.3-1.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/jcodings-1.0.8.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/jersey-client-1.9.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/jersey-core-1.9.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/jersey-guice-1.9.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/jersey-json-1.9.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/jersey-server-1.9.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/jets3t-0.9.0.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/jettison-1.3.3.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/jetty-6.1.26.hwx.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/jetty-sslengine-6.1.26.hwx.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/jetty-util-6.1.26.hwx.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/joni-2.1.2.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/jruby-complete-1.6.8.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/jsch-0.1.42.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/jsp-2.1-6.1.14.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/jsp-api-2.1-6.1.14.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/jsr305-1.3.9.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/junit-4.11.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/leveldbjni-all-1.8.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/libthrift-0.9.0.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/log4j-1.2.17.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/metrics-core-2.2.0.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/microsoft-windowsazure-storage-sdk-0.6.0.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/netty-3.2.4.Final.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/netty-all-4.0.23.Final.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/ojdbc6.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/okhttp-2.4.0.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/okio-1.4.0.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/paranamer-2.3.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/phoenix-4.8.0-HBase-1.1-server.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/phoenix-server.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/protobuf-java-2.5.0.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/ranger-hbase-plugin-shim-0.5.0.2.4.2.0-258.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/ranger-plugin-classloader-0.5.0.2.4.2.0-258.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/servlet-api-2.5-6.1.14.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/servlet-api-2.5.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/slf4j-api-1.7.7.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/snappy-java-1.0.4.1.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/spymemcached-2.11.6.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/xercesImpl-2.9.1.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/xml-apis-1.3.04.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/xmlenc-0.52.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/xz-1.0.jar:/usr/hdp/current/hbase-regionserver/bin/../lib/zookeeper.jar:/usr/hdp/2.4.2.0-258/hadoop/conf:/usr/hdp/2.4.2.0-258/hadoop/lib/jackson-core-asl-1.9.13.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/jackson-databind-2.2.3.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/httpclient-4.2.5.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/commons-collections-3.2.2.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/jaxb-api-2.2.2.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/log4j-1.2.17.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/jettison-1.1.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/jackson-jaxrs-1.9.13.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/hamcrest-core-1.3.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/servlet-api-2.5.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/gson-2.2.4.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/spark-yarn-shuffle.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/aws-java-sdk-1.7.4.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/htrace-core-3.1.0-incubating.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/ranger-hdfs-plugin-shim-0.5.0.2.4.2.0-258.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/paranamer-2.3.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/jersey-core-1.9.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/commons-codec-1.4.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/avro-1.7.4.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/commons-math3-3.1.1.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/commons-logging-1.1.3.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/microsoft-windowsazure-storage-sdk-0.6.0.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/apacheds-i18n-2.0.0-M15.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/jetty-6.1.26.hwx.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/httpcore-4.2.5.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/api-util-1.0.0-M20.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/jackson-annotations-2.2.3.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/commons-compress-1.4.1.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/commons-digester-1.8.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/commons-beanutils-1.7.0.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/jackson-mapper-asl-1.9.13.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/slf4j-log4j12-1.7.10.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/curator-framework-2.7.1.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/jets3t-0.9.0.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/commons-httpclient-3.1.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/apacheds-kerberos-codec-2.0.0-M15.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/snappy-java-1.0.4.1.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/junit-4.11.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/azure-storage-2.2.0.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/ojdbc6.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/ranger-plugin-classloader-0.5.0.2.4.2.0-258.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/activation-1.1.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/jackson-xc-1.9.13.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/jersey-server-1.9.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/mockito-all-1.8.5.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/guava-11.0.2.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/netty-3.6.2.Final.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/ranger-yarn-plugin-shim-0.5.0.2.4.2.0-258.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/protobuf-java-2.5.0.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/xmlenc-0.52.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/api-asn1-api-1.0.0-M20.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/commons-configuration-1.6.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/jsp-api-2.1.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/xz-1.0.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/zookeeper-3.4.6.2.4.2.0-258.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/commons-io-2.4.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/curator-client-2.7.1.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/commons-cli-1.2.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/jersey-json-1.9.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/jsch-0.1.42.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/commons-net-3.1.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/java-xmlbuilder-0.4.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/curator-recipes-2.7.1.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/asm-3.2.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/slf4j-api-1.7.10.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/jaxb-impl-2.2.3-1.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/jsr305-3.0.0.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/stax-api-1.0-2.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/commons-lang-2.6.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/commons-beanutils-core-1.8.0.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/jetty-util-6.1.26.hwx.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/jackson-core-2.2.3.jar:/usr/hdp/2.4.2.0-258/hadoop/.//hadoop-nfs-2.7.1.2.4.2.0-258.jar:/usr/hdp/2.4.2.0-258/hadoop/.//hadoop-common-2.7.1.2.4.2.0-258-tests.jar:/usr/hdp/2.4.2.0-258/hadoop/.//hadoop-annotations-2.7.1.2.4.2.0-258.jar:/usr/hdp/2.4.2.0-258/hadoop/.//hadoop-auth-2.7.1.2.4.2.0-258.jar:/usr/hdp/2.4.2.0-258/hadoop/.//hadoop-azure-2.7.1.2.4.2.0-258.jar:/usr/hdp/2.4.2.0-258/hadoop/.//hadoop-aws-2.7.1.2.4.2.0-258.jar:/usr/hdp/2.4.2.0-258/hadoop/.//hadoop-common-tests.jar:/usr/hdp/2.4.2.0-258/hadoop/.//hadoop-common.jar:/usr/hdp/2.4.2.0-258/hadoop/.//hadoop-auth.jar:/usr/hdp/2.4.2.0-258/hadoop/.//hadoop-aws.jar:/usr/hdp/2.4.2.0-258/hadoop/.//hadoop-common-2.7.1.2.4.2.0-258.jar:/usr/hdp/2.4.2.0-258/hadoop/.//hadoop-annotations.jar:/usr/hdp/2.4.2.0-258/hadoop/.//hadoop-azure.jar:/usr/hdp/2.4.2.0-258/hadoop/.//hadoop-nfs.jar:/usr/hdp/2.4.2.0-258/hadoop-hdfs/./:/usr/hdp/2.4.2.0-258/hadoop-hdfs/lib/jackson-core-asl-1.9.13.jar:/usr/hdp/2.4.2.0-258/hadoop-hdfs/lib/okio-1.4.0.jar:/usr/hdp/2.4.2.0-258/hadoop-hdfs/lib/log4j-1.2.17.jar:/usr/hdp/2.4.2.0-258/hadoop-hdfs/lib/servlet-api-2.5.jar:/usr/hdp/2.4.2.0-258/hadoop-hdfs/lib/htrace-core-3.1.0-incubating.jar:/usr/hdp/2.4.2.0-258/hadoop-hdfs/lib/jersey-core-1.9.jar:/usr/hdp/2.4.2.0-258/hadoop-hdfs/lib/commons-codec-1.4.jar:/usr/hdp/2.4.2.0-258/hadoop-hdfs/lib/commons-logging-1.1.3.jar:/usr/hdp/2.4.2.0-258/hadoop-hdfs/lib/jetty-6.1.26.hwx.jar:/usr/hdp/2.4.2.0-258/hadoop-hdfs/lib/jackson-mapper-asl-1.9.13.jar:/usr/hdp/2.4.2.0-258/hadoop-hdfs/lib/xercesImpl-2.9.1.jar:/usr/hdp/2.4.2.0-258/hadoop-hdfs/lib/leveldbjni-all-1.8.jar:/usr/hdp/2.4.2.0-258/hadoop-hdfs/lib/jersey-server-1.9.jar:/usr/hdp/2.4.2.0-258/hadoop-hdfs/lib/guava-11.0.2.jar:/usr/hdp/2.4.2.0-258/hadoop-hdfs/lib/netty-3.6.2.Final.jar:/usr/hdp/2.4.2.0-258/hadoop-hdfs/lib/protobuf-java-2.5.0.jar:/usr/hdp/2.4.2.0-258/hadoop-hdfs/lib/xmlenc-0.52.jar:/usr/hdp/2.4.2.0-258/hadoop-hdfs/lib/xml-apis-1.3.04.jar:/usr/hdp/2.4.2.0-258/hadoop-hdfs/lib/netty-all-4.0.23.Final.jar:/usr/hdp/2.4.2.0-258/hadoop-hdfs/lib/commons-io-2.4.jar:/usr/hdp/2.4.2.0-258/hadoop-hdfs/lib/commons-cli-1.2.jar:/usr/hdp/2.4.2.0-258/hadoop-hdfs/lib/asm-3.2.jar:/usr/hdp/2.4.2.0-258/hadoop-hdfs/lib/jsr305-3.0.0.jar:/usr/hdp/2.4.2.0-258/hadoop-hdfs/lib/commons-lang-2.6.jar:/usr/hdp/2.4.2.0-258/hadoop-hdfs/lib/jetty-util-6.1.26.hwx.jar:/usr/hdp/2.4.2.0-258/hadoop-hdfs/lib/okhttp-2.4.0.jar:/usr/hdp/2.4.2.0-258/hadoop-hdfs/lib/commons-daemon-1.0.13.jar:/usr/hdp/2.4.2.0-258/hadoop-hdfs/.//hadoop-hdfs-2.7.1.2.4.2.0-258-tests.jar:/usr/hdp/2.4.2.0-258/hadoop-hdfs/.//hadoop-hdfs-tests.jar:/usr/hdp/2.4.2.0-258/hadoop-hdfs/.//hadoop-hdfs-nfs-2.7.1.2.4.2.0-258.jar:/usr/hdp/2.4.2.0-258/hadoop-hdfs/.//hadoop-hdfs-nfs.jar:/usr/hdp/2.4.2.0-258/hadoop-hdfs/.//hadoop-hdfs-2.7.1.2.4.2.0-258.jar:/usr/hdp/2.4.2.0-258/hadoop-hdfs/.//hadoop-hdfs.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/lib/jackson-core-asl-1.9.13.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/lib/jackson-databind-2.2.3.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/lib/httpclient-4.2.5.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/lib/commons-collections-3.2.2.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/lib/jaxb-api-2.2.2.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/lib/log4j-1.2.17.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/lib/jettison-1.1.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/lib/jackson-jaxrs-1.9.13.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/lib/guice-3.0.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/lib/servlet-api-2.5.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/lib/gson-2.2.4.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/lib/htrace-core-3.1.0-incubating.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/lib/paranamer-2.3.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/lib/jersey-core-1.9.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/lib/commons-codec-1.4.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/lib/avro-1.7.4.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/lib/commons-math3-3.1.1.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/lib/commons-logging-1.1.3.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/lib/jersey-guice-1.9.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/lib/microsoft-windowsazure-storage-sdk-0.6.0.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/lib/apacheds-i18n-2.0.0-M15.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/lib/jetty-6.1.26.hwx.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/lib/httpcore-4.2.5.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/lib/spark-assembly-hadoop2.7.1.2.4.2.0-258-1.6.2.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/lib/api-util-1.0.0-M20.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/lib/jackson-annotations-2.2.3.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/lib/commons-compress-1.4.1.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/lib/guice-servlet-3.0.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/lib/commons-digester-1.8.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/lib/javassist-3.18.1-GA.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/lib/commons-beanutils-1.7.0.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/lib/jackson-mapper-asl-1.9.13.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/lib/curator-framework-2.7.1.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/lib/jets3t-0.9.0.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/lib/commons-httpclient-3.1.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/lib/apacheds-kerberos-codec-2.0.0-M15.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/lib/snappy-java-1.0.4.1.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/lib/jersey-client-1.9.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/lib/activation-1.1.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/lib/jackson-xc-1.9.13.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/lib/leveldbjni-all-1.8.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/lib/jersey-server-1.9.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/lib/zookeeper-3.4.6.2.4.2.0-258-tests.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/lib/guava-11.0.2.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/lib/netty-3.6.2.Final.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/lib/protobuf-java-2.5.0.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/lib/xmlenc-0.52.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/lib/api-asn1-api-1.0.0-M20.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/lib/commons-configuration-1.6.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/lib/jsp-api-2.1.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/lib/fst-2.24.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/lib/xz-1.0.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/lib/aopalliance-1.0.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/lib/zookeeper-3.4.6.2.4.2.0-258.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/lib/commons-io-2.4.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/lib/curator-client-2.7.1.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/lib/commons-cli-1.2.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/lib/jersey-json-1.9.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/lib/jsch-0.1.42.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/lib/commons-net-3.1.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/lib/objenesis-2.1.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/lib/java-xmlbuilder-0.4.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/lib/curator-recipes-2.7.1.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/lib/asm-3.2.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/lib/jaxb-impl-2.2.3-1.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/lib/jsr305-3.0.0.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/lib/stax-api-1.0-2.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/lib/commons-lang-2.6.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/lib/javax.inject-1.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/lib/commons-beanutils-core-1.8.0.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/lib/jetty-util-6.1.26.hwx.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/lib/jackson-core-2.2.3.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/.//hadoop-yarn-server-common.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/.//hadoop-yarn-applications-distributedshell-2.7.1.2.4.2.0-258.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/.//hadoop-yarn-server-nodemanager-2.7.1.2.4.2.0-258.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/.//hadoop-yarn-server-common-2.7.1.2.4.2.0-258.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/.//hadoop-yarn-server-sharedcachemanager-2.7.1.2.4.2.0-258.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/.//hadoop-yarn-server-timeline-plugins-2.7.1.2.4.2.0-258.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/.//hadoop-yarn-registry-2.7.1.2.4.2.0-258.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/.//hadoop-yarn-server-applicationhistoryservice.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/.//hadoop-yarn-server-sharedcachemanager.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/.//hadoop-yarn-common-2.7.1.2.4.2.0-258.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/.//hadoop-yarn-server-web-proxy-2.7.1.2.4.2.0-258.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/.//hadoop-yarn-client.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/.//hadoop-yarn-common.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/.//hadoop-yarn-client-2.7.1.2.4.2.0-258.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/.//hadoop-yarn-registry.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/.//hadoop-yarn-api.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/.//hadoop-yarn-server-tests-2.7.1.2.4.2.0-258.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/.//hadoop-yarn-server-nodemanager.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/.//hadoop-yarn-server-resourcemanager.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/.//hadoop-yarn-server-resourcemanager-2.7.1.2.4.2.0-258.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/.//hadoop-yarn-applications-distributedshell.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/.//hadoop-yarn-applications-unmanaged-am-launcher-2.7.1.2.4.2.0-258.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/.//hadoop-yarn-applications-unmanaged-am-launcher.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/.//hadoop-yarn-server-applicationhistoryservice-2.7.1.2.4.2.0-258.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/.//hadoop-yarn-server-web-proxy.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/.//hadoop-yarn-api-2.7.1.2.4.2.0-258.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/.//hadoop-yarn-server-tests.jar:/usr/hdp/2.4.2.0-258/hadoop-yarn/.//hadoop-yarn-server-timeline-plugins.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/lib/jackson-core-asl-1.9.13.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/lib/log4j-1.2.17.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/lib/hamcrest-core-1.3.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/lib/guice-3.0.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/lib/paranamer-2.3.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/lib/jersey-core-1.9.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/lib/avro-1.7.4.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/lib/jersey-guice-1.9.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/lib/commons-compress-1.4.1.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/lib/guice-servlet-3.0.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/lib/jackson-mapper-asl-1.9.13.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/lib/snappy-java-1.0.4.1.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/lib/junit-4.11.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/lib/leveldbjni-all-1.8.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/lib/jersey-server-1.9.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/lib/netty-3.6.2.Final.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/lib/protobuf-java-2.5.0.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/lib/xz-1.0.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/lib/aopalliance-1.0.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/lib/commons-io-2.4.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/lib/asm-3.2.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/lib/javax.inject-1.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//jackson-core-asl-1.9.13.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//hadoop-openstack.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//httpclient-4.2.5.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//joda-time-2.9.3.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//commons-collections-3.2.2.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//hadoop-sls.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//hadoop-mapreduce-client-core.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//jaxb-api-2.2.2.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//hadoop-mapreduce-examples-2.7.1.2.4.2.0-258.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//hadoop-mapreduce-client-app.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//log4j-1.2.17.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//jettison-1.1.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//jackson-jaxrs-1.9.13.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//hamcrest-core-1.3.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//servlet-api-2.5.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//hadoop-ant-2.7.1.2.4.2.0-258.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//hadoop-mapreduce-client-jobclient-2.7.1.2.4.2.0-258.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//hadoop-auth-2.7.1.2.4.2.0-258.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//hadoop-sls-2.7.1.2.4.2.0-258.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//gson-2.2.4.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//htrace-core-3.1.0-incubating.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//hadoop-mapreduce-client-core-2.7.1.2.4.2.0-258.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//paranamer-2.3.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//jersey-core-1.9.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//hadoop-rumen.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//commons-codec-1.4.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//avro-1.7.4.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//commons-math3-3.1.1.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//hadoop-datajoin-2.7.1.2.4.2.0-258.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//commons-logging-1.1.3.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//hadoop-datajoin.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//hadoop-streaming.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//microsoft-windowsazure-storage-sdk-0.6.0.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//apacheds-i18n-2.0.0-M15.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//jetty-6.1.26.hwx.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//httpcore-4.2.5.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//hadoop-mapreduce-client-hs-plugins.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//api-util-1.0.0-M20.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//commons-compress-1.4.1.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//commons-digester-1.8.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//commons-beanutils-1.7.0.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//jackson-mapper-asl-1.9.13.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//curator-framework-2.7.1.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//jets3t-0.9.0.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//commons-httpclient-3.1.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//apacheds-kerberos-codec-2.0.0-M15.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//hadoop-mapreduce-client-hs-plugins-2.7.1.2.4.2.0-258.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//snappy-java-1.0.4.1.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//junit-4.11.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//hadoop-mapreduce-client-common.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//hadoop-auth.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//metrics-core-3.0.1.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//activation-1.1.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//jackson-xc-1.9.13.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//jersey-server-1.9.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//mockito-all-1.8.5.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//hadoop-mapreduce-client-hs.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//guava-11.0.2.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//netty-3.6.2.Final.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//protobuf-java-2.5.0.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//xmlenc-0.52.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//api-asn1-api-1.0.0-M20.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//hadoop-archives-2.7.1.2.4.2.0-258.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//commons-configuration-1.6.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//hadoop-mapreduce-client-hs-2.7.1.2.4.2.0-258.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//jsp-api-2.1.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//xz-1.0.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//zookeeper-3.4.6.2.4.2.0-258.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//commons-io-2.4.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//curator-client-2.7.1.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//commons-cli-1.2.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//jersey-json-1.9.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//jsch-0.1.42.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//hadoop-mapreduce-client-jobclient-tests.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//commons-net-3.1.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//hadoop-extras-2.7.1.2.4.2.0-258.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//hadoop-rumen-2.7.1.2.4.2.0-258.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//hadoop-mapreduce-client-app-2.7.1.2.4.2.0-258.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//hadoop-openstack-2.7.1.2.4.2.0-258.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//java-xmlbuilder-0.4.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//hadoop-archives.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//hadoop-gridmix.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//hadoop-ant.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//curator-recipes-2.7.1.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//hadoop-mapreduce-client-shuffle.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//asm-3.2.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//hadoop-gridmix-2.7.1.2.4.2.0-258.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//jaxb-impl-2.2.3-1.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//jsr305-3.0.0.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//hadoop-streaming-2.7.1.2.4.2.0-258.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//hadoop-distcp.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//stax-api-1.0-2.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//commons-lang-2.6.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//hadoop-mapreduce-client-jobclient.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//commons-beanutils-core-1.8.0.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//jetty-util-6.1.26.hwx.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//commons-lang3-3.3.2.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//hadoop-mapreduce-examples.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//jackson-core-2.2.3.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//hadoop-mapreduce-client-shuffle-2.7.1.2.4.2.0-258.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//hadoop-mapreduce-client-common-2.7.1.2.4.2.0-258.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//hadoop-mapreduce-client-jobclient-2.7.1.2.4.2.0-258-tests.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//hadoop-distcp-2.7.1.2.4.2.0-258.jar:/usr/hdp/2.4.2.0-258/hadoop-mapreduce/.//hadoop-extras.jar::mysql-connector-java-5.1.17.jar:mysql-connector-java-5.1.17.jar_bkp:mysql-connector-java.jar:/usr/hdp/2.4.2.0-258/tez/tez-yarn-timeline-history-with-acls-0.7.0.2.4.2.0-258.jar:/usr/hdp/2.4.2.0-258/tez/tez-runtime-library-0.7.0.2.4.2.0-258.jar:/usr/hdp/2.4.2.0-258/tez/tez-common-0.7.0.2.4.2.0-258.jar:/usr/hdp/2.4.2.0-258/tez/tez-dag-0.7.0.2.4.2.0-258.jar:/usr/hdp/2.4.2.0-258/tez/tez-api-0.7.0.2.4.2.0-258.jar:/usr/hdp/2.4.2.0-258/tez/tez-tests-0.7.0.2.4.2.0-258.jar:/usr/hdp/2.4.2.0-258/tez/tez-examples-0.7.0.2.4.2.0-258.jar:/usr/hdp/2.4.2.0-258/tez/tez-runtime-internals-0.7.0.2.4.2.0-258.jar:/usr/hdp/2.4.2.0-258/tez/tez-mapreduce-0.7.0.2.4.2.0-258.jar:/usr/hdp/2.4.2.0-258/tez/tez-yarn-timeline-history-0.7.0.2.4.2.0-258.jar:/usr/hdp/2.4.2.0-258/tez/tez-history-parser-0.7.0.2.4.2.0-258.jar:/usr/hdp/2.4.2.0-258/tez/tez-yarn-timeline-history-with-fs-0.7.0.2.4.2.0-258.jar:/usr/hdp/2.4.2.0-258/tez/tez-yarn-timeline-cache-plugin-0.7.0.2.4.2.0-258.jar:/usr/hdp/2.4.2.0-258/tez/lib/commons-collections-3.2.2.jar:/usr/hdp/2.4.2.0-258/tez/lib/servlet-api-2.5.jar:/usr/hdp/2.4.2.0-258/tez/lib/hadoop-annotations-2.7.1.2.4.2.0-258.jar:/usr/hdp/2.4.2.0-258/tez/lib/hadoop-yarn-server-timeline-plugins-2.7.1.2.4.2.0-258.jar:/usr/hdp/2.4.2.0-258/tez/lib/slf4j-api-1.7.5.jar:/usr/hdp/2.4.2.0-258/tez/lib/hadoop-mapreduce-client-core-2.7.1.2.4.2.0-258.jar:/usr/hdp/2.4.2.0-258/tez/lib/commons-codec-1.4.jar:/usr/hdp/2.4.2.0-258/tez/lib/commons-math3-3.1.1.jar:/usr/hdp/2.4.2.0-258/tez/lib/hadoop-azure-2.7.1.2.4.2.0-258.jar:/usr/hdp/2.4.2.0-258/tez/lib/hadoop-aws-2.7.1.2.4.2.0-258.jar:/usr/hdp/2.4.2.0-258/tez/lib/jetty-6.1.26.hwx.jar:/usr/hdp/2.4.2.0-258/tez/lib/hadoop-yarn-server-web-proxy-2.7.1.2.4.2.0-258.jar:/usr/hdp/2.4.2.0-258/tez/lib/jersey-client-1.9.jar:/usr/hdp/2.4.2.0-258/tez/lib/commons-collections4-4.1.jar:/usr/hdp/2.4.2.0-258/tez/lib/guava-11.0.2.jar:/usr/hdp/2.4.2.0-258/tez/lib/protobuf-java-2.5.0.jar:/usr/hdp/2.4.2.0-258/tez/lib/jettison-1.3.4.jar:/usr/hdp/2.4.2.0-258/tez/lib/commons-io-2.4.jar:/usr/hdp/2.4.2.0-258/tez/lib/commons-cli-1.2.jar:/usr/hdp/2.4.2.0-258/tez/lib/jersey-json-1.9.jar:/usr/hdp/2.4.2.0-258/tez/lib/jsr305-2.0.3.jar:/usr/hdp/2.4.2.0-258/tez/lib/commons-lang-2.6.jar:/usr/hdp/2.4.2.0-258/tez/lib/jetty-util-6.1.26.hwx.jar:/usr/hdp/2.4.2.0-258/tez/lib/hadoop-mapreduce-client-common-2.7.1.2.4.2.0-258.jar:/usr/hdp/2.4.2.0-258/tez/conf:/usr/hdp/2.4.2.0-258/hadoop/conf:/usr/hdp/2.4.2.0-258/hadoop/hadoop-nfs-2.7.1.2.4.2.0-258.jar:/usr/hdp/2.4.2.0-258/hadoop/hadoop-common-2.7.1.2.4.2.0-258-tests.jar:/usr/hdp/2.4.2.0-258/hadoop/hadoop-annotations-2.7.1.2.4.2.0-258.jar:/usr/hdp/2.4.2.0-258/hadoop/hadoop-auth-2.7.1.2.4.2.0-258.jar:/usr/hdp/2.4.2.0-258/hadoop/hadoop-azure-2.7.1.2.4.2.0-258.jar:/usr/hdp/2.4.2.0-258/hadoop/hadoop-aws-2.7.1.2.4.2.0-258.jar:/usr/hdp/2.4.2.0-258/hadoop/hadoop-common-tests.jar:/usr/hdp/2.4.2.0-258/hadoop/hadoop-common.jar:/usr/hdp/2.4.2.0-258/hadoop/hadoop-auth.jar:/usr/hdp/2.4.2.0-258/hadoop/hadoop-aws.jar:/usr/hdp/2.4.2.0-258/hadoop/hadoop-common-2.7.1.2.4.2.0-258.jar:/usr/hdp/2.4.2.0-258/hadoop/hadoop-annotations.jar:/usr/hdp/2.4.2.0-258/hadoop/hadoop-azure.jar:/usr/hdp/2.4.2.0-258/hadoop/hadoop-nfs.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/jackson-core-asl-1.9.13.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/jackson-databind-2.2.3.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/httpclient-4.2.5.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/commons-collections-3.2.2.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/jaxb-api-2.2.2.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/log4j-1.2.17.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/jettison-1.1.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/jackson-jaxrs-1.9.13.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/hamcrest-core-1.3.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/servlet-api-2.5.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/gson-2.2.4.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/spark-yarn-shuffle.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/aws-java-sdk-1.7.4.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/htrace-core-3.1.0-incubating.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/ranger-hdfs-plugin-shim-0.5.0.2.4.2.0-258.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/paranamer-2.3.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/jersey-core-1.9.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/commons-codec-1.4.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/avro-1.7.4.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/commons-math3-3.1.1.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/commons-logging-1.1.3.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/microsoft-windowsazure-storage-sdk-0.6.0.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/apacheds-i18n-2.0.0-M15.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/jetty-6.1.26.hwx.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/httpcore-4.2.5.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/api-util-1.0.0-M20.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/jackson-annotations-2.2.3.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/commons-compress-1.4.1.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/commons-digester-1.8.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/commons-beanutils-1.7.0.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/jackson-mapper-asl-1.9.13.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/slf4j-log4j12-1.7.10.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/curator-framework-2.7.1.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/jets3t-0.9.0.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/commons-httpclient-3.1.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/apacheds-kerberos-codec-2.0.0-M15.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/snappy-java-1.0.4.1.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/junit-4.11.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/azure-storage-2.2.0.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/ojdbc6.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/ranger-plugin-classloader-0.5.0.2.4.2.0-258.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/activation-1.1.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/jackson-xc-1.9.13.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/jersey-server-1.9.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/mockito-all-1.8.5.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/guava-11.0.2.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/netty-3.6.2.Final.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/ranger-yarn-plugin-shim-0.5.0.2.4.2.0-258.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/protobuf-java-2.5.0.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/xmlenc-0.52.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/api-asn1-api-1.0.0-M20.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/commons-configuration-1.6.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/jsp-api-2.1.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/xz-1.0.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/zookeeper-3.4.6.2.4.2.0-258.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/commons-io-2.4.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/curator-client-2.7.1.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/commons-cli-1.2.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/jersey-json-1.9.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/jsch-0.1.42.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/commons-net-3.1.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/java-xmlbuilder-0.4.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/curator-recipes-2.7.1.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/asm-3.2.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/slf4j-api-1.7.10.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/jaxb-impl-2.2.3-1.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/jsr305-3.0.0.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/stax-api-1.0-2.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/commons-lang-2.6.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/commons-beanutils-core-1.8.0.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/jetty-util-6.1.26.hwx.jar:/usr/hdp/2.4.2.0-258/hadoop/lib/jackson-core-2.2.3.jar:/usr/hdp/2.4.2.0-258/zookeeper/zookeeper.jar:/usr/hdp/2.4.2.0-258/zookeeper/zookeeper-3.4.6.2.4.2.0-258.jar:/usr/hdp/2.4.2.0-258/zookeeper/lib/netty-3.7.0.Final.jar:/usr/hdp/2.4.2.0-258/zookeeper/lib/plexus-interpolation-1.11.jar:/usr/hdp/2.4.2.0-258/zookeeper/lib/nekohtml-1.9.6.2.jar:/usr/hdp/2.4.2.0-258/zookeeper/lib/commons-codec-1.6.jar:/usr/hdp/2.4.2.0-258/zookeeper/lib/maven-plugin-registry-2.2.1.jar:/usr/hdp/2.4.2.0-258/zookeeper/lib/jline-0.9.94.jar:/usr/hdp/2.4.2.0-258/zookeeper/lib/wagon-http-2.4.jar:/usr/hdp/2.4.2.0-258/zookeeper/lib/wagon-http-shared-1.0-beta-6.jar:/usr/hdp/2.4.2.0-258/zookeeper/lib/jsoup-1.7.1.jar:/usr/hdp/2.4.2.0-258/zookeeper/lib/maven-artifact-manager-2.2.1.jar:/usr/hdp/2.4.2.0-258/zookeeper/lib/maven-profile-2.2.1.jar:/usr/hdp/2.4.2.0-258/zookeeper/lib/commons-io-2.2.jar:/usr/hdp/2.4.2.0-258/zookeeper/lib/ant-1.8.0.jar:/usr/hdp/2.4.2.0-258/zookeeper/lib/log4j-1.2.16.jar:/usr/hdp/2.4.2.0-258/zookeeper/lib/maven-model-2.2.1.jar:/usr/hdp/2.4.2.0-258/zookeeper/lib/maven-ant-tasks-2.1.3.jar:/usr/hdp/2.4.2.0-258/zookeeper/lib/maven-error-diagnostics-2.2.1.jar:/usr/hdp/2.4.2.0-258/zookeeper/lib/httpclient-4.2.3.jar:/usr/hdp/2.4.2.0-258/zookeeper/lib/slf4j-api-1.6.1.jar:/usr/hdp/2.4.2.0-258/zookeeper/lib/wagon-http-shared4-2.4.jar:/usr/hdp/2.4.2.0-258/zookeeper/lib/backport-util-concurrent-3.1.jar:/usr/hdp/2.4.2.0-258/zookeeper/lib/plexus-utils-3.0.8.jar:/usr/hdp/2.4.2.0-258/zookeeper/lib/maven-settings-2.2.1.jar:/usr/hdp/2.4.2.0-258/zookeeper/lib/slf4j-log4j12-1.6.1.jar:/usr/hdp/2.4.2.0-258/zookeeper/lib/xercesMinimal-1.9.6.2.jar:/usr/hdp/2.4.2.0-258/zookeeper/lib/plexus-container-default-1.0-alpha-9-stable-1.jar:/usr/hdp/2.4.2.0-258/zookeeper/lib/ant-launcher-1.8.0.jar:/usr/hdp/2.4.2.0-258/zookeeper/lib/wagon-file-1.0-beta-6.jar:/usr/hdp/2.4.2.0-258/zookeeper/lib/maven-artifact-2.2.1.jar:/usr/hdp/2.4.2.0-258/zookeeper/lib/commons-logging-1.1.1.jar:/usr/hdp/2.4.2.0-258/zookeeper/lib/classworlds-1.1-alpha-2.jar:/usr/hdp/2.4.2.0-258/zookeeper/lib/wagon-http-lightweight-1.0-beta-6.jar:/usr/hdp/2.4.2.0-258/zookeeper/lib/wagon-provider-api-2.4.jar:/usr/hdp/2.4.2.0-258/zookeeper/lib/maven-repository-metadata-2.2.1.jar:/usr/hdp/2.4.2.0-258/zookeeper/lib/maven-project-2.2.1.jar:/usr/hdp/2.4.2.0-258/zookeeper/lib/httpcore-4.2.3.jar::/opt/splice/default/lib/concurrentlinkedhashmap-lru-1.4.2.jar:/opt/splice/default/lib/db-client-2.0.1.28.jar:/opt/splice/default/lib/db-drda-2.0.1.28.jar:/opt/splice/default/lib/db-engine-2.0.1.28.jar:/opt/splice/default/lib/db-shared-2.0.1.28.jar:/opt/splice/default/lib/db-tools-i18n-2.0.1.28.jar:/opt/splice/default/lib/db-tools-ij-2.0.1.28.jar:/opt/splice/default/lib/disruptor-3.2.1.jar:/opt/splice/default/lib/gson-2.2.2.jar:/opt/splice/default/lib/hbase_pipeline-hdp2.4.2-2.0.1.28.jar:/opt/splice/default/lib/hbase_sql-hdp2.4.2-2.0.1.28.jar:/opt/splice/default/lib/hbase_storage-hdp2.4.2-2.0.1.28.jar:/opt/splice/default/lib/hppc-0.5.2.jar:/opt/splice/default/lib/kryo-2.21.jar:/opt/splice/default/lib/kryo-serializers-0.26.jar:/opt/splice/default/lib/lucene-core-4.3.1.jar:/opt/splice/default/lib/opencsv-2.3.jar:/opt/splice/default/lib/pipeline_api-2.0.1.28.jar:/opt/splice/default/lib/protobuf-java-2.5.0.jar:/opt/splice/default/lib/spark-assembly-hadoop2.7.1.2.4.2.0-258-1.6.2.jar:/opt/splice/default/lib/splice_access_api-2.0.1.28.jar:/opt/splice/default/lib/splice_auth-2.0.1.28.jar:/opt/splice/default/lib/splice_backup-hdp2.4.2-2.0.1.28.jar:/opt/splice/default/lib/splice_colperms-2.0.1.28.jar:/opt/splice/default/lib/splice_ee-hdp2.4.2-2.0.1.28.jar:/opt/splice/default/lib/splice_encoding-2.0.1.28.jar:/opt/splice/default/lib/splice_encryption-2.0.1.28.jar:/opt/splice/default/lib/splice_machine-2.0.1.28.jar:/opt/splice/default/lib/splice_protocol-2.0.1.28.jar:/opt/splice/default/lib/splice_si_api-2.0.1.28.jar:/opt/splice/default/lib/splice_timestamp_api-2.0.1.28.jar:/opt/splice/default/lib/stats-2.0.1.28.jar:/opt/splice/default/lib/super-csv-2.4.0.jar:/opt/splice/default/lib/utilities-2.0.1.28.jar
2016-12-02 19:20:18,173 INFO  [main] zookeeper.ZooKeeper: Client environment:java.library.path=:/usr/hdp/2.4.2.0-258/hadoop/lib/native/Linux-amd64-64:/usr/hdp/2.4.2.0-258/hadoop/lib/native
2016-12-02 19:20:18,173 INFO  [main] zookeeper.ZooKeeper: Client environment:java.io.tmpdir=/tmp
2016-12-02 19:20:18,173 INFO  [main] zookeeper.ZooKeeper: Client environment:java.compiler=<NA>
2016-12-02 19:20:18,173 INFO  [main] zookeeper.ZooKeeper: Client environment:os.name=Linux
2016-12-02 19:20:18,173 INFO  [main] zookeeper.ZooKeeper: Client environment:os.arch=amd64
2016-12-02 19:20:18,173 INFO  [main] zookeeper.ZooKeeper: Client environment:os.version=2.6.32-504.el6.x86_64
2016-12-02 19:20:18,173 INFO  [main] zookeeper.ZooKeeper: Client environment:user.name=hbase
2016-12-02 19:20:18,173 INFO  [main] zookeeper.ZooKeeper: Client environment:user.home=/home/hbase
2016-12-02 19:20:18,173 INFO  [main] zookeeper.ZooKeeper: Client environment:user.dir=/home/hbase
2016-12-02 19:20:18,174 INFO  [main] zookeeper.ZooKeeper: Initiating client connection, connectString=hscale-dev1-dn1:2181,hscale-dev1-dn3:2181,hscale-dev1-dn2:2181,hscale-dev1-dn4:2181 sessionTimeout=120000 watcher=regionserver:160200x0, quorum=hscale-dev1-dn1:2181,hscale-dev1-dn3:2181,hscale-dev1-dn2:2181,hscale-dev1-dn4:2181, baseZNode=/hbase-unsecure
2016-12-02 19:20:18,183 INFO  [main-SendThread(hscale-dev1-dn4:2181)] zookeeper.ClientCnxn: Opening socket connection to server hscale-dev1-dn4/10.60.70.14:2181. Will not attempt to authenticate using SASL (unknown error)
2016-12-02 19:20:18,187 INFO  [main-SendThread(hscale-dev1-dn4:2181)] zookeeper.ClientCnxn: Socket connection established to hscale-dev1-dn4/10.60.70.14:2181, initiating session
2016-12-02 19:20:18,203 INFO  [main-SendThread(hscale-dev1-dn4:2181)] zookeeper.ClientCnxn: Session establishment complete on server hscale-dev1-dn4/10.60.70.14:2181, sessionid = 0x458bf3f64d4002f, negotiated timeout = 120000
2016-12-02 19:20:18,212 INFO  [RpcServer.listener,port=16020] ipc.RpcServer: RpcServer.listener,port=16020: starting
2016-12-02 19:20:18,212 INFO  [RpcServer.responder] ipc.RpcServer: RpcServer.responder: starting
2016-12-02 19:20:18,274 INFO  [main] mortbay.log: Logging to org.slf4j.impl.Log4jLoggerAdapter(org.mortbay.log) via org.mortbay.log.Slf4jLog
2016-12-02 19:20:18,277 INFO  [main] http.HttpRequestLog: Http request log for http.requests.regionserver is not defined
2016-12-02 19:20:18,284 INFO  [main] http.HttpServer: Added global filter 'safety' (class=org.apache.hadoop.hbase.http.HttpServer$QuotingInputFilter)
2016-12-02 19:20:18,285 INFO  [main] http.HttpServer: Added filter static_user_filter (class=org.apache.hadoop.hbase.http.lib.StaticUserWebFilter$StaticUserFilter) to context regionserver
2016-12-02 19:20:18,286 INFO  [main] http.HttpServer: Added filter static_user_filter (class=org.apache.hadoop.hbase.http.lib.StaticUserWebFilter$StaticUserFilter) to context static
2016-12-02 19:20:18,286 INFO  [main] http.HttpServer: Added filter static_user_filter (class=org.apache.hadoop.hbase.http.lib.StaticUserWebFilter$StaticUserFilter) to context logs
2016-12-02 19:20:18,294 INFO  [main] http.HttpServer: Jetty bound to port 16030
2016-12-02 19:20:18,294 INFO  [main] mortbay.log: jetty-6.1.26.hwx
2016-12-02 19:20:18,541 INFO  [main] mortbay.log: Started SelectChannelConnector@0.0.0.0:16030
2016-12-02 19:20:18,580 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] zookeeper.RecoverableZooKeeper: Process identifier=hconnection-0x557164c8 connecting to ZooKeeper ensemble=hscale-dev1-dn1:2181,hscale-dev1-dn3:2181,hscale-dev1-dn2:2181,hscale-dev1-dn4:2181
2016-12-02 19:20:18,580 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] zookeeper.ZooKeeper: Initiating client connection, connectString=hscale-dev1-dn1:2181,hscale-dev1-dn3:2181,hscale-dev1-dn2:2181,hscale-dev1-dn4:2181 sessionTimeout=120000 watcher=hconnection-0x557164c80x0, quorum=hscale-dev1-dn1:2181,hscale-dev1-dn3:2181,hscale-dev1-dn2:2181,hscale-dev1-dn4:2181, baseZNode=/hbase-unsecure
2016-12-02 19:20:18,580 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020-SendThread(hscale-dev1-dn2:2181)] zookeeper.ClientCnxn: Opening socket connection to server hscale-dev1-dn2/10.60.70.12:2181. Will not attempt to authenticate using SASL (unknown error)
2016-12-02 19:20:18,581 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020-SendThread(hscale-dev1-dn2:2181)] zookeeper.ClientCnxn: Socket connection established to hscale-dev1-dn2/10.60.70.12:2181, initiating session
2016-12-02 19:20:18,597 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020-SendThread(hscale-dev1-dn2:2181)] zookeeper.ClientCnxn: Session establishment complete on server hscale-dev1-dn2/10.60.70.12:2181, sessionid = 0x258bf3f6848002a, negotiated timeout = 120000
2016-12-02 19:20:18,647 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] regionserver.HRegionServer: ClusterId : e53c756c-5d17-4349-aaa8-6d5c497d937d
2016-12-02 19:20:18,718 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] regionserver.MemStoreFlusher: globalMemStoreLimit=768 M, globalMemStoreLimitLowMark=691.2 M, maxHeap=3 G
2016-12-02 19:20:18,720 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] regionserver.HRegionServer: CompactionChecker runs every 10sec
2016-12-02 19:20:18,748 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] regionserver.RegionServerCoprocessorHost: System coprocessor loading is enabled
2016-12-02 19:20:18,748 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] regionserver.RegionServerCoprocessorHost: Table coprocessor loading is enabled
2016-12-02 19:20:18,816 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config: Created Splice configuration.
2016-12-02 19:20:18,819 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  LOG = [org.apache.log4j.Logger@1e9c026e]
2016-12-02 19:20:18,819 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  activeTransactionCacheSize = [4096]
2016-12-02 19:20:18,820 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  authentication = [NATIVE]
2016-12-02 19:20:18,820 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  authenticationCustomProvider = [null]
2016-12-02 19:20:18,820 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  authenticationLdapSearchauthdn = [null]
2016-12-02 19:20:18,820 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  authenticationLdapSearchauthpw = [null]
2016-12-02 19:20:18,820 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  authenticationLdapSearchbase = [null]
2016-12-02 19:20:18,820 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  authenticationLdapSearchfilter = [null]
2016-12-02 19:20:18,820 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  authenticationLdapServer = [null]
2016-12-02 19:20:18,820 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  authenticationNativeAlgorithm = [SHA-512]
2016-12-02 19:20:18,820 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  authenticationNativeCreateCredentialsDatabase = [true]
2016-12-02 19:20:18,820 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  backupParallelism = [16]
2016-12-02 19:20:18,820 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  backupPath = [/backup]
2016-12-02 19:20:18,820 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  batchOnceBatchSize = [50000]
2016-12-02 19:20:18,820 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  broadcastRegionMbThreshold = [30]
2016-12-02 19:20:18,820 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  broadcastRegionRowThreshold = [1000000]
2016-12-02 19:20:18,820 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  cardinalityPrecision = [14]
2016-12-02 19:20:18,820 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  clientPause = [1000]
2016-12-02 19:20:18,820 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  compactionReservedSlots = [1]
2016-12-02 19:20:18,820 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  completedTxnCacheSize = [131072]
2016-12-02 19:20:18,820 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  completedTxnConcurrency = [128]
2016-12-02 19:20:18,820 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  compressionAlgorithm = [snappy]
2016-12-02 19:20:18,820 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  configSource = [com.splicemachine.access.HBaseConfigurationSource@1b23bf0c]
2016-12-02 19:20:18,820 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  coreWriterThreads = [2]
2016-12-02 19:20:18,820 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  ddlDrainingInitialWait = [1000]
2016-12-02 19:20:18,820 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  ddlDrainingMaximumWait = [100000]
2016-12-02 19:20:18,820 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  ddlRefreshInterval = [10000]
2016-12-02 19:20:18,821 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  debugDumpBindTree = [false]
2016-12-02 19:20:18,821 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  debugDumpClassFile = [false]
2016-12-02 19:20:18,821 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  debugDumpOptimizedTree = [false]
2016-12-02 19:20:18,821 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  debugLogStatementContext = [false]
2016-12-02 19:20:18,821 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.block.access.key.update.interval = [600]
2016-12-02 19:20:18,821 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.block.access.token.enable = [true]
2016-12-02 19:20:18,821 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.block.access.token.lifetime = [600]
2016-12-02 19:20:18,821 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.block.scanner.volume.bytes.per.second = [1048576]
2016-12-02 19:20:18,821 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.blockreport.initialDelay = [120]
2016-12-02 19:20:18,821 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.blockreport.intervalMsec = [21600000]
2016-12-02 19:20:18,821 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.blockreport.split.threshold = [1000000]
2016-12-02 19:20:18,821 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.blocksize = [134217728]
2016-12-02 19:20:18,821 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.bytes-per-checksum = [512]
2016-12-02 19:20:18,821 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.cachereport.intervalMsec = [10000]
2016-12-02 19:20:18,821 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.client-write-packet-size = [65536]
2016-12-02 19:20:18,821 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.client.block.write.replace-datanode-on-failure.best-effort = [false]
2016-12-02 19:20:18,821 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.client.block.write.replace-datanode-on-failure.enable = [true]
2016-12-02 19:20:18,821 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.client.block.write.replace-datanode-on-failure.policy = [DEFAULT]
2016-12-02 19:20:18,821 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.client.block.write.retries = [3]
2016-12-02 19:20:18,821 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.client.cached.conn.retry = [3]
2016-12-02 19:20:18,821 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.client.context = [default]
2016-12-02 19:20:18,821 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.client.datanode-restart.timeout = [30]
2016-12-02 19:20:18,821 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.client.domain.socket.data.traffic = [false]
2016-12-02 19:20:18,821 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.client.failover.connection.retries = [0]
2016-12-02 19:20:18,821 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.client.failover.connection.retries.on.timeouts = [0]
2016-12-02 19:20:18,821 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.client.failover.max.attempts = [15]
2016-12-02 19:20:18,821 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.client.failover.sleep.base.millis = [500]
2016-12-02 19:20:18,821 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.client.failover.sleep.max.millis = [15000]
2016-12-02 19:20:18,821 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.client.file-block-storage-locations.num-threads = [10]
2016-12-02 19:20:18,822 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.client.file-block-storage-locations.timeout.millis = [1000]
2016-12-02 19:20:18,822 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.client.https.keystore.resource = [ssl-client.xml]
2016-12-02 19:20:18,822 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.client.https.need-auth = [false]
2016-12-02 19:20:18,822 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.client.mmap.cache.size = [256]
2016-12-02 19:20:18,822 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.client.mmap.cache.timeout.ms = [3600000]
2016-12-02 19:20:18,822 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.client.mmap.enabled = [true]
2016-12-02 19:20:18,822 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.client.mmap.retry.timeout.ms = [300000]
2016-12-02 19:20:18,822 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.client.read.shortcircuit = [true]
2016-12-02 19:20:18,822 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.client.read.shortcircuit.buffer.size = [131072]
2016-12-02 19:20:18,822 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.client.read.shortcircuit.skip.checksum = [false]
2016-12-02 19:20:18,822 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.client.read.shortcircuit.streams.cache.expiry.ms = [300000]
2016-12-02 19:20:18,822 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.client.read.shortcircuit.streams.cache.size = [4096]
2016-12-02 19:20:18,822 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.client.retry.policy.enabled = [false]
2016-12-02 19:20:18,822 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.client.short.circuit.replica.stale.threshold.ms = [1800000]
2016-12-02 19:20:18,822 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.client.slow.io.warning.threshold.ms = [30000]
2016-12-02 19:20:18,822 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.client.use.datanode.hostname = [false]
2016-12-02 19:20:18,822 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.client.use.legacy.blockreader.local = [false]
2016-12-02 19:20:18,822 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.client.write.exclude.nodes.cache.expiry.interval.millis = [600000]
2016-12-02 19:20:18,822 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.cluster.administrators = [ hdfs]
2016-12-02 19:20:18,822 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.content-summary.limit = [5000]
2016-12-02 19:20:18,822 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.datanode.address = [0.0.0.0:50010]
2016-12-02 19:20:18,822 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.datanode.available-space-volume-choosing-policy.balanced-space-preference-fraction = [0.75f]
2016-12-02 19:20:18,822 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.datanode.available-space-volume-choosing-policy.balanced-space-threshold = [10737418240]
2016-12-02 19:20:18,822 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.datanode.balance.bandwidthPerSec = [6250000]
2016-12-02 19:20:18,822 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.datanode.block-pinning.enabled = [false]
2016-12-02 19:20:18,822 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.datanode.block.id.layout.upgrade.threads = [12]
2016-12-02 19:20:18,822 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.datanode.bp-ready.timeout = [20]
2016-12-02 19:20:18,822 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.datanode.cache.revocation.polling.ms = [500]
2016-12-02 19:20:18,823 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.datanode.cache.revocation.timeout.ms = [900000]
2016-12-02 19:20:18,823 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.datanode.data.dir = [/hadoop/hdfs/data]
2016-12-02 19:20:18,823 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.datanode.data.dir.perm = [750]
2016-12-02 19:20:18,823 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.datanode.directoryscan.interval = [21600]
2016-12-02 19:20:18,823 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.datanode.directoryscan.threads = [1]
2016-12-02 19:20:18,823 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.datanode.dns.interface = [default]
2016-12-02 19:20:18,823 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.datanode.dns.nameserver = [default]
2016-12-02 19:20:18,823 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.datanode.drop.cache.behind.reads = [false]
2016-12-02 19:20:18,823 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.datanode.drop.cache.behind.writes = [false]
2016-12-02 19:20:18,823 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.datanode.du.reserved = [1073741824]
2016-12-02 19:20:18,823 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.datanode.failed.volumes.tolerated = [0]
2016-12-02 19:20:18,823 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.datanode.fsdatasetcache.max.threads.per.volume = [4]
2016-12-02 19:20:18,823 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.datanode.handler.count = [20]
2016-12-02 19:20:18,823 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.datanode.hdfs-blocks-metadata.enabled = [false]
2016-12-02 19:20:18,823 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.datanode.http.address = [0.0.0.0:50075]
2016-12-02 19:20:18,823 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.datanode.https.address = [0.0.0.0:50475]
2016-12-02 19:20:18,823 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.datanode.ipc.address = [0.0.0.0:8010]
2016-12-02 19:20:18,823 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.datanode.max.locked.memory = [0]
2016-12-02 19:20:18,823 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.datanode.max.transfer.threads = [1024]
2016-12-02 19:20:18,823 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.datanode.readahead.bytes = [4193404]
2016-12-02 19:20:18,823 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.datanode.scan.period.hours = [504]
2016-12-02 19:20:18,823 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.datanode.shared.file.descriptor.paths = [/dev/shm,/tmp]
2016-12-02 19:20:18,823 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.datanode.slow.io.warning.threshold.ms = [300]
2016-12-02 19:20:18,823 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.datanode.sync.behind.writes = [false]
2016-12-02 19:20:18,823 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.datanode.transfer.socket.recv.buffer.size = [131072]
2016-12-02 19:20:18,823 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.datanode.transfer.socket.send.buffer.size = [131072]
2016-12-02 19:20:18,823 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.datanode.use.datanode.hostname = [false]
2016-12-02 19:20:18,823 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.default.chunk.view.size = [32768]
2016-12-02 19:20:18,823 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.domain.socket.path = [/var/lib/hadoop-hdfs/dn_socket]
2016-12-02 19:20:18,823 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.encrypt.data.transfer = [false]
2016-12-02 19:20:18,824 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.encrypt.data.transfer.cipher.key.bitlength = [128]
2016-12-02 19:20:18,824 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.encrypt.data.transfer.cipher.suites = [AES/CTR/NoPadding]
2016-12-02 19:20:18,824 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.ha.automatic-failover.enabled = [false]
2016-12-02 19:20:18,824 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.ha.fencing.ssh.connect-timeout = [30000]
2016-12-02 19:20:18,824 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.ha.log-roll.period = [120]
2016-12-02 19:20:18,824 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.ha.tail-edits.period = [60]
2016-12-02 19:20:18,824 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.heartbeat.interval = [3]
2016-12-02 19:20:18,824 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.hosts.exclude = [/etc/hadoop/conf/dfs.exclude]
2016-12-02 19:20:18,824 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.http.policy = [HTTP_ONLY]
2016-12-02 19:20:18,824 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.https.port = [50470]
2016-12-02 19:20:18,824 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.https.server.keystore.resource = [ssl-server.xml]
2016-12-02 19:20:18,824 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.image.compress = [false]
2016-12-02 19:20:18,824 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.image.compression.codec = [org.apache.hadoop.io.compress.DefaultCodec]
2016-12-02 19:20:18,824 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.image.transfer.bandwidthPerSec = [0]
2016-12-02 19:20:18,824 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.image.transfer.chunksize = [65536]
2016-12-02 19:20:18,824 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.image.transfer.timeout = [60000]
2016-12-02 19:20:18,824 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.journalnode.edits.dir = [/hadoop/hdfs/journalnode]
2016-12-02 19:20:18,824 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.journalnode.http-address = [0.0.0.0:8480]
2016-12-02 19:20:18,824 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.journalnode.https-address = [0.0.0.0:8481]
2016-12-02 19:20:18,824 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.journalnode.rpc-address = [0.0.0.0:8485]
2016-12-02 19:20:18,824 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.namenode.accesstime.precision = [0]
2016-12-02 19:20:18,824 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.namenode.acls.enabled = [false]
2016-12-02 19:20:18,824 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.namenode.audit.log.async = [true]
2016-12-02 19:20:18,824 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.namenode.audit.loggers = [default]
2016-12-02 19:20:18,824 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.namenode.avoid.read.stale.datanode = [true]
2016-12-02 19:20:18,824 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.namenode.avoid.write.stale.datanode = [true]
2016-12-02 19:20:18,824 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.namenode.backup.address = [0.0.0.0:50100]
2016-12-02 19:20:18,824 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.namenode.backup.http-address = [0.0.0.0:50105]
2016-12-02 19:20:18,824 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.namenode.blocks.per.postponedblocks.rescan = [10000]
2016-12-02 19:20:18,824 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.namenode.checkpoint.check.period = [60]
2016-12-02 19:20:18,824 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.namenode.checkpoint.dir = [/hadoop/hdfs/namesecondary]
2016-12-02 19:20:18,824 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.namenode.checkpoint.edits.dir = [/hadoop/hdfs/namesecondary]
2016-12-02 19:20:18,824 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.namenode.checkpoint.max-retries = [3]
2016-12-02 19:20:18,824 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.namenode.checkpoint.period = [21600]
2016-12-02 19:20:18,824 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.namenode.checkpoint.txns = [1000000]
2016-12-02 19:20:18,825 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.namenode.datanode.registration.ip-hostname-check = [true]
2016-12-02 19:20:18,825 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.namenode.decommission.blocks.per.interval = [500000]
2016-12-02 19:20:18,825 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.namenode.decommission.interval = [30]
2016-12-02 19:20:18,825 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.namenode.decommission.max.concurrent.tracked.nodes = [100]
2016-12-02 19:20:18,825 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.namenode.delegation.key.update-interval = [86400000]
2016-12-02 19:20:18,825 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.namenode.delegation.token.max-lifetime = [604800000]
2016-12-02 19:20:18,825 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.namenode.delegation.token.renew-interval = [86400000]
2016-12-02 19:20:18,825 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.namenode.edit.log.autoroll.check.interval.ms = [300000]
2016-12-02 19:20:18,825 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.namenode.edit.log.autoroll.multiplier.threshold = [2.0]
2016-12-02 19:20:18,825 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.namenode.edits.dir = [/hadoop/hdfs/namenode]
2016-12-02 19:20:18,825 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.namenode.edits.journal-plugin.qjournal = [org.apache.hadoop.hdfs.qjournal.client.QuorumJournalManager]
2016-12-02 19:20:18,825 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.namenode.edits.noeditlogchannelflush = [false]
2016-12-02 19:20:18,825 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.namenode.enable.retrycache = [true]
2016-12-02 19:20:18,825 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.namenode.fs-limits.max-blocks-per-file = [1048576]
2016-12-02 19:20:18,825 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.namenode.fs-limits.max-component-length = [255]
2016-12-02 19:20:18,825 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.namenode.fs-limits.max-directory-items = [1048576]
2016-12-02 19:20:18,825 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.namenode.fs-limits.max-xattr-size = [16384]
2016-12-02 19:20:18,825 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.namenode.fs-limits.max-xattrs-per-inode = [32]
2016-12-02 19:20:18,825 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.namenode.fs-limits.min-block-size = [1048576]
2016-12-02 19:20:18,825 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.namenode.fslock.fair = [false]
2016-12-02 19:20:18,825 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.namenode.full.block.report.lease.length.ms = [300000]
2016-12-02 19:20:18,825 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.namenode.handler.count = [100]
2016-12-02 19:20:18,825 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.namenode.heartbeat.recheck-interval = [300000]
2016-12-02 19:20:18,825 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.namenode.http-address = [hscale-dev1-nn:50070]
2016-12-02 19:20:18,825 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.namenode.https-address = [hscale-dev1-nn:50470]
2016-12-02 19:20:18,825 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.namenode.inotify.max.events.per.rpc = [1000]
2016-12-02 19:20:18,825 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.namenode.invalidate.work.pct.per.iteration = [0.32f]
2016-12-02 19:20:18,825 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.namenode.kerberos.internal.spnego.principal = [${dfs.web.authentication.kerberos.principal}]
2016-12-02 19:20:18,825 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.namenode.kerberos.principal.pattern = [*]
2016-12-02 19:20:18,825 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.namenode.lazypersist.file.scrub.interval.sec = [300]
2016-12-02 19:20:18,825 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.namenode.lifeline.handler.ratio = [0.10]
2016-12-02 19:20:18,826 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.namenode.list.cache.directives.num.responses = [100]
2016-12-02 19:20:18,826 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.namenode.list.cache.pools.num.responses = [100]
2016-12-02 19:20:18,826 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.namenode.list.encryption.zones.num.responses = [100]
2016-12-02 19:20:18,826 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.namenode.max.extra.edits.segments.retained = [10000]
2016-12-02 19:20:18,826 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.namenode.max.full.block.report.leases = [6]
2016-12-02 19:20:18,826 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.namenode.max.objects = [0]
2016-12-02 19:20:18,826 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.namenode.metrics.logger.period.seconds = [600]
2016-12-02 19:20:18,826 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.namenode.name.dir = [/hadoop/hdfs/namenode]
2016-12-02 19:20:18,826 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.namenode.name.dir.restore = [true]
2016-12-02 19:20:18,826 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.namenode.num.checkpoints.retained = [2]
2016-12-02 19:20:18,826 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.namenode.num.extra.edits.retained = [1000000]
2016-12-02 19:20:18,826 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.namenode.path.based.cache.block.map.allocation.percent = [0.25]
2016-12-02 19:20:18,826 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.namenode.path.based.cache.refresh.interval.ms = [30000]
2016-12-02 19:20:18,826 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.namenode.path.based.cache.retry.interval.ms = [30000]
2016-12-02 19:20:18,826 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.namenode.reject-unresolved-dn-topology-mapping = [false]
2016-12-02 19:20:18,826 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.namenode.replication.considerLoad = [true]
2016-12-02 19:20:18,826 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.namenode.replication.interval = [3]
2016-12-02 19:20:18,826 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.namenode.replication.min = [1]
2016-12-02 19:20:18,826 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.namenode.replication.work.multiplier.per.iteration = [2]
2016-12-02 19:20:18,826 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.namenode.resource.check.interval = [5000]
2016-12-02 19:20:18,826 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.namenode.resource.checked.volumes.minimum = [1]
2016-12-02 19:20:18,826 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.namenode.resource.du.reserved = [104857600]
2016-12-02 19:20:18,826 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.namenode.retrycache.expirytime.millis = [600000]
2016-12-02 19:20:18,826 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.namenode.retrycache.heap.percent = [0.03f]
2016-12-02 19:20:18,826 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.namenode.rpc-address = [hscale-dev1-nn:8020]
2016-12-02 19:20:18,826 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.namenode.safemode.extension = [30000]
2016-12-02 19:20:18,826 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.namenode.safemode.min.datanodes = [0]
2016-12-02 19:20:18,826 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.namenode.safemode.threshold-pct = [0.999]
2016-12-02 19:20:18,826 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.namenode.secondary.http-address = [hscale-dev1-dn1:50090]
2016-12-02 19:20:18,826 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.namenode.secondary.https-address = [0.0.0.0:50091]
2016-12-02 19:20:18,826 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.namenode.stale.datanode.interval = [30000]
2016-12-02 19:20:18,826 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.namenode.startup.delay.block.deletion.sec = [3600]
2016-12-02 19:20:18,826 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.namenode.support.allow.format = [true]
2016-12-02 19:20:18,826 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.namenode.top.enabled = [true]
2016-12-02 19:20:18,826 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.namenode.top.num.users = [10]
2016-12-02 19:20:18,826 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.namenode.top.window.num.buckets = [10]
2016-12-02 19:20:18,826 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.namenode.top.windows.minutes = [1,5,25]
2016-12-02 19:20:18,827 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.namenode.write.stale.datanode.ratio = [1.0f]
2016-12-02 19:20:18,827 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.namenode.xattrs.enabled = [true]
2016-12-02 19:20:18,827 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.permissions.enabled = [true]
2016-12-02 19:20:18,827 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.permissions.superusergroup = [hdfs]
2016-12-02 19:20:18,827 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.replication = [3]
2016-12-02 19:20:18,827 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.replication.max = [50]
2016-12-02 19:20:18,827 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.secondary.namenode.kerberos.internal.spnego.principal = [${dfs.web.authentication.kerberos.principal}]
2016-12-02 19:20:18,827 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.short.circuit.shared.memory.watcher.interrupt.check.ms = [60000]
2016-12-02 19:20:18,827 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.storage.policy.enabled = [true]
2016-12-02 19:20:18,827 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.stream-buffer-size = [4096]
2016-12-02 19:20:18,827 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.support.append = [true]
2016-12-02 19:20:18,827 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.user.home.dir.prefix = [/user]
2016-12-02 19:20:18,827 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.webhdfs.enabled = [true]
2016-12-02 19:20:18,827 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.webhdfs.rest-csrf.browser-useragents-regex = [^Mozilla.*,^Opera.*]
2016-12-02 19:20:18,827 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.webhdfs.rest-csrf.custom-header = [X-XSRF-HEADER]
2016-12-02 19:20:18,827 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.webhdfs.rest-csrf.enabled = [false]
2016-12-02 19:20:18,827 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.webhdfs.rest-csrf.methods-to-ignore = [GET,OPTIONS,HEAD,TRACE]
2016-12-02 19:20:18,827 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.webhdfs.ugi.expire.after.access = [600000]
2016-12-02 19:20:18,827 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  dfs.webhdfs.user.provider.user.pattern = [^[A-Za-z_][A-Za-z0-9._-]*[$]?$]
2016-12-02 19:20:18,827 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  fallbackLocalLatency = [1]
2016-12-02 19:20:18,827 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  fallbackMinimumRowCount = [20]
2016-12-02 19:20:18,827 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  fallbackNullFraction = [0.1]
2016-12-02 19:20:18,827 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  fallbackOpencloseLatency = [2000]
2016-12-02 19:20:18,827 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  fallbackRegionRowCount = [5000000]
2016-12-02 19:20:18,827 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  fallbackRemoteLatencyRatio = [10]
2016-12-02 19:20:18,827 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  fallbackRowWidth = [170]
2016-12-02 19:20:18,827 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  file.blocksize = [67108864]
2016-12-02 19:20:18,827 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  file.bytes-per-checksum = [512]
2016-12-02 19:20:18,827 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  file.client-write-packet-size = [65536]
2016-12-02 19:20:18,827 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  file.replication = [1]
2016-12-02 19:20:18,827 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  file.stream-buffer-size = [4096]
2016-12-02 19:20:18,827 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  fs.AbstractFileSystem.file.impl = [org.apache.hadoop.fs.local.LocalFs]
2016-12-02 19:20:18,827 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  fs.AbstractFileSystem.ftp.impl = [org.apache.hadoop.fs.ftp.FtpFs]
2016-12-02 19:20:18,828 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  fs.AbstractFileSystem.har.impl = [org.apache.hadoop.fs.HarFs]
2016-12-02 19:20:18,828 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  fs.AbstractFileSystem.hdfs.impl = [org.apache.hadoop.fs.Hdfs]
2016-12-02 19:20:18,828 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  fs.AbstractFileSystem.swebhdfs.impl = [org.apache.hadoop.fs.SWebHdfs]
2016-12-02 19:20:18,828 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  fs.AbstractFileSystem.viewfs.impl = [org.apache.hadoop.fs.viewfs.ViewFs]
2016-12-02 19:20:18,828 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  fs.AbstractFileSystem.webhdfs.impl = [org.apache.hadoop.fs.WebHdfs]
2016-12-02 19:20:18,828 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  fs.automatic.close = [true]
2016-12-02 19:20:18,828 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  fs.client.resolve.remote.symlinks = [true]
2016-12-02 19:20:18,828 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  fs.defaultFS = [hdfs://hscale-dev1-nn:8020]
2016-12-02 19:20:18,828 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  fs.df.interval = [60000]
2016-12-02 19:20:18,828 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  fs.du.interval = [600000]
2016-12-02 19:20:18,828 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  fs.ftp.host = [0.0.0.0]
2016-12-02 19:20:18,828 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  fs.ftp.host.port = [21]
2016-12-02 19:20:18,828 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  fs.har.impl.disable.cache = [true]
2016-12-02 19:20:18,828 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  fs.permissions.umask-mode = [022]
2016-12-02 19:20:18,828 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  fs.s3.block.size = [67108864]
2016-12-02 19:20:18,828 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  fs.s3.buffer.dir = [/tmp/hadoop-hbase/s3]
2016-12-02 19:20:18,828 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  fs.s3.maxRetries = [4]
2016-12-02 19:20:18,828 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  fs.s3.sleepTimeSeconds = [10]
2016-12-02 19:20:18,828 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  fs.s3a.attempts.maximum = [10]
2016-12-02 19:20:18,828 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  fs.s3a.buffer.dir = [/tmp/hadoop-hbase/s3a]
2016-12-02 19:20:18,828 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  fs.s3a.connection.establish.timeout = [5000]
2016-12-02 19:20:18,828 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  fs.s3a.connection.maximum = [15]
2016-12-02 19:20:18,828 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  fs.s3a.connection.ssl.enabled = [true]
2016-12-02 19:20:18,828 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  fs.s3a.connection.timeout = [50000]
2016-12-02 19:20:18,828 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  fs.s3a.fast.buffer.size = [1048576]
2016-12-02 19:20:18,828 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  fs.s3a.fast.upload = [false]
2016-12-02 19:20:18,828 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  fs.s3a.impl = [org.apache.hadoop.fs.s3a.S3AFileSystem]
2016-12-02 19:20:18,828 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  fs.s3a.max.total.tasks = [1000]
2016-12-02 19:20:18,828 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  fs.s3a.multipart.purge = [false]
2016-12-02 19:20:18,828 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  fs.s3a.multipart.purge.age = [86400]
2016-12-02 19:20:18,828 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  fs.s3a.multipart.size = [104857600]
2016-12-02 19:20:18,828 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  fs.s3a.multipart.threshold = [2147483647]
2016-12-02 19:20:18,828 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  fs.s3a.paging.maximum = [5000]
2016-12-02 19:20:18,828 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  fs.s3a.threads.core = [15]
2016-12-02 19:20:18,828 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  fs.s3a.threads.keepalivetime = [60]
2016-12-02 19:20:18,828 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  fs.s3a.threads.max = [256]
2016-12-02 19:20:18,828 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  fs.s3n.block.size = [67108864]
2016-12-02 19:20:18,828 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  fs.s3n.multipart.copy.block.size = [5368709120]
2016-12-02 19:20:18,828 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  fs.s3n.multipart.uploads.block.size = [67108864]
2016-12-02 19:20:18,829 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  fs.s3n.multipart.uploads.enabled = [false]
2016-12-02 19:20:18,829 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  fs.swift.impl = [org.apache.hadoop.fs.swift.snative.SwiftNativeFileSystem]
2016-12-02 19:20:18,829 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  fs.trash.checkpoint.interval = [0]
2016-12-02 19:20:18,829 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  fs.trash.interval = [360]
2016-12-02 19:20:18,829 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  ftp.blocksize = [67108864]
2016-12-02 19:20:18,829 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  ftp.bytes-per-checksum = [512]
2016-12-02 19:20:18,829 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  ftp.client-write-packet-size = [65536]
2016-12-02 19:20:18,829 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  ftp.replication = [3]
2016-12-02 19:20:18,829 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  ftp.stream-buffer-size = [4096]
2016-12-02 19:20:18,829 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  ha.failover-controller.active-standby-elector.zk.op.retries = [120]
2016-12-02 19:20:18,829 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  ha.failover-controller.cli-check.rpc-timeout.ms = [20000]
2016-12-02 19:20:18,829 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  ha.failover-controller.graceful-fence.connection.retries = [1]
2016-12-02 19:20:18,829 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  ha.failover-controller.graceful-fence.rpc-timeout.ms = [5000]
2016-12-02 19:20:18,829 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  ha.failover-controller.new-active.rpc-timeout.ms = [60000]
2016-12-02 19:20:18,829 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  ha.health-monitor.check-interval.ms = [1000]
2016-12-02 19:20:18,829 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  ha.health-monitor.connect-retry-interval.ms = [1000]
2016-12-02 19:20:18,829 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  ha.health-monitor.rpc-timeout.ms = [45000]
2016-12-02 19:20:18,829 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  ha.health-monitor.sleep-after-disconnect.ms = [1000]
2016-12-02 19:20:18,829 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  ha.zookeeper.acl = [world:anyone:rwcda]
2016-12-02 19:20:18,829 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  ha.zookeeper.parent-znode = [/hadoop-ha]
2016-12-02 19:20:18,829 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  ha.zookeeper.session-timeout.ms = [5000]
2016-12-02 19:20:18,829 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hadoop.common.configuration.version = [0.23.0]
2016-12-02 19:20:18,829 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hadoop.fuse.connection.timeout = [300]
2016-12-02 19:20:18,829 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hadoop.fuse.timer.period = [5]
2016-12-02 19:20:18,829 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hadoop.hdfs.configuration.version = [1]
2016-12-02 19:20:18,829 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hadoop.http.authentication.kerberos.keytab = [/home/hbase/hadoop.keytab]
2016-12-02 19:20:18,829 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hadoop.http.authentication.kerberos.principal = [HTTP/_HOST@LOCALHOST]
2016-12-02 19:20:18,829 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hadoop.http.authentication.signature.secret.file = [/home/hbase/hadoop-http-auth-signature-secret]
2016-12-02 19:20:18,829 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hadoop.http.authentication.simple.anonymous.allowed = [true]
2016-12-02 19:20:18,829 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hadoop.http.authentication.token.validity = [36000]
2016-12-02 19:20:18,829 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hadoop.http.authentication.type = [simple]
2016-12-02 19:20:18,829 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hadoop.http.cross-origin.allowed-headers = [X-Requested-With,Content-Type,Accept,Origin]
2016-12-02 19:20:18,829 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hadoop.http.cross-origin.allowed-methods = [GET,POST,HEAD]
2016-12-02 19:20:18,829 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hadoop.http.cross-origin.allowed-origins = [*]
2016-12-02 19:20:18,829 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hadoop.http.cross-origin.enabled = [false]
2016-12-02 19:20:18,829 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hadoop.http.cross-origin.max-age = [1800]
2016-12-02 19:20:18,829 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hadoop.http.filter.initializers = [org.apache.hadoop.http.lib.StaticUserWebFilter,org.apache.hadoop.security.HttpCrossOriginFilterInitializer]
2016-12-02 19:20:18,829 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hadoop.http.staticuser.user = [dr.who]
2016-12-02 19:20:18,829 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hadoop.jetty.logs.serve.aliases = [true]
2016-12-02 19:20:18,829 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hadoop.kerberos.kinit.command = [kinit]
2016-12-02 19:20:18,829 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hadoop.policy.file = [hbase-policy.xml]
2016-12-02 19:20:18,829 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hadoop.proxyuser.hbase.groups = [*]
2016-12-02 19:20:18,829 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hadoop.proxyuser.hbase.hosts = [*]
2016-12-02 19:20:18,830 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hadoop.proxyuser.hcat.groups = [*]
2016-12-02 19:20:18,830 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hadoop.proxyuser.hcat.hosts = [hscale-dev1-dn2]
2016-12-02 19:20:18,830 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hadoop.proxyuser.hdfs.groups = [*]
2016-12-02 19:20:18,830 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hadoop.proxyuser.hdfs.hosts = [*]
2016-12-02 19:20:18,830 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hadoop.proxyuser.hive.groups = [*]
2016-12-02 19:20:18,830 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hadoop.proxyuser.hive.hosts = [hscale-dev1-dn2]
2016-12-02 19:20:18,830 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hadoop.proxyuser.oozie.groups = [*]
2016-12-02 19:20:18,830 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hadoop.proxyuser.oozie.hosts = [hscale-dev1-dn3]
2016-12-02 19:20:18,830 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hadoop.registry.jaas.context = [Client]
2016-12-02 19:20:18,830 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hadoop.registry.rm.enabled = [false]
2016-12-02 19:20:18,830 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hadoop.registry.secure = [false]
2016-12-02 19:20:18,830 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hadoop.registry.system.acls = [sasl:yarn@, sasl:mapred@, sasl:hdfs@]
2016-12-02 19:20:18,830 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hadoop.registry.zk.connection.timeout.ms = [15000]
2016-12-02 19:20:18,830 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hadoop.registry.zk.quorum = [hscale-dev1-dn1:2181,hscale-dev1-dn3:2181,hscale-dev1-dn2:2181,hscale-dev1-dn4:2181]
2016-12-02 19:20:18,830 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hadoop.registry.zk.retry.ceiling.ms = [60000]
2016-12-02 19:20:18,830 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hadoop.registry.zk.retry.interval.ms = [1000]
2016-12-02 19:20:18,830 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hadoop.registry.zk.retry.times = [5]
2016-12-02 19:20:18,830 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hadoop.registry.zk.root = [/registry]
2016-12-02 19:20:18,830 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hadoop.registry.zk.session.timeout.ms = [60000]
2016-12-02 19:20:18,830 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hadoop.rpc.protection = [authentication]
2016-12-02 19:20:18,830 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hadoop.rpc.socket.factory.class.default = [org.apache.hadoop.net.StandardSocketFactory]
2016-12-02 19:20:18,830 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hadoop.security.auth_to_local = [RULE:[1:$1@$0](ambari-qa@HSCALE.COM)s/.*/ambari-qa/
RULE:[1:$1@$0](hbase@HSCALE.COM)s/.*/hbase/
RULE:[1:$1@$0](hdfs@HSCALE.COM)s/.*/hdfs/
RULE:[1:$1@$0](spark@HSCALE.COM)s/.*/spark/
RULE:[1:$1@$0](.*@HSCALE.COM)s/@.*//
RULE:[2:$1@$0](amshbase@HSCALE.COM)s/.*/ams/
RULE:[2:$1@$0](amszk@HSCALE.COM)s/.*/ams/
RULE:[2:$1@$0](dn@HSCALE.COM)s/.*/hdfs/
RULE:[2:$1@$0](hbase@HSCALE.COM)s/.*/hbase/
RULE:[2:$1@$0](hive@HSCALE.COM)s/.*/hive/
RULE:[2:$1@$0](jhs@HSCALE.COM)s/.*/mapred/
RULE:[2:$1@$0](nm@HSCALE.COM)s/.*/yarn/
RULE:[2:$1@$0](nn@HSCALE.COM)s/.*/hdfs/
RULE:[2:$1@$0](oozie@HSCALE.COM)s/.*/oozie/
RULE:[2:$1@$0](rm@HSCALE.COM)s/.*/yarn/
RULE:[2:$1@$0](yarn@HSCALE.COM)s/.*/yarn/
DEFAULT]
2016-12-02 19:20:18,830 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hadoop.security.authentication = [simple]
2016-12-02 19:20:18,830 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hadoop.security.authorization = [false]
2016-12-02 19:20:18,830 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hadoop.security.crypto.buffer.size = [8192]
2016-12-02 19:20:18,830 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hadoop.security.crypto.cipher.suite = [AES/CTR/NoPadding]
2016-12-02 19:20:18,830 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hadoop.security.crypto.codec.classes.aes.ctr.nopadding = [org.apache.hadoop.crypto.OpensslAesCtrCryptoCodec,org.apache.hadoop.crypto.JceAesCtrCryptoCodec]
2016-12-02 19:20:18,830 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hadoop.security.dns.log-slow-lookups.enabled = [false]
2016-12-02 19:20:18,830 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hadoop.security.dns.log-slow-lookups.threshold.ms = [1000]
2016-12-02 19:20:18,830 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hadoop.security.group.mapping = [org.apache.hadoop.security.JniBasedUnixGroupsMappingWithFallback]
2016-12-02 19:20:18,830 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hadoop.security.group.mapping.ldap.directory.search.timeout = [10000]
2016-12-02 19:20:18,830 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hadoop.security.group.mapping.ldap.search.attr.group.name = [cn]
2016-12-02 19:20:18,830 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hadoop.security.group.mapping.ldap.search.attr.member = [member]
2016-12-02 19:20:18,830 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hadoop.security.group.mapping.ldap.search.filter.group = [(objectClass=group)]
2016-12-02 19:20:18,830 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hadoop.security.group.mapping.ldap.search.filter.user = [(&(objectClass=user)(sAMAccountName={0}))]
2016-12-02 19:20:18,830 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hadoop.security.group.mapping.ldap.ssl = [false]
2016-12-02 19:20:18,830 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hadoop.security.groups.cache.secs = [300]
2016-12-02 19:20:18,830 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hadoop.security.groups.cache.warn.after.ms = [5000]
2016-12-02 19:20:18,830 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hadoop.security.groups.negative-cache.secs = [30]
2016-12-02 19:20:18,830 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hadoop.security.instrumentation.requires.admin = [false]
2016-12-02 19:20:18,830 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hadoop.security.java.secure.random.algorithm = [SHA1PRNG]
2016-12-02 19:20:18,831 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hadoop.security.kms.client.authentication.retry-count = [1]
2016-12-02 19:20:18,831 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hadoop.security.kms.client.encrypted.key.cache.expiry = [43200000]
2016-12-02 19:20:18,831 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hadoop.security.kms.client.encrypted.key.cache.low-watermark = [0.3f]
2016-12-02 19:20:18,831 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hadoop.security.kms.client.encrypted.key.cache.num.refill.threads = [2]
2016-12-02 19:20:18,831 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hadoop.security.kms.client.encrypted.key.cache.size = [500]
2016-12-02 19:20:18,831 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hadoop.security.random.device.file.path = [/dev/urandom]
2016-12-02 19:20:18,831 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hadoop.security.uid.cache.secs = [14400]
2016-12-02 19:20:18,831 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hadoop.shell.safely.delete.limit.num.files = [100]
2016-12-02 19:20:18,831 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hadoop.ssl.client.conf = [ssl-client.xml]
2016-12-02 19:20:18,831 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hadoop.ssl.enabled = [false]
2016-12-02 19:20:18,831 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hadoop.ssl.enabled.protocols = [TLSv1]
2016-12-02 19:20:18,831 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hadoop.ssl.hostname.verifier = [DEFAULT]
2016-12-02 19:20:18,831 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hadoop.ssl.keystores.factory.class = [org.apache.hadoop.security.ssl.FileBasedKeyStoresFactory]
2016-12-02 19:20:18,831 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hadoop.ssl.require.client.cert = [false]
2016-12-02 19:20:18,831 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hadoop.ssl.server.conf = [ssl-server.xml]
2016-12-02 19:20:18,831 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hadoop.tmp.dir = [/tmp/hadoop-hbase]
2016-12-02 19:20:18,831 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hadoop.user.group.static.mapping.overrides = [dr.who=;]
2016-12-02 19:20:18,831 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hadoop.util.hash.type = [murmur]
2016-12-02 19:20:18,831 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hadoop.work.around.non.threadsafe.getpwuid = [false]
2016-12-02 19:20:18,831 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.auth.key.update.interval = [86400000]
2016-12-02 19:20:18,831 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.auth.token.max.lifetime = [604800000]
2016-12-02 19:20:18,831 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.balancer.period = [60000]
2016-12-02 19:20:18,831 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.bulkload.retries.number = [10]
2016-12-02 19:20:18,831 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.bulkload.staging.dir = [/apps/hbase/staging]
2016-12-02 19:20:18,831 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.cells.scanned.per.heartbeat.check = [10000]
2016-12-02 19:20:18,831 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.client.ipc.pool.size = [10]
2016-12-02 19:20:18,831 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.client.keyvalue.maxsize = [1048576]
2016-12-02 19:20:18,831 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.client.localityCheck.threadPoolSize = [2]
2016-12-02 19:20:18,831 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.client.max.perregion.tasks = [100]
2016-12-02 19:20:18,831 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.client.max.perserver.tasks = [5]
2016-12-02 19:20:18,831 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.client.max.total.tasks = [100]
2016-12-02 19:20:18,831 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.client.pause = [100]
2016-12-02 19:20:18,831 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.client.retries.number = [5]
2016-12-02 19:20:18,831 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.client.scanner.caching = [1000]
2016-12-02 19:20:18,831 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.client.scanner.timeout.period = [60000]
2016-12-02 19:20:18,831 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.client.write.buffer = [2097152]
2016-12-02 19:20:18,831 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.cluster.distributed = [true]
2016-12-02 19:20:18,831 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.column.max.version = [1]
2016-12-02 19:20:18,832 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.config.read.zookeeper.config = [false]
2016-12-02 19:20:18,832 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.coordinated.state.manager.class = [org.apache.hadoop.hbase.coordination.ZkCoordinatedStateManager]
2016-12-02 19:20:18,832 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.coprocessor.abortonerror = [true]
2016-12-02 19:20:18,832 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.coprocessor.enabled = [true]
2016-12-02 19:20:18,832 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.coprocessor.master.classes = [com.splicemachine.hbase.SpliceMasterObserver]
2016-12-02 19:20:18,832 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.coprocessor.region.classes = [org.apache.hadoop.hbase.security.access.SecureBulkLoadEndpoint,com.splicemachine.hbase.MemstoreAwareObserver,com.splicemachine.derby.hbase.SpliceIndexObserver,com.splicemachine.derby.hbase.SpliceIndexEndpoint,com.splicemachine.hbase.RegionSizeEndpoint,com.splicemachine.si.data.hbase.coprocessor.TxnLifecycleEndpoint,com.splicemachine.si.data.hbase.coprocessor.SIObserver,com.splicemachine.hbase.BackupEndpointObserver]
2016-12-02 19:20:18,832 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.coprocessor.regionserver.classes = [com.splicemachine.hbase.RegionServerLifecycleObserver]
2016-12-02 19:20:18,832 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.coprocessor.user.enabled = [true]
2016-12-02 19:20:18,832 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.data.umask = [000]
2016-12-02 19:20:18,832 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.data.umask.enable = [false]
2016-12-02 19:20:18,832 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.defaults.for.version = [1.1.2.2.4.2.0-258]
2016-12-02 19:20:18,832 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.defaults.for.version.skip = [true]
2016-12-02 19:20:18,832 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.dfs.client.read.shortcircuit.buffer.size = [131072]
2016-12-02 19:20:18,832 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.dynamic.jars.dir = [hdfs://hscale-dev1-nn:8020/apps/hbase/data/lib]
2016-12-02 19:20:18,832 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.fs.tmp.dir = [/user/hbase/hbase-staging]
2016-12-02 19:20:18,832 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.hregion.majorcompaction = [604800000]
2016-12-02 19:20:18,832 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.hregion.majorcompaction.jitter = [0.50]
2016-12-02 19:20:18,832 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.hregion.max.filesize = [10737418240]
2016-12-02 19:20:18,832 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.hregion.memstore.block.multiplier = [4]
2016-12-02 19:20:18,832 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.hregion.memstore.flush.size = [134217728]
2016-12-02 19:20:18,832 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.hregion.memstore.mslab.enabled = [true]
2016-12-02 19:20:18,832 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.hregion.percolumnfamilyflush.size.lower.bound = [16777216]
2016-12-02 19:20:18,832 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.hregion.preclose.flush.size = [5242880]
2016-12-02 19:20:18,832 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.hstore.blockingStoreFiles = [20]
2016-12-02 19:20:18,832 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.hstore.blockingWaitTime = [90000]
2016-12-02 19:20:18,832 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.hstore.bytes.per.checksum = [16384]
2016-12-02 19:20:18,832 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.hstore.checksum.algorithm = [CRC32]
2016-12-02 19:20:18,832 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.hstore.compaction.kv.max = [10]
2016-12-02 19:20:18,832 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.hstore.compaction.max = [10]
2016-12-02 19:20:18,832 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.hstore.compaction.max.size = [260046848]
2016-12-02 19:20:18,832 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.hstore.compaction.min = [5]
2016-12-02 19:20:18,832 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.hstore.compaction.min.size = [16777216]
2016-12-02 19:20:18,832 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.hstore.compactionThreshold = [3]
2016-12-02 19:20:18,832 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.hstore.defaultengine.compactionpolicy.class = [com.splicemachine.compactions.SpliceDefaultCompactionPolicy]
2016-12-02 19:20:18,832 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.hstore.defaultengine.compactor.class = [com.splicemachine.compactions.SpliceDefaultCompactor]
2016-12-02 19:20:18,832 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.hstore.flusher.count = [2]
2016-12-02 19:20:18,832 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.hstore.time.to.purge.deletes = [0]
2016-12-02 19:20:18,832 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.htable.threads.max = [96]
2016-12-02 19:20:18,832 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.http.filter.initializers = [org.apache.hadoop.hbase.http.lib.StaticUserWebFilter]
2016-12-02 19:20:18,832 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.http.max.threads = [10]
2016-12-02 19:20:18,832 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.http.staticuser.user = [dr.stack]
2016-12-02 19:20:18,832 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.ipc.client.fallback-to-simple-auth-allowed = [false]
2016-12-02 19:20:18,832 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.ipc.client.tcpnodelay = [true]
2016-12-02 19:20:18,833 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.ipc.server.callqueue.handler.factor = [0.1]
2016-12-02 19:20:18,833 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.ipc.server.callqueue.read.ratio = [0]
2016-12-02 19:20:18,833 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.ipc.server.callqueue.scan.ratio = [0]
2016-12-02 19:20:18,833 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.ipc.warn.response.size = [-1]
2016-12-02 19:20:18,833 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.ipc.warn.response.time = [-1]
2016-12-02 19:20:18,833 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.lease.recovery.dfs.timeout = [64000]
2016-12-02 19:20:18,833 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.lease.recovery.timeout = [900000]
2016-12-02 19:20:18,833 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.local.dir = [/tmp/hbase-hbase/local]
2016-12-02 19:20:18,833 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.master.catalog.timeout = [600000]
2016-12-02 19:20:18,833 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.master.distributed.log.replay = [false]
2016-12-02 19:20:18,833 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.master.hfilecleaner.plugins = [org.apache.hadoop.hbase.master.cleaner.TimeToLiveHFileCleaner]
2016-12-02 19:20:18,833 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.master.info.bindAddress = [0.0.0.0]
2016-12-02 19:20:18,833 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.master.info.port = [16010]
2016-12-02 19:20:18,833 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.master.infoserver.redirect = [true]
2016-12-02 19:20:18,833 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.master.loadbalance.bytable = [TRUE]
2016-12-02 19:20:18,833 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.master.loadbalancer.class = [org.apache.hadoop.hbase.master.balancer.StochasticLoadBalancer]
2016-12-02 19:20:18,833 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.master.logcleaner.plugins = [org.apache.hadoop.hbase.master.cleaner.TimeToLiveLogCleaner]
2016-12-02 19:20:18,833 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.master.logcleaner.ttl = [600000]
2016-12-02 19:20:18,833 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.master.normalizer.class = [org.apache.hadoop.hbase.master.normalizer.SimpleRegionNormalizer]
2016-12-02 19:20:18,833 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.master.port = [16000]
2016-12-02 19:20:18,833 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.metrics.exposeOperationTimes = [true]
2016-12-02 19:20:18,833 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.metrics.showTableName = [true]
2016-12-02 19:20:18,833 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.mvcc.impl = [org.apache.hadoop.hbase.regionserver.SIMultiVersionConsistencyControl]
2016-12-02 19:20:18,833 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.normalizer.enabled = [false]
2016-12-02 19:20:18,833 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.normalizer.period = [1800000]
2016-12-02 19:20:18,833 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.online.schema.update.enable = [true]
2016-12-02 19:20:18,833 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.region.replica.replication.enabled = [false]
2016-12-02 19:20:18,833 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.regions.slop = [0.01]
2016-12-02 19:20:18,833 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.regionserver.catalog.timeout = [600000]
2016-12-02 19:20:18,833 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.regionserver.checksum.verify = [true]
2016-12-02 19:20:18,833 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.regionserver.dns.interface = [default]
2016-12-02 19:20:18,833 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.regionserver.dns.nameserver = [default]
2016-12-02 19:20:18,833 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.regionserver.global.memstore.size = [0.25]
2016-12-02 19:20:18,833 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.regionserver.global.memstore.size.lower.limit = [0.9]
2016-12-02 19:20:18,833 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.regionserver.handler.abort.on.error.percent = [0.5]
2016-12-02 19:20:18,833 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.regionserver.handler.count = [100]
2016-12-02 19:20:18,833 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.regionserver.hlog.reader.impl = [org.apache.hadoop.hbase.regionserver.wal.ProtobufLogReader]
2016-12-02 19:20:18,833 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.regionserver.hlog.writer.impl = [org.apache.hadoop.hbase.regionserver.wal.ProtobufLogWriter]
2016-12-02 19:20:18,833 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.regionserver.info.bindAddress = [0.0.0.0]
2016-12-02 19:20:18,833 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.regionserver.info.port = [16030]
2016-12-02 19:20:18,833 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.regionserver.info.port.auto = [false]
2016-12-02 19:20:18,833 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.regionserver.lease.period = [1200000]
2016-12-02 19:20:18,833 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.regionserver.logroll.errors.tolerated = [2]
2016-12-02 19:20:18,833 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.regionserver.logroll.period = [3600000]
2016-12-02 19:20:18,833 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.regionserver.maxlogs = [48]
2016-12-02 19:20:18,833 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.regionserver.msginterval = [3000]
2016-12-02 19:20:18,833 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.regionserver.optionalcacheflushinterval = [3600000]
2016-12-02 19:20:18,834 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.regionserver.port = [16020]
2016-12-02 19:20:18,834 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.regionserver.region.split.policy = [org.apache.hadoop.hbase.regionserver.IncreasingToUpperBoundRegionSplitPolicy]
2016-12-02 19:20:18,834 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.regionserver.regionSplitLimit = [1000]
2016-12-02 19:20:18,834 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.regionserver.storefile.refresh.period = [0]
2016-12-02 19:20:18,834 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.regionserver.thread.compaction.large = [1]
2016-12-02 19:20:18,834 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.regionserver.thread.compaction.small = [4]
2016-12-02 19:20:18,834 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.regionserver.thrift.compact = [false]
2016-12-02 19:20:18,834 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.regionserver.thrift.framed = [false]
2016-12-02 19:20:18,834 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.regionserver.thrift.framed.max_frame_size_in_mb = [2]
2016-12-02 19:20:18,834 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.regionserver.wal.codec = [org.apache.hadoop.hbase.regionserver.wal.WALCellCodec]
2016-12-02 19:20:18,834 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.regionserver.wal.enablecompression = [TRUE]
2016-12-02 19:20:18,834 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.replication.rpc.codec = [org.apache.hadoop.hbase.codec.KeyValueCodecWithTags]
2016-12-02 19:20:18,834 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.rest.filter.classes = [org.apache.hadoop.hbase.rest.filter.GzipFilter]
2016-12-02 19:20:18,834 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.rest.port = [8080]
2016-12-02 19:20:18,834 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.rest.readonly = [false]
2016-12-02 19:20:18,834 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.rest.support.proxyuser = [false]
2016-12-02 19:20:18,834 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.rest.threads.max = [100]
2016-12-02 19:20:18,834 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.rest.threads.min = [2]
2016-12-02 19:20:18,834 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.rootdir = [hdfs://hscale-dev1-nn:8020/apps/hbase/data]
2016-12-02 19:20:18,834 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.rootdir.perms = [700]
2016-12-02 19:20:18,834 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.rpc.protection = [authentication]
2016-12-02 19:20:18,834 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.rpc.shortoperation.timeout = [10000]
2016-12-02 19:20:18,834 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.rpc.timeout = [1200000]
2016-12-02 19:20:18,834 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.rs.cacheblocksonwrite = [false]
2016-12-02 19:20:18,834 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.security.authentication = [simple]
2016-12-02 19:20:18,834 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.security.authorization = [false]
2016-12-02 19:20:18,834 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.security.exec.permission.checks = [false]
2016-12-02 19:20:18,834 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.security.visibility.mutations.checkauths = [false]
2016-12-02 19:20:18,834 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.server.compactchecker.interval.multiplier = [1000]
2016-12-02 19:20:18,834 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.server.scanner.max.result.size = [104857600]
2016-12-02 19:20:18,834 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.server.thread.wakefrequency = [10000]
2016-12-02 19:20:18,834 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.server.versionfile.writeattempts = [3]
2016-12-02 19:20:18,834 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.snapshot.enabled = [true]
2016-12-02 19:20:18,834 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.snapshot.master.timeout.millis = [300000]
2016-12-02 19:20:18,834 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.snapshot.region.timeout = [300000]
2016-12-02 19:20:18,834 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.snapshot.restore.failsafe.name = [hbase-failsafe-{snapshot.name}-{restore.timestamp}]
2016-12-02 19:20:18,834 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.snapshot.restore.take.failsafe.snapshot = [true]
2016-12-02 19:20:18,834 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.splitlog.manager.timeout = [3000]
2016-12-02 19:20:18,834 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.status.listener.class = [org.apache.hadoop.hbase.client.ClusterStatusListener$MulticastListener]
2016-12-02 19:20:18,834 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.status.multicast.address.ip = [226.1.1.3]
2016-12-02 19:20:18,834 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.status.multicast.address.port = [16100]
2016-12-02 19:20:18,834 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.status.multicast.port = [16100]
2016-12-02 19:20:18,834 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.status.published = [false]
2016-12-02 19:20:18,834 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.status.publisher.class = [org.apache.hadoop.hbase.master.ClusterStatusPublisher$MulticastPublisher]
2016-12-02 19:20:18,834 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.storescanner.parallel.seek.enable = [false]
2016-12-02 19:20:18,834 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.storescanner.parallel.seek.threads = [10]
2016-12-02 19:20:18,834 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.superuser = [hbase]
2016-12-02 19:20:18,834 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.table.lock.enable = [true]
2016-12-02 19:20:18,835 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.table.max.rowsize = [1073741824]
2016-12-02 19:20:18,835 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.thrift.htablepool.size.max = [1000]
2016-12-02 19:20:18,835 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.thrift.maxQueuedRequests = [1000]
2016-12-02 19:20:18,835 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.thrift.maxWorkerThreads = [1000]
2016-12-02 19:20:18,835 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.thrift.minWorkerThreads = [16]
2016-12-02 19:20:18,835 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.tmp.dir = [/tmp/hbase-hbase]
2016-12-02 19:20:18,835 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.wal.disruptor.batch = [TRUE]
2016-12-02 19:20:18,835 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.wal.provider = [multiwal]
2016-12-02 19:20:18,835 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.wal.regiongrouping.numgroups = [16]
2016-12-02 19:20:18,835 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.zookeeper.dns.interface = [default]
2016-12-02 19:20:18,835 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.zookeeper.dns.nameserver = [default]
2016-12-02 19:20:18,835 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.zookeeper.leaderport = [3888]
2016-12-02 19:20:18,835 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.zookeeper.peerport = [2888]
2016-12-02 19:20:18,835 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.zookeeper.property.clientPort = [2181]
2016-12-02 19:20:18,835 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.zookeeper.property.dataDir = [/tmp/hbase-hbase/zookeeper]
2016-12-02 19:20:18,835 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.zookeeper.property.initLimit = [10]
2016-12-02 19:20:18,835 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.zookeeper.property.maxClientCnxns = [300]
2016-12-02 19:20:18,835 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.zookeeper.property.syncLimit = [5]
2016-12-02 19:20:18,835 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.zookeeper.property.tickTime = [6000]
2016-12-02 19:20:18,835 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.zookeeper.quorum = [hscale-dev1-dn1,hscale-dev1-dn3,hscale-dev1-dn2,hscale-dev1-dn4]
2016-12-02 19:20:18,835 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbase.zookeeper.useMulti = [true]
2016-12-02 19:20:18,835 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbaseSecurityAuthentication = [false]
2016-12-02 19:20:18,835 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hbaseSecurityAuthorization = [simple]
2016-12-02 19:20:18,835 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hfile.block.bloom.cacheonwrite = [TRUE]
2016-12-02 19:20:18,835 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hfile.block.cache.size = [0.25]
2016-12-02 19:20:18,835 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hfile.block.index.cacheonwrite = [false]
2016-12-02 19:20:18,835 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hfile.format.version = [3]
2016-12-02 19:20:18,835 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  hfile.index.block.max.size = [131072]
2016-12-02 19:20:18,835 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  ignoreSavePoints = [false]
2016-12-02 19:20:18,835 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  importMaxQuotedColumnLines = [50000]
2016-12-02 19:20:18,835 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  indexBatchSize = [4000]
2016-12-02 19:20:18,835 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  indexFetchSampleSize = [128]
2016-12-02 19:20:18,835 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  indexLookupBlocks = [5]
2016-12-02 19:20:18,835 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  io.compression.codec.bzip2.library = [system-native]
2016-12-02 19:20:18,835 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  io.compression.codecs = [org.apache.hadoop.io.compress.GzipCodec,org.apache.hadoop.io.compress.DefaultCodec,org.apache.hadoop.io.compress.SnappyCodec]
2016-12-02 19:20:18,835 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  io.file.buffer.size = [131072]
2016-12-02 19:20:18,835 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  io.map.index.interval = [128]
2016-12-02 19:20:18,835 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  io.map.index.skip = [0]
2016-12-02 19:20:18,835 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  io.mapfile.bloom.error.rate = [0.005]
2016-12-02 19:20:18,835 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  io.mapfile.bloom.size = [1048576]
2016-12-02 19:20:18,835 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  io.native.lib.available = [true]
2016-12-02 19:20:18,835 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  io.seqfile.compress.blocksize = [1000000]
2016-12-02 19:20:18,835 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  io.seqfile.lazydecompress = [true]
2016-12-02 19:20:18,835 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  io.seqfile.local.dir = [/tmp/hadoop-hbase/io/local]
2016-12-02 19:20:18,835 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  io.seqfile.sorter.recordlimit = [1000000]
2016-12-02 19:20:18,835 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  io.serializations = [org.apache.hadoop.io.serializer.WritableSerialization]
2016-12-02 19:20:18,835 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  io.skip.checksum.errors = [false]
2016-12-02 19:20:18,835 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  io.storefile.bloom.block.size = [131072]
2016-12-02 19:20:18,835 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  io.storefile.bloom.error.rate = [0.005]
2016-12-02 19:20:18,835 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  ipc.client.connect.max.retries = [50]
2016-12-02 19:20:18,835 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  ipc.client.connect.max.retries.on.timeouts = [45]
2016-12-02 19:20:18,835 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  ipc.client.connect.retry.interval = [1000]
2016-12-02 19:20:18,836 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  ipc.client.connect.timeout = [20000]
2016-12-02 19:20:18,836 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  ipc.client.connection.maxidletime = [30000]
2016-12-02 19:20:18,836 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  ipc.client.fallback-to-simple-auth-allowed = [false]
2016-12-02 19:20:18,836 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  ipc.client.idlethreshold = [8000]
2016-12-02 19:20:18,836 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  ipc.client.kill.max = [10]
2016-12-02 19:20:18,836 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  ipc.server.listen.queue.size = [128]
2016-12-02 19:20:18,836 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  ipc.server.log.slow.rpc = [false]
2016-12-02 19:20:18,836 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  ipc.server.max.connections = [0]
2016-12-02 19:20:18,836 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  ipc.server.tcpnodelay = [true]
2016-12-02 19:20:18,836 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  ipcThreads = [100]
2016-12-02 19:20:18,836 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  kryoPoolSize = [1100]
2016-12-02 19:20:18,836 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  map.sort.class = [org.apache.hadoop.util.QuickSort]
2016-12-02 19:20:18,836 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapred.child.java.opts = [-Xmx200m]
2016-12-02 19:20:18,836 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.admin.map.child.java.opts = [-server -XX:NewRatio=8 -Djava.net.preferIPv4Stack=true -Dhdp.version=2.4.2.0-258]
2016-12-02 19:20:18,836 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.admin.reduce.child.java.opts = [-server -XX:NewRatio=8 -Djava.net.preferIPv4Stack=true -Dhdp.version=2.4.2.0-258]
2016-12-02 19:20:18,836 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.admin.user.env = [LD_LIBRARY_PATH=/usr/hdp/2.4.2.0-258/hadoop/lib/native:/usr/hdp/2.4.2.0-258/hadoop/lib/native/Linux-amd64-64]
2016-12-02 19:20:18,836 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.am.max-attempts = [2]
2016-12-02 19:20:18,836 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.app-submission.cross-platform = [false]
2016-12-02 19:20:18,836 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.application.classpath = [$PWD/mr-framework/hadoop/share/hadoop/mapreduce/*:$PWD/mr-framework/hadoop/share/hadoop/mapreduce/lib/*:$PWD/mr-framework/hadoop/share/hadoop/common/*:$PWD/mr-framework/hadoop/share/hadoop/common/lib/*:$PWD/mr-framework/hadoop/share/hadoop/yarn/*:$PWD/mr-framework/hadoop/share/hadoop/yarn/lib/*:$PWD/mr-framework/hadoop/share/hadoop/hdfs/*:$PWD/mr-framework/hadoop/share/hadoop/hdfs/lib/*:$PWD/mr-framework/hadoop/share/hadoop/tools/lib/*:/usr/hdp/2.4.2.0-258/hadoop/lib/hadoop-lzo-0.6.0.2.4.2.0-258.jar:/etc/hadoop/conf/secure]
2016-12-02 19:20:18,836 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.application.framework.path = [/hdp/apps/2.4.2.0-258/mapreduce/mapreduce.tar.gz#mr-framework]
2016-12-02 19:20:18,836 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.client.completion.pollinterval = [5000]
2016-12-02 19:20:18,836 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.client.output.filter = [FAILED]
2016-12-02 19:20:18,836 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.client.progressmonitor.pollinterval = [1000]
2016-12-02 19:20:18,836 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.client.submit.file.replication = [10]
2016-12-02 19:20:18,836 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.cluster.acls.enabled = [false]
2016-12-02 19:20:18,836 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.cluster.administrators = [ hadoop]
2016-12-02 19:20:18,836 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.cluster.local.dir = [/tmp/hadoop-hbase/mapred/local]
2016-12-02 19:20:18,836 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.cluster.temp.dir = [/tmp/hadoop-hbase/mapred/temp]
2016-12-02 19:20:18,836 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.fileoutputcommitter.algorithm.version = [1]
2016-12-02 19:20:18,836 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.framework.name = [yarn]
2016-12-02 19:20:18,836 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.ifile.readahead = [true]
2016-12-02 19:20:18,836 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.ifile.readahead.bytes = [4194304]
2016-12-02 19:20:18,836 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.input.fileinputformat.list-status.num-threads = [1]
2016-12-02 19:20:18,836 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.input.fileinputformat.split.minsize = [0]
2016-12-02 19:20:18,836 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.input.lineinputformat.linespermap = [1]
2016-12-02 19:20:18,836 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.job.acl-modify-job = [ ]
2016-12-02 19:20:18,836 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.job.acl-view-job = [ ]
2016-12-02 19:20:18,836 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.job.classloader = [false]
2016-12-02 19:20:18,836 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.job.committer.setup.cleanup.needed = [true]
2016-12-02 19:20:18,836 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.job.complete.cancel.delegation.tokens = [true]
2016-12-02 19:20:18,836 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.job.counters.max = [130]
2016-12-02 19:20:18,836 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.job.emit-timeline-data = [false]
2016-12-02 19:20:18,836 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.job.end-notification.max.attempts = [5]
2016-12-02 19:20:18,836 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.job.end-notification.max.retry.interval = [5000]
2016-12-02 19:20:18,836 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.job.end-notification.retry.attempts = [0]
2016-12-02 19:20:18,836 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.job.end-notification.retry.interval = [1000]
2016-12-02 19:20:18,836 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.job.hdfs-servers = [hdfs://hscale-dev1-nn:8020]
2016-12-02 19:20:18,836 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.job.jvm.numtasks = [1]
2016-12-02 19:20:18,836 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.job.map.output.collector.class = [org.apache.hadoop.mapred.MapTask$MapOutputBuffer]
2016-12-02 19:20:18,836 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.job.maps = [2]
2016-12-02 19:20:18,836 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.job.max.split.locations = [10]
2016-12-02 19:20:18,837 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.job.maxtaskfailures.per.tracker = [3]
2016-12-02 19:20:18,837 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.job.queuename = [default]
2016-12-02 19:20:18,837 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.job.reduce.shuffle.consumer.plugin.class = [org.apache.hadoop.mapreduce.task.reduce.Shuffle]
2016-12-02 19:20:18,837 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.job.reduce.slowstart.completedmaps = [0.05]
2016-12-02 19:20:18,837 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.job.reducer.preempt.delay.sec = [0]
2016-12-02 19:20:18,837 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.job.reducer.unconditional-preempt.delay.sec = [300]
2016-12-02 19:20:18,837 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.job.reduces = [1]
2016-12-02 19:20:18,837 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.job.running.map.limit = [0]
2016-12-02 19:20:18,837 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.job.running.reduce.limit = [0]
2016-12-02 19:20:18,837 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.job.speculative.minimum-allowed-tasks = [10]
2016-12-02 19:20:18,837 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.job.speculative.retry-after-no-speculate = [1000]
2016-12-02 19:20:18,837 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.job.speculative.retry-after-speculate = [15000]
2016-12-02 19:20:18,837 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.job.speculative.slowtaskthreshold = [1.0]
2016-12-02 19:20:18,837 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.job.speculative.speculative-cap-running-tasks = [0.1]
2016-12-02 19:20:18,837 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.job.speculative.speculative-cap-total-tasks = [0.01]
2016-12-02 19:20:18,837 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.job.split.metainfo.maxsize = [10000000]
2016-12-02 19:20:18,837 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.job.token.tracking.ids.enabled = [false]
2016-12-02 19:20:18,837 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.job.ubertask.enable = [false]
2016-12-02 19:20:18,837 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.job.ubertask.maxmaps = [9]
2016-12-02 19:20:18,837 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.job.ubertask.maxreduces = [1]
2016-12-02 19:20:18,837 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.job.userlog.retain.hours = [24]
2016-12-02 19:20:18,837 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.jobhistory.address = [hscale-dev1-nn:10020]
2016-12-02 19:20:18,837 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.jobhistory.admin.acl = [*]
2016-12-02 19:20:18,837 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.jobhistory.admin.address = [0.0.0.0:10033]
2016-12-02 19:20:18,837 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.jobhistory.bind-host = [0.0.0.0]
2016-12-02 19:20:18,837 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.jobhistory.cleaner.enable = [true]
2016-12-02 19:20:18,837 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.jobhistory.cleaner.interval-ms = [86400000]
2016-12-02 19:20:18,837 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.jobhistory.client.thread-count = [10]
2016-12-02 19:20:18,837 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.jobhistory.datestring.cache.size = [200000]
2016-12-02 19:20:18,837 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.jobhistory.done-dir = [/mr-history/done]
2016-12-02 19:20:18,837 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.jobhistory.http.policy = [HTTP_ONLY]
2016-12-02 19:20:18,837 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.jobhistory.intermediate-done-dir = [/mr-history/tmp]
2016-12-02 19:20:18,837 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.jobhistory.joblist.cache.size = [20000]
2016-12-02 19:20:18,837 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.jobhistory.keytab = [/etc/security/keytabs/jhs.service.keytab]
2016-12-02 19:20:18,837 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.jobhistory.loadedjobs.cache.size = [5]
2016-12-02 19:20:18,837 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.jobhistory.max-age-ms = [604800000]
2016-12-02 19:20:18,837 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.jobhistory.minicluster.fixed.ports = [false]
2016-12-02 19:20:18,837 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.jobhistory.move.interval-ms = [180000]
2016-12-02 19:20:18,837 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.jobhistory.move.thread-count = [3]
2016-12-02 19:20:18,837 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.jobhistory.principal = [jhs/_HOST@HSCALE.COM]
2016-12-02 19:20:18,837 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.jobhistory.recovery.enable = [true]
2016-12-02 19:20:18,837 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.jobhistory.recovery.store.class = [org.apache.hadoop.mapreduce.v2.hs.HistoryServerLeveldbStateStoreService]
2016-12-02 19:20:18,837 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.jobhistory.recovery.store.fs.uri = [/tmp/hadoop-hbase/mapred/history/recoverystore]
2016-12-02 19:20:18,837 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.jobhistory.recovery.store.leveldb.path = [/hadoop/mapreduce/jhs]
2016-12-02 19:20:18,837 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.jobhistory.webapp.address = [hscale-dev1-nn:19888]
2016-12-02 19:20:18,837 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.jobhistory.webapp.rest-csrf.custom-header = [X-XSRF-Header]
2016-12-02 19:20:18,837 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.jobhistory.webapp.rest-csrf.enabled = [false]
2016-12-02 19:20:18,837 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.jobhistory.webapp.rest-csrf.methods-to-ignore = [GET,OPTIONS,HEAD]
2016-12-02 19:20:18,837 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.jobhistory.webapp.spnego-keytab-file = [/etc/security/keytabs/spnego.service.keytab]
2016-12-02 19:20:18,837 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.jobhistory.webapp.spnego-principal = [HTTP/_HOST@HSCALE.COM]
2016-12-02 19:20:18,837 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.jobtracker.address = [local]
2016-12-02 19:20:18,837 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.jobtracker.expire.trackers.interval = [600000]
2016-12-02 19:20:18,837 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.jobtracker.handler.count = [10]
2016-12-02 19:20:18,837 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.jobtracker.heartbeats.in.second = [100]
2016-12-02 19:20:18,837 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.jobtracker.http.address = [0.0.0.0:50030]
2016-12-02 19:20:18,837 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.jobtracker.instrumentation = [org.apache.hadoop.mapred.JobTrackerMetricsInst]
2016-12-02 19:20:18,837 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.jobtracker.jobhistory.block.size = [3145728]
2016-12-02 19:20:18,837 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.jobtracker.jobhistory.lru.cache.size = [5]
2016-12-02 19:20:18,838 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.jobtracker.jobhistory.task.numberprogresssplits = [12]
2016-12-02 19:20:18,838 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.jobtracker.maxtasks.perjob = [-1]
2016-12-02 19:20:18,838 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.jobtracker.persist.jobstatus.active = [true]
2016-12-02 19:20:18,838 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.jobtracker.persist.jobstatus.dir = [/jobtracker/jobsInfo]
2016-12-02 19:20:18,838 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.jobtracker.persist.jobstatus.hours = [1]
2016-12-02 19:20:18,838 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.jobtracker.restart.recover = [false]
2016-12-02 19:20:18,838 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.jobtracker.retiredjobs.cache.size = [1000]
2016-12-02 19:20:18,838 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.jobtracker.staging.root.dir = [/tmp/hadoop-hbase/mapred/staging]
2016-12-02 19:20:18,838 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.jobtracker.system.dir = [/tmp/hadoop-hbase/mapred/system]
2016-12-02 19:20:18,838 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.jobtracker.taskcache.levels = [2]
2016-12-02 19:20:18,838 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.jobtracker.taskscheduler = [org.apache.hadoop.mapred.JobQueueTaskScheduler]
2016-12-02 19:20:18,838 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.jobtracker.tasktracker.maxblacklists = [4]
2016-12-02 19:20:18,838 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.jobtracker.webinterface.trusted = [false]
2016-12-02 19:20:18,838 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.local.clientfactory.class.name = [org.apache.hadoop.mapred.LocalClientFactory]
2016-12-02 19:20:18,838 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.map.cpu.vcores = [1]
2016-12-02 19:20:18,838 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.map.java.opts = [-Xmx1228m]
2016-12-02 19:20:18,838 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.map.log.level = [INFO]
2016-12-02 19:20:18,838 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.map.maxattempts = [4]
2016-12-02 19:20:18,838 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.map.memory.mb = [1536]
2016-12-02 19:20:18,838 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.map.output.compress = [false]
2016-12-02 19:20:18,838 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.map.output.compress.codec = [org.apache.hadoop.io.compress.DefaultCodec]
2016-12-02 19:20:18,838 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.map.skip.maxrecords = [0]
2016-12-02 19:20:18,838 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.map.skip.proc.count.autoincr = [true]
2016-12-02 19:20:18,838 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.map.sort.spill.percent = [0.7]
2016-12-02 19:20:18,838 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.map.speculative = [false]
2016-12-02 19:20:18,838 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.output.fileoutputformat.compress = [false]
2016-12-02 19:20:18,838 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.output.fileoutputformat.compress.codec = [org.apache.hadoop.io.compress.DefaultCodec]
2016-12-02 19:20:18,838 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.output.fileoutputformat.compress.type = [BLOCK]
2016-12-02 19:20:18,838 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.reduce.cpu.vcores = [1]
2016-12-02 19:20:18,838 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.reduce.input.buffer.percent = [0.0]
2016-12-02 19:20:18,838 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.reduce.java.opts = [-Xmx1228m]
2016-12-02 19:20:18,838 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.reduce.log.level = [INFO]
2016-12-02 19:20:18,838 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.reduce.markreset.buffer.percent = [0.0]
2016-12-02 19:20:18,838 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.reduce.maxattempts = [4]
2016-12-02 19:20:18,838 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.reduce.memory.mb = [1536]
2016-12-02 19:20:18,838 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.reduce.merge.inmem.threshold = [1000]
2016-12-02 19:20:18,838 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.reduce.shuffle.connect.timeout = [180000]
2016-12-02 19:20:18,838 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.reduce.shuffle.fetch.retry.enabled = [1]
2016-12-02 19:20:18,838 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.reduce.shuffle.fetch.retry.interval-ms = [1000]
2016-12-02 19:20:18,838 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.reduce.shuffle.fetch.retry.timeout-ms = [30000]
2016-12-02 19:20:18,838 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.reduce.shuffle.input.buffer.percent = [0.7]
2016-12-02 19:20:18,838 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.reduce.shuffle.memory.limit.percent = [0.25]
2016-12-02 19:20:18,838 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.reduce.shuffle.merge.percent = [0.66]
2016-12-02 19:20:18,838 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.reduce.shuffle.parallelcopies = [30]
2016-12-02 19:20:18,838 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.reduce.shuffle.read.timeout = [180000]
2016-12-02 19:20:18,838 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.reduce.shuffle.retry-delay.max.ms = [60000]
2016-12-02 19:20:18,838 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.reduce.skip.maxgroups = [0]
2016-12-02 19:20:18,838 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.reduce.skip.proc.count.autoincr = [true]
2016-12-02 19:20:18,838 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.reduce.speculative = [false]
2016-12-02 19:20:18,838 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.shuffle.connection-keep-alive.enable = [false]
2016-12-02 19:20:18,838 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.shuffle.connection-keep-alive.timeout = [5]
2016-12-02 19:20:18,839 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.shuffle.max.connections = [0]
2016-12-02 19:20:18,839 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.shuffle.max.threads = [0]
2016-12-02 19:20:18,839 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.shuffle.port = [13562]
2016-12-02 19:20:18,839 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.shuffle.ssl.enabled = [false]
2016-12-02 19:20:18,839 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.shuffle.ssl.file.buffer.size = [65536]
2016-12-02 19:20:18,839 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.shuffle.transfer.buffer.size = [131072]
2016-12-02 19:20:18,839 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.task.combine.progress.records = [10000]
2016-12-02 19:20:18,839 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.task.files.preserve.failedtasks = [false]
2016-12-02 19:20:18,839 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.task.io.sort.factor = [100]
2016-12-02 19:20:18,839 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.task.io.sort.mb = [859]
2016-12-02 19:20:18,839 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.task.merge.progress.records = [10000]
2016-12-02 19:20:18,839 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.task.profile = [false]
2016-12-02 19:20:18,839 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.task.profile.map.params = [-agentlib:hprof=cpu=samples,heap=sites,force=n,thread=y,verbose=n,file=%s]
2016-12-02 19:20:18,839 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.task.profile.maps = [0-2]
2016-12-02 19:20:18,839 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.task.profile.params = [-agentlib:hprof=cpu=samples,heap=sites,force=n,thread=y,verbose=n,file=%s]
2016-12-02 19:20:18,839 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.task.profile.reduce.params = [-agentlib:hprof=cpu=samples,heap=sites,force=n,thread=y,verbose=n,file=%s]
2016-12-02 19:20:18,839 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.task.profile.reduces = [0-2]
2016-12-02 19:20:18,839 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.task.skip.start.attempts = [2]
2016-12-02 19:20:18,839 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.task.timeout = [300000]
2016-12-02 19:20:18,839 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.task.userlog.limit.kb = [0]
2016-12-02 19:20:18,839 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.tasktracker.dns.interface = [default]
2016-12-02 19:20:18,839 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.tasktracker.dns.nameserver = [default]
2016-12-02 19:20:18,839 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.tasktracker.healthchecker.interval = [60000]
2016-12-02 19:20:18,839 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.tasktracker.healthchecker.script.timeout = [600000]
2016-12-02 19:20:18,839 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.tasktracker.http.address = [0.0.0.0:50060]
2016-12-02 19:20:18,839 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.tasktracker.http.threads = [40]
2016-12-02 19:20:18,839 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.tasktracker.indexcache.mb = [10]
2016-12-02 19:20:18,839 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.tasktracker.instrumentation = [org.apache.hadoop.mapred.TaskTrackerMetricsInst]
2016-12-02 19:20:18,839 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.tasktracker.local.dir.minspacekill = [0]
2016-12-02 19:20:18,839 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.tasktracker.local.dir.minspacestart = [0]
2016-12-02 19:20:18,839 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.tasktracker.map.tasks.maximum = [2]
2016-12-02 19:20:18,839 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.tasktracker.outofband.heartbeat = [false]
2016-12-02 19:20:18,839 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.tasktracker.reduce.tasks.maximum = [2]
2016-12-02 19:20:18,839 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.tasktracker.report.address = [127.0.0.1:0]
2016-12-02 19:20:18,839 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.tasktracker.taskcontroller = [org.apache.hadoop.mapred.DefaultTaskController]
2016-12-02 19:20:18,839 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.tasktracker.taskmemorymanager.monitoringinterval = [5000]
2016-12-02 19:20:18,839 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  mapreduce.tasktracker.tasks.sleeptimebeforesigkill = [5000]
2016-12-02 19:20:18,839 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  maxBufferEntries = [1000]
2016-12-02 19:20:18,839 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  maxBufferHeapSize = [3145728]
2016-12-02 19:20:18,839 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  maxDdlWait = [60000]
2016-12-02 19:20:18,839 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  maxDependentWrites = [60000]
2016-12-02 19:20:18,839 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  maxIndependentWrites = [60000]
2016-12-02 19:20:18,839 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  maxRetries = [5]
2016-12-02 19:20:18,839 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  maxWriterThreads = [5]
2016-12-02 19:20:18,839 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  namespace = [splice]
2016-12-02 19:20:18,839 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  nestedLoopJoinBatchSize = [10]
2016-12-02 19:20:18,839 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  net.topology.impl = [org.apache.hadoop.net.NetworkTopology]
2016-12-02 19:20:18,839 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  net.topology.node.switch.mapping.impl = [org.apache.hadoop.net.ScriptBasedMapping]
2016-12-02 19:20:18,839 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  net.topology.script.file.name = [/etc/hadoop/conf/topology_script.py]
2016-12-02 19:20:18,839 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  net.topology.script.number.args = [100]
2016-12-02 19:20:18,839 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  networkBindAddress = [0.0.0.0]
2016-12-02 19:20:18,839 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  networkBindPort = [1527]
2016-12-02 19:20:18,839 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  nfs.allow.insecure.ports = [true]
2016-12-02 19:20:18,839 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  nfs.dump.dir = [/tmp/.hdfs-nfs]
2016-12-02 19:20:18,839 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  nfs.exports.allowed.hosts = [* rw]
2016-12-02 19:20:18,839 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  nfs.file.dump.dir = [/tmp/.hdfs-nfs]
2016-12-02 19:20:18,839 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  nfs.mountd.port = [4242]
2016-12-02 19:20:18,840 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  nfs.rtmax = [1048576]
2016-12-02 19:20:18,840 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  nfs.server.port = [2049]
2016-12-02 19:20:18,840 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  nfs.wtmax = [1048576]
2016-12-02 19:20:18,840 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  olapClientTickTime = [1000]
2016-12-02 19:20:18,840 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  olapClientWaitTime = [900000]
2016-12-02 19:20:18,840 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  olapServerBindPort = [60014]
2016-12-02 19:20:18,840 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  olapServerThreads = [16]
2016-12-02 19:20:18,840 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  olapServerTickLimit = [120]
2016-12-02 19:20:18,840 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  optimizerExtraQualifierMultiplier = [0.9]
2016-12-02 19:20:18,840 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  optimizerPlanMaximumTimeout = [9223372036854775807]
2016-12-02 19:20:18,840 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  optimizerPlanMinimumTimeout = [0]
2016-12-02 19:20:18,840 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  partitionCacheExpiration = [60000]
2016-12-02 19:20:18,840 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  partitionserverJmxPort = [10102]
2016-12-02 19:20:18,840 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  partitionserverPort = [16020]
2016-12-02 19:20:18,840 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  phoenix.connection.autoCommit = [true]
2016-12-02 19:20:18,840 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  phoenix.functions.allowUserDefinedFunctions = [ ]
2016-12-02 19:20:18,840 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  phoenix.query.timeoutMs = [60000]
2016-12-02 19:20:18,840 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  pipelineKryoPoolSize = [1024]
2016-12-02 19:20:18,840 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  readResolverQueueSize = [-1]
2016-12-02 19:20:18,840 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  readResolverThreads = [4]
2016-12-02 19:20:18,840 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  regionLoadUpdateInterval = [5]
2016-12-02 19:20:18,840 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  regionMaxFileSize = [10737418240]
2016-12-02 19:20:18,840 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  regionServerHandlerCount = [100]
2016-12-02 19:20:18,840 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  reservedSlotsTimeout = [60]
2016-12-02 19:20:18,840 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  rpc.metrics.quantile.enable = [false]
2016-12-02 19:20:18,840 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  s3.blocksize = [67108864]
2016-12-02 19:20:18,840 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  s3.bytes-per-checksum = [512]
2016-12-02 19:20:18,840 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  s3.client-write-packet-size = [65536]
2016-12-02 19:20:18,840 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  s3.replication = [3]
2016-12-02 19:20:18,840 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  s3.stream-buffer-size = [4096]
2016-12-02 19:20:18,840 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  s3native.blocksize = [67108864]
2016-12-02 19:20:18,840 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  s3native.bytes-per-checksum = [512]
2016-12-02 19:20:18,840 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  s3native.client-write-packet-size = [65536]
2016-12-02 19:20:18,840 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  s3native.replication = [3]
2016-12-02 19:20:18,840 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  s3native.stream-buffer-size = [4096]
2016-12-02 19:20:18,840 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  sequenceBlockSize = [1000]
2016-12-02 19:20:18,840 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  sparkIoCompressionCodec = [lz4]
2016-12-02 19:20:18,840 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  sparkResultStreamingBatchSize = [1024]
2016-12-02 19:20:18,840 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  sparkResultStreamingBatches = [10]
2016-12-02 19:20:18,840 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  splice.authentication = [NATIVE]
2016-12-02 19:20:18,840 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  splice.authentication.native.algorithm = [SHA-512]
2016-12-02 19:20:18,840 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  splice.authentication.native.create.credentials.database = [true]
2016-12-02 19:20:18,840 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  splice.client.numConnections = [1]
2016-12-02 19:20:18,840 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  splice.client.write.maxDependentWrites = [60000]
2016-12-02 19:20:18,840 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  splice.client.write.maxIndependentWrites = [60000]
2016-12-02 19:20:18,840 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  splice.compression = [snappy]
2016-12-02 19:20:18,840 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  splice.marshal.kryoPoolSize = [1100]
2016-12-02 19:20:18,840 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  splice.olap_server.clientWaitTime = [900000]
2016-12-02 19:20:18,840 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  splice.ring.bufferSize = [131072]
2016-12-02 19:20:18,840 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  splice.splitBlockSize = [67108864]
2016-12-02 19:20:18,840 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  splice.timestamp_server.clientWaitTime = [120000]
2016-12-02 19:20:18,840 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  splice.txn.activeTxns.cacheSize = [10240]
2016-12-02 19:20:18,840 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  splice.txn.completedTxns.concurrency = [128]
2016-12-02 19:20:18,840 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  splice.txn.concurrencyLevel = [4096]
2016-12-02 19:20:18,840 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  spliceRootPath = [/splice]
2016-12-02 19:20:18,840 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  splitBlockSize = [67108864]
2016-12-02 19:20:18,840 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  startupLockWaitPeriod = [1000]
2016-12-02 19:20:18,840 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  storageFactoryHome = [hdfs://hscale-dev1-nn:8020/apps/hbase/data]
2016-12-02 19:20:18,840 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  tableSplitSleepInterval = [500]
2016-12-02 19:20:18,840 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  tfile.fs.input.buffer.size = [262144]
2016-12-02 19:20:18,840 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  tfile.fs.output.buffer.size = [262144]
2016-12-02 19:20:18,840 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  tfile.io.chunk.size = [1048576]
2016-12-02 19:20:18,840 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  threadKeepaliveTime = [60]
2016-12-02 19:20:18,841 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  timestampBlockSize = [8192]
2016-12-02 19:20:18,841 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  timestampClientWaitTime = [120000]
2016-12-02 19:20:18,841 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  timestampServerBindPort = [60012]
2016-12-02 19:20:18,841 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  topkSize = [10]
2016-12-02 19:20:18,841 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  transactionKeepAliveInterval = [15000]
2016-12-02 19:20:18,841 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  transactionKeepAliveThreads = [4]
2016-12-02 19:20:18,841 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  transactionLockStripes = [100]
2016-12-02 19:20:18,841 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  transactionTimeout = [150000]
2016-12-02 19:20:18,841 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  upgradeForced = [false]
2016-12-02 19:20:18,841 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  upgradeForcedFrom = [null]
2016-12-02 19:20:18,841 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  writeMaxFlushesPerRegion = [5]
2016-12-02 19:20:18,841 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.acl.enable = [false]
2016-12-02 19:20:18,841 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.admin.acl = [yarn]
2016-12-02 19:20:18,841 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.am.liveness-monitor.expiry-interval-ms = [600000]
2016-12-02 19:20:18,841 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.app.mapreduce.am.admin-command-opts = [-Dhdp.version=2.4.2.0-258]
2016-12-02 19:20:18,841 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.app.mapreduce.am.command-opts = [-Xmx410m]
2016-12-02 19:20:18,841 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.app.mapreduce.am.container.log.backups = [0]
2016-12-02 19:20:18,841 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.app.mapreduce.am.container.log.limit.kb = [0]
2016-12-02 19:20:18,841 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.app.mapreduce.am.containerlauncher.threadpool-initial-size = [10]
2016-12-02 19:20:18,841 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.app.mapreduce.am.hard-kill-timeout-ms = [10000]
2016-12-02 19:20:18,841 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.app.mapreduce.am.job.committer.cancel-timeout = [60000]
2016-12-02 19:20:18,841 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.app.mapreduce.am.job.committer.commit-window = [10000]
2016-12-02 19:20:18,841 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.app.mapreduce.am.job.task.listener.thread-count = [30]
2016-12-02 19:20:18,841 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.app.mapreduce.am.log.level = [INFO]
2016-12-02 19:20:18,841 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.app.mapreduce.am.resource.cpu-vcores = [1]
2016-12-02 19:20:18,841 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.app.mapreduce.am.resource.mb = [512]
2016-12-02 19:20:18,841 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.app.mapreduce.am.scheduler.heartbeat.interval-ms = [1000]
2016-12-02 19:20:18,841 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.app.mapreduce.am.staging-dir = [/user]
2016-12-02 19:20:18,841 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.app.mapreduce.client-am.ipc.max-retries = [3]
2016-12-02 19:20:18,841 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.app.mapreduce.client-am.ipc.max-retries-on-timeouts = [3]
2016-12-02 19:20:18,841 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.app.mapreduce.client.job.max-retries = [0]
2016-12-02 19:20:18,841 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.app.mapreduce.client.job.retry-interval = [2000]
2016-12-02 19:20:18,841 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.app.mapreduce.client.max-retries = [3]
2016-12-02 19:20:18,841 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.app.mapreduce.shuffle.log.backups = [0]
2016-12-02 19:20:18,841 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.app.mapreduce.shuffle.log.limit.kb = [0]
2016-12-02 19:20:18,841 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.app.mapreduce.shuffle.log.separate = [true]
2016-12-02 19:20:18,841 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.app.mapreduce.task.container.log.backups = [0]
2016-12-02 19:20:18,841 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.application.classpath = [$HADOOP_CONF_DIR, /usr/hdp/current/hadoop-client/*, /usr/hdp/current/hadoop-client/lib/*, /usr/hdp/current/hadoop-hdfs-client/*, /usr/hdp/current/hadoop-hdfs-client/lib/*, /usr/hdp/current/hadoop-yarn-client/*, /usr/hdp/current/hadoop-yarn-client/lib/*, /usr/hdp/current/hadoop-mapreduce-client/*, /usr/hdp/current/hadoop-mapreduce-client/lib/*, /usr/hdp/current/hbase-regionserver/*, /usr/hdp/current/hbase-regionserver/lib/*, /opt/splice/default/lib/*]
2016-12-02 19:20:18,841 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.client.application-client-protocol.poll-interval-ms = [200]
2016-12-02 19:20:18,841 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.client.failover-proxy-provider = [org.apache.hadoop.yarn.client.ConfiguredRMFailoverProxyProvider]
2016-12-02 19:20:18,841 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.client.failover-retries = [0]
2016-12-02 19:20:18,841 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.client.failover-retries-on-socket-timeouts = [0]
2016-12-02 19:20:18,841 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.client.max-cached-nodemanagers-proxies = [0]
2016-12-02 19:20:18,841 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.client.nodemanager-client-async.thread-pool-max-size = [500]
2016-12-02 19:20:18,841 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.client.nodemanager-connect.max-wait-ms = [60000]
2016-12-02 19:20:18,841 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.client.nodemanager-connect.retry-interval-ms = [10000]
2016-12-02 19:20:18,841 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.dispatcher.drain-events.timeout = [300000]
2016-12-02 19:20:18,841 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.fail-fast = [false]
2016-12-02 19:20:18,841 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.http.policy = [HTTP_ONLY]
2016-12-02 19:20:18,841 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.ipc.rpc.class = [org.apache.hadoop.yarn.ipc.HadoopYarnProtoRPC]
2016-12-02 19:20:18,841 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.log-aggregation-enable = [true]
2016-12-02 19:20:18,841 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.log-aggregation.retain-check-interval-seconds = [-1]
2016-12-02 19:20:18,841 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.log-aggregation.retain-seconds = [2592000]
2016-12-02 19:20:18,841 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.log.server.url = [http://hscale-dev1-nn:19888/jobhistory/logs]
2016-12-02 19:20:18,841 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.nm.liveness-monitor.expiry-interval-ms = [600000]
2016-12-02 19:20:18,841 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.node-labels.enabled = [false]
2016-12-02 19:20:18,841 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.node-labels.fs-store.impl.class = [org.apache.hadoop.yarn.nodelabels.FileSystemNodeLabelsStore]
2016-12-02 19:20:18,841 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.node-labels.fs-store.retry-policy-spec = [2000, 500]
2016-12-02 19:20:18,841 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.node-labels.fs-store.root-dir = [/system/yarn/node-labels]
2016-12-02 19:20:18,841 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.nodemanager.address = [0.0.0.0:45454]
2016-12-02 19:20:18,841 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.nodemanager.admin-env = [MALLOC_ARENA_MAX=$MALLOC_ARENA_MAX]
2016-12-02 19:20:18,841 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.nodemanager.aux-services = [mapreduce_shuffle,spark_shuffle]
2016-12-02 19:20:18,841 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.nodemanager.aux-services.mapreduce_shuffle.class = [org.apache.hadoop.mapred.ShuffleHandler]
2016-12-02 19:20:18,841 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.nodemanager.aux-services.spark_shuffle.class = [org.apache.spark.network.yarn.YarnShuffleService]
2016-12-02 19:20:18,841 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.nodemanager.bind-host = [0.0.0.0]
2016-12-02 19:20:18,841 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.nodemanager.container-executor.class = [org.apache.hadoop.yarn.server.nodemanager.DefaultContainerExecutor]
2016-12-02 19:20:18,841 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.nodemanager.container-manager.thread-count = [20]
2016-12-02 19:20:18,841 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.nodemanager.container-metrics.unregister-delay-ms = [10000]
2016-12-02 19:20:18,842 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.nodemanager.container-monitor.interval-ms = [3000]
2016-12-02 19:20:18,842 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.nodemanager.container-monitor.procfs-tree.smaps-based-rss.enabled = [false]
2016-12-02 19:20:18,842 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.nodemanager.delete.debug-delay-sec = [86400]
2016-12-02 19:20:18,842 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.nodemanager.delete.thread-count = [4]
2016-12-02 19:20:18,842 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.nodemanager.disk-health-checker.interval-ms = [120000]
2016-12-02 19:20:18,842 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.nodemanager.disk-health-checker.max-disk-utilization-per-disk-percentage = [90]
2016-12-02 19:20:18,842 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.nodemanager.disk-health-checker.min-free-space-per-disk-mb = [1000]
2016-12-02 19:20:18,842 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.nodemanager.disk-health-checker.min-healthy-disks = [0.25]
2016-12-02 19:20:18,842 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.nodemanager.docker-container-executor.exec-name = [/usr/bin/docker]
2016-12-02 19:20:18,842 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.nodemanager.env-whitelist = [JAVA_HOME,HADOOP_COMMON_HOME,HADOOP_HDFS_HOME,HADOOP_CONF_DIR,HADOOP_YARN_HOME]
2016-12-02 19:20:18,842 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.nodemanager.health-checker.interval-ms = [135000]
2016-12-02 19:20:18,842 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.nodemanager.health-checker.script.timeout-ms = [60000]
2016-12-02 19:20:18,842 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.nodemanager.hostname = [0.0.0.0]
2016-12-02 19:20:18,842 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.nodemanager.keytab = [/etc/krb5.keytab]
2016-12-02 19:20:18,842 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.nodemanager.linux-container-executor.cgroups.hierarchy = [hadoop-yarn]
2016-12-02 19:20:18,842 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.nodemanager.linux-container-executor.cgroups.mount = [false]
2016-12-02 19:20:18,842 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.nodemanager.linux-container-executor.cgroups.strict-resource-usage = [false]
2016-12-02 19:20:18,842 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.nodemanager.linux-container-executor.group = [hadoop]
2016-12-02 19:20:18,842 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.nodemanager.linux-container-executor.nonsecure-mode.limit-users = [true]
2016-12-02 19:20:18,842 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.nodemanager.linux-container-executor.nonsecure-mode.local-user = [nobody]
2016-12-02 19:20:18,842 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.nodemanager.linux-container-executor.nonsecure-mode.user-pattern = [^[_.A-Za-z0-9][-@_.A-Za-z0-9]{0,255}?[$]?$]
2016-12-02 19:20:18,842 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.nodemanager.linux-container-executor.resources-handler.class = [org.apache.hadoop.yarn.server.nodemanager.util.DefaultLCEResourcesHandler]
2016-12-02 19:20:18,842 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.nodemanager.local-cache.max-files-per-directory = [8192]
2016-12-02 19:20:18,842 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.nodemanager.local-dirs = [/hadoop/yarn/local]
2016-12-02 19:20:18,842 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.nodemanager.localizer.address = [0.0.0.0:8040]
2016-12-02 19:20:18,842 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.nodemanager.localizer.cache.cleanup.interval-ms = [600000]
2016-12-02 19:20:18,842 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.nodemanager.localizer.cache.target-size-mb = [10240]
2016-12-02 19:20:18,842 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.nodemanager.localizer.client.thread-count = [5]
2016-12-02 19:20:18,842 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.nodemanager.localizer.fetch.thread-count = [4]
2016-12-02 19:20:18,842 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.nodemanager.log-aggregation.compression-type = [gz]
2016-12-02 19:20:18,842 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.nodemanager.log-aggregation.debug-enabled = [false]
2016-12-02 19:20:18,842 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.nodemanager.log-aggregation.num-log-files-per-app = [30]
2016-12-02 19:20:18,842 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.nodemanager.log-aggregation.roll-monitoring-interval-seconds = [-1]
2016-12-02 19:20:18,842 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.nodemanager.log-container-debug-info.enabled = [true]
2016-12-02 19:20:18,842 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.nodemanager.log-dirs = [/hadoop/yarn/log]
2016-12-02 19:20:18,842 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.nodemanager.log.retain-second = [604800]
2016-12-02 19:20:18,842 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.nodemanager.log.retain-seconds = [10800]
2016-12-02 19:20:18,842 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.nodemanager.pmem-check-enabled = [true]
2016-12-02 19:20:18,842 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.nodemanager.process-kill-wait.ms = [2000]
2016-12-02 19:20:18,842 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.nodemanager.recovery.dir = [/var/log/hadoop-yarn/nodemanager/recovery-state]
2016-12-02 19:20:18,842 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.nodemanager.recovery.enabled = [true]
2016-12-02 19:20:18,842 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.nodemanager.remote-app-log-dir = [/app-logs]
2016-12-02 19:20:18,842 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.nodemanager.remote-app-log-dir-suffix = [logs]
2016-12-02 19:20:18,842 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.nodemanager.resource.cpu-vcores = [8]
2016-12-02 19:20:18,842 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.nodemanager.resource.memory-mb = [5120]
2016-12-02 19:20:18,842 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.nodemanager.resource.percentage-physical-cpu-limit = [80]
2016-12-02 19:20:18,842 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.nodemanager.resourcemanager.minimum.version = [NONE]
2016-12-02 19:20:18,842 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.nodemanager.sleep-delay-before-sigkill.ms = [250]
2016-12-02 19:20:18,842 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.nodemanager.vmem-check-enabled = [false]
2016-12-02 19:20:18,842 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.nodemanager.vmem-pmem-ratio = [2.1]
2016-12-02 19:20:18,842 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.nodemanager.webapp.address = [0.0.0.0:8042]
2016-12-02 19:20:18,842 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.nodemanager.webapp.cross-origin.enabled = [false]
2016-12-02 19:20:18,842 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.nodemanager.webapp.rest-csrf.custom-header = [X-XSRF-Header]
2016-12-02 19:20:18,842 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.nodemanager.webapp.rest-csrf.enabled = [false]
2016-12-02 19:20:18,842 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.nodemanager.webapp.rest-csrf.methods-to-ignore = [GET,OPTIONS,HEAD]
2016-12-02 19:20:18,842 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.nodemanager.windows-container.cpu-limit.enabled = [false]
2016-12-02 19:20:18,842 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.nodemanager.windows-container.memory-limit.enabled = [false]
2016-12-02 19:20:18,842 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.resourcemanager.address = [hscale-dev1-nn:8050]
2016-12-02 19:20:18,842 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.resourcemanager.admin.address = [hscale-dev1-nn:8141]
2016-12-02 19:20:18,842 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.resourcemanager.admin.client.thread-count = [1]
2016-12-02 19:20:18,842 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.resourcemanager.am-rm-tokens.master-key-rolling-interval-secs = [86400]
2016-12-02 19:20:18,842 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.resourcemanager.am.max-attempts = [2]
2016-12-02 19:20:18,842 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.resourcemanager.amlauncher.thread-count = [50]
2016-12-02 19:20:18,842 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.resourcemanager.bind-host = [0.0.0.0]
2016-12-02 19:20:18,842 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.resourcemanager.client.thread-count = [50]
2016-12-02 19:20:18,842 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.resourcemanager.configuration.provider-class = [org.apache.hadoop.yarn.LocalConfigurationProvider]
2016-12-02 19:20:18,842 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.resourcemanager.connect.max-wait.ms = [900000]
2016-12-02 19:20:18,842 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.resourcemanager.connect.retry-interval.ms = [30000]
2016-12-02 19:20:18,842 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.resourcemanager.container-tokens.master-key-rolling-interval-secs = [86400]
2016-12-02 19:20:18,842 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.resourcemanager.container.liveness-monitor.interval-ms = [600000]
2016-12-02 19:20:18,842 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.resourcemanager.delayed.delegation-token.removal-interval-ms = [30000]
2016-12-02 19:20:18,842 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.resourcemanager.fail-fast = [false]
2016-12-02 19:20:18,842 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.resourcemanager.fs.state-store.num-retries = [0]
2016-12-02 19:20:18,842 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.resourcemanager.fs.state-store.retry-interval-ms = [1000]
2016-12-02 19:20:18,843 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.resourcemanager.fs.state-store.retry-policy-spec = [2000, 500]
2016-12-02 19:20:18,843 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.resourcemanager.fs.state-store.uri = [ ]
2016-12-02 19:20:18,843 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.resourcemanager.ha.automatic-failover.embedded = [true]
2016-12-02 19:20:18,843 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.resourcemanager.ha.automatic-failover.enabled = [true]
2016-12-02 19:20:18,843 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.resourcemanager.ha.automatic-failover.zk-base-path = [/yarn-leader-election]
2016-12-02 19:20:18,843 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.resourcemanager.ha.enabled = [false]
2016-12-02 19:20:18,843 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.resourcemanager.hostname = [hscale-dev1-nn]
2016-12-02 19:20:18,843 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.resourcemanager.keytab = [/etc/krb5.keytab]
2016-12-02 19:20:18,843 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.resourcemanager.leveldb-state-store.path = [/tmp/hadoop-hbase/yarn/system/rmstore]
2016-12-02 19:20:18,843 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.resourcemanager.max-completed-applications = [10000]
2016-12-02 19:20:18,843 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.resourcemanager.max-log-aggregation-diagnostics-in-memory = [10]
2016-12-02 19:20:18,843 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.resourcemanager.nodemanager-connect-retries = [10]
2016-12-02 19:20:18,843 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.resourcemanager.nodemanager.minimum.version = [NONE]
2016-12-02 19:20:18,843 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.resourcemanager.nodemanagers.heartbeat-interval-ms = [1000]
2016-12-02 19:20:18,843 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.resourcemanager.nodes.exclude-path = [/etc/hadoop/conf/yarn.exclude]
2016-12-02 19:20:18,843 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.resourcemanager.proxy-user-privileges.enabled = [false]
2016-12-02 19:20:18,843 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.resourcemanager.recovery.enabled = [true]
2016-12-02 19:20:18,843 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.resourcemanager.resource-tracker.address = [hscale-dev1-nn:8025]
2016-12-02 19:20:18,843 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.resourcemanager.resource-tracker.client.thread-count = [50]
2016-12-02 19:20:18,843 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.resourcemanager.scheduler.address = [hscale-dev1-nn:8030]
2016-12-02 19:20:18,843 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.resourcemanager.scheduler.class = [org.apache.hadoop.yarn.server.resourcemanager.scheduler.capacity.CapacityScheduler]
2016-12-02 19:20:18,843 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.resourcemanager.scheduler.client.thread-count = [50]
2016-12-02 19:20:18,843 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.resourcemanager.scheduler.monitor.enable = [false]
2016-12-02 19:20:18,843 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.resourcemanager.scheduler.monitor.policies = [org.apache.hadoop.yarn.server.resourcemanager.monitor.capacity.ProportionalCapacityPreemptionPolicy]
2016-12-02 19:20:18,843 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.resourcemanager.state-store.max-completed-applications = [10000]
2016-12-02 19:20:18,843 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.resourcemanager.store.class = [org.apache.hadoop.yarn.server.resourcemanager.recovery.ZKRMStateStore]
2016-12-02 19:20:18,843 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.resourcemanager.system-metrics-publisher.dispatcher.pool-size = [10]
2016-12-02 19:20:18,843 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.resourcemanager.system-metrics-publisher.enabled = [true]
2016-12-02 19:20:18,843 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.resourcemanager.webapp.address = [hscale-dev1-nn:8088]
2016-12-02 19:20:18,843 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.resourcemanager.webapp.cross-origin.enabled = [true]
2016-12-02 19:20:18,843 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.resourcemanager.webapp.delegation-token-auth-filter.enabled = [false]
2016-12-02 19:20:18,843 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.resourcemanager.webapp.https.address = [hscale-dev1-nn:8090]
2016-12-02 19:20:18,843 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.resourcemanager.webapp.rest-csrf.custom-header = [X-XSRF-Header]
2016-12-02 19:20:18,843 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.resourcemanager.webapp.rest-csrf.enabled = [false]
2016-12-02 19:20:18,843 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.resourcemanager.webapp.rest-csrf.methods-to-ignore = [GET,OPTIONS,HEAD]
2016-12-02 19:20:18,843 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.resourcemanager.work-preserving-recovery.enabled = [true]
2016-12-02 19:20:18,843 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.resourcemanager.work-preserving-recovery.scheduling-wait-ms = [10000]
2016-12-02 19:20:18,843 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.resourcemanager.zk-acl = [world:anyone:rwcda]
2016-12-02 19:20:18,843 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.resourcemanager.zk-address = [hscale-dev1-dn1:2181,hscale-dev1-dn3:2181,hscale-dev1-dn2:2181,hscale-dev1-dn4:2181]
2016-12-02 19:20:18,843 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.resourcemanager.zk-num-retries = [1000]
2016-12-02 19:20:18,843 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.resourcemanager.zk-retry-interval-ms = [1000]
2016-12-02 19:20:18,843 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.resourcemanager.zk-state-store.parent-path = [/rmstore]
2016-12-02 19:20:18,843 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.resourcemanager.zk-timeout-ms = [10000]
2016-12-02 19:20:18,843 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.scheduler.maximum-allocation-mb = [5120]
2016-12-02 19:20:18,843 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.scheduler.maximum-allocation-vcores = [8]
2016-12-02 19:20:18,843 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.scheduler.minimum-allocation-mb = [512]
2016-12-02 19:20:18,843 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.scheduler.minimum-allocation-vcores = [1]
2016-12-02 19:20:18,843 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.sharedcache.admin.address = [0.0.0.0:8047]
2016-12-02 19:20:18,843 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.sharedcache.admin.thread-count = [1]
2016-12-02 19:20:18,843 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.sharedcache.app-checker.class = [org.apache.hadoop.yarn.server.sharedcachemanager.RemoteAppChecker]
2016-12-02 19:20:18,843 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.sharedcache.checksum.algo.impl = [org.apache.hadoop.yarn.sharedcache.ChecksumSHA256Impl]
2016-12-02 19:20:18,843 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.sharedcache.cleaner.initial-delay-mins = [10]
2016-12-02 19:20:18,843 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.sharedcache.cleaner.period-mins = [1440]
2016-12-02 19:20:18,843 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.sharedcache.cleaner.resource-sleep-ms = [0]
2016-12-02 19:20:18,843 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.sharedcache.client-server.address = [0.0.0.0:8045]
2016-12-02 19:20:18,843 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.sharedcache.client-server.thread-count = [50]
2016-12-02 19:20:18,843 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.sharedcache.enabled = [false]
2016-12-02 19:20:18,843 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.sharedcache.nested-level = [3]
2016-12-02 19:20:18,843 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.sharedcache.nm.uploader.replication.factor = [10]
2016-12-02 19:20:18,843 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.sharedcache.nm.uploader.thread-count = [20]
2016-12-02 19:20:18,843 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.sharedcache.root-dir = [/sharedcache]
2016-12-02 19:20:18,843 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.sharedcache.store.class = [org.apache.hadoop.yarn.server.sharedcachemanager.store.InMemorySCMStore]
2016-12-02 19:20:18,843 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.sharedcache.store.in-memory.check-period-mins = [720]
2016-12-02 19:20:18,843 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.sharedcache.store.in-memory.initial-delay-mins = [10]
2016-12-02 19:20:18,843 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.sharedcache.store.in-memory.staleness-period-mins = [10080]
2016-12-02 19:20:18,843 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.sharedcache.uploader.server.address = [0.0.0.0:8046]
2016-12-02 19:20:18,843 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.sharedcache.uploader.server.thread-count = [50]
2016-12-02 19:20:18,843 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.sharedcache.webapp.address = [0.0.0.0:8788]
2016-12-02 19:20:18,843 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.timeline-service.address = [hscale-dev1-nn:10200]
2016-12-02 19:20:18,843 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.timeline-service.bind-host = [0.0.0.0]
2016-12-02 19:20:18,843 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.timeline-service.client.best-effort = [false]
2016-12-02 19:20:18,843 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.timeline-service.client.max-retries = [30]
2016-12-02 19:20:18,843 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.timeline-service.client.retry-interval-ms = [1000]
2016-12-02 19:20:18,843 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.timeline-service.enabled = [true]
2016-12-02 19:20:18,843 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.timeline-service.entity-group-fs-store.active-dir = [/ats/active/]
2016-12-02 19:20:18,844 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.timeline-service.entity-group-fs-store.cleaner-interval-seconds = [3600]
2016-12-02 19:20:18,844 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.timeline-service.entity-group-fs-store.done-dir = [/ats/done/]
2016-12-02 19:20:18,844 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.timeline-service.entity-group-fs-store.retain-seconds = [604800]
2016-12-02 19:20:18,844 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.timeline-service.entity-group-fs-store.scan-interval-seconds = [60]
2016-12-02 19:20:18,844 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.timeline-service.entity-group-fs-store.summary-store = [org.apache.hadoop.yarn.server.timeline.RollingLevelDBTimelineStore]
2016-12-02 19:20:18,844 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.timeline-service.generic-application-history.max-applications = [10000]
2016-12-02 19:20:18,844 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.timeline-service.generic-application-history.save-non-am-container-meta-info = [false]
2016-12-02 19:20:18,844 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.timeline-service.generic-application-history.store-class = [org.apache.hadoop.yarn.server.applicationhistoryservice.NullApplicationHistoryStore]
2016-12-02 19:20:18,844 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.timeline-service.handler-thread-count = [10]
2016-12-02 19:20:18,844 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.timeline-service.hostname = [0.0.0.0]
2016-12-02 19:20:18,844 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.timeline-service.http-authentication.simple.anonymous.allowed = [true]
2016-12-02 19:20:18,844 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.timeline-service.http-authentication.type = [simple]
2016-12-02 19:20:18,844 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.timeline-service.keytab = [/etc/krb5.keytab]
2016-12-02 19:20:18,844 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.timeline-service.leveldb-state-store.path = [/hadoop/yarn/timeline]
2016-12-02 19:20:18,844 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.timeline-service.leveldb-timeline-store.path = [/hadoop/yarn/timeline]
2016-12-02 19:20:18,844 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.timeline-service.leveldb-timeline-store.read-cache-size = [104857600]
2016-12-02 19:20:18,844 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.timeline-service.leveldb-timeline-store.start-time-read-cache-size = [10000]
2016-12-02 19:20:18,844 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.timeline-service.leveldb-timeline-store.start-time-write-cache-size = [10000]
2016-12-02 19:20:18,844 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.timeline-service.leveldb-timeline-store.ttl-interval-ms = [300000]
2016-12-02 19:20:18,844 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.timeline-service.recovery.enabled = [true]
2016-12-02 19:20:18,844 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.timeline-service.state-store-class = [org.apache.hadoop.yarn.server.timeline.recovery.LeveldbTimelineStateStore]
2016-12-02 19:20:18,844 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.timeline-service.store-class = [org.apache.hadoop.yarn.server.timeline.EntityGroupFSTimelineStore]
2016-12-02 19:20:18,844 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.timeline-service.ttl-enable = [true]
2016-12-02 19:20:18,844 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.timeline-service.ttl-ms = [2678400000]
2016-12-02 19:20:18,844 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.timeline-service.version = [1.5]
2016-12-02 19:20:18,844 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.timeline-service.webapp.address = [hscale-dev1-nn:8188]
2016-12-02 19:20:18,844 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.timeline-service.webapp.https.address = [hscale-dev1-nn:8190]
2016-12-02 19:20:18,844 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.timeline-service.webapp.rest-csrf.custom-header = [X-XSRF-Header]
2016-12-02 19:20:18,844 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.timeline-service.webapp.rest-csrf.enabled = [false]
2016-12-02 19:20:18,844 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  yarn.timeline-service.webapp.rest-csrf.methods-to-ignore = [GET,OPTIONS,HEAD]
2016-12-02 19:20:18,844 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  zookeeper.session.timeout = [120000]
2016-12-02 19:20:18,844 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  zookeeper.znode.acl.parent = [acl]
2016-12-02 19:20:18,844 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  zookeeper.znode.parent = [/hbase-unsecure]
2016-12-02 19:20:18,844 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] splice.config:  zookeeper.znode.rootserver = [root-region-server]
2016-12-02 19:20:18,844 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] zookeeper.RecoverableZooKeeper: Process identifier=spliceconnection connecting to ZooKeeper ensemble=hscale-dev1-dn1:2181,hscale-dev1-dn3:2181,hscale-dev1-dn2:2181,hscale-dev1-dn4:2181
2016-12-02 19:20:18,844 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] zookeeper.ZooKeeper: Initiating client connection, connectString=hscale-dev1-dn1:2181,hscale-dev1-dn3:2181,hscale-dev1-dn2:2181,hscale-dev1-dn4:2181 sessionTimeout=120000 watcher=spliceconnection0x0, quorum=hscale-dev1-dn1:2181,hscale-dev1-dn3:2181,hscale-dev1-dn2:2181,hscale-dev1-dn4:2181, baseZNode=/hbase-unsecure
2016-12-02 19:20:18,845 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020-SendThread(hscale-dev1-dn2:2181)] zookeeper.ClientCnxn: Opening socket connection to server hscale-dev1-dn2/10.60.70.12:2181. Will not attempt to authenticate using SASL (unknown error)
2016-12-02 19:20:18,846 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020-SendThread(hscale-dev1-dn2:2181)] zookeeper.ClientCnxn: Socket connection established to hscale-dev1-dn2/10.60.70.12:2181, initiating session
2016-12-02 19:20:18,855 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] hbase.ZkTimestampSource: Creating the TimestampClient...
2016-12-02 19:20:18,857 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] zookeeper.RecoverableZooKeeper: Process identifier=hconnection-0x3327a80e connecting to ZooKeeper ensemble=hscale-dev1-dn1:2181,hscale-dev1-dn3:2181,hscale-dev1-dn2:2181,hscale-dev1-dn4:2181
2016-12-02 19:20:18,857 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] zookeeper.ZooKeeper: Initiating client connection, connectString=hscale-dev1-dn1:2181,hscale-dev1-dn3:2181,hscale-dev1-dn2:2181,hscale-dev1-dn4:2181 sessionTimeout=120000 watcher=hconnection-0x3327a80e0x0, quorum=hscale-dev1-dn1:2181,hscale-dev1-dn3:2181,hscale-dev1-dn2:2181,hscale-dev1-dn4:2181, baseZNode=/hbase-unsecure
2016-12-02 19:20:18,858 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020-SendThread(hscale-dev1-dn2:2181)] zookeeper.ClientCnxn: Opening socket connection to server hscale-dev1-dn2/10.60.70.12:2181. Will not attempt to authenticate using SASL (unknown error)
2016-12-02 19:20:18,858 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020-SendThread(hscale-dev1-dn2:2181)] zookeeper.ClientCnxn: Socket connection established to hscale-dev1-dn2/10.60.70.12:2181, initiating session
2016-12-02 19:20:18,860 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020-SendThread(hscale-dev1-dn2:2181)] zookeeper.ClientCnxn: Session establishment complete on server hscale-dev1-dn2/10.60.70.12:2181, sessionid = 0x258bf3f6848002b, negotiated timeout = 120000
2016-12-02 19:20:18,885 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020-SendThread(hscale-dev1-dn2:2181)] zookeeper.ClientCnxn: Session establishment complete on server hscale-dev1-dn2/10.60.70.12:2181, sessionid = 0x258bf3f6848002c, negotiated timeout = 120000
2016-12-02 19:20:18,935 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] impl.TimestampClient: TimestampClient on region server successfully registered with JMX
2016-12-02 19:20:19,548 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] coprocessor.CoprocessorHost: System coprocessor com.splicemachine.hbase.RegionServerLifecycleObserver was loaded successfully with priority (536870911).
2016-12-02 19:20:19,549 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] regionserver.HRegionServer: reportForDuty to master=hscale-dev1-nn,16000,1480686229143 with port=16020, startcode=1480686617338
2016-12-02 19:20:19,559 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] hfile.CacheConfig: Allocating LruBlockCache size=768 MB, blockSize=64 KB
2016-12-02 19:20:19,565 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] hfile.CacheConfig: blockCache=LruBlockCache{blockCount=0, currentSize=827112, freeSize=804479256, maxSize=805306368, heapSize=827112, minSize=765041024, minFactor=0.95, multiSize=382520512, multiFactor=0.5, singleSize=191260256, singleFactor=0.25}, cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=true, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false
2016-12-02 19:20:19,672 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] wal.WALFactory: Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.BoundedRegionGroupingProvider
2016-12-02 19:20:19,673 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] wal.RegionGroupingProvider: Instantiating RegionGroupingStrategy of type class org.apache.hadoop.hbase.wal.RegionGroupingProvider$IdentityGroupingStrategy
2016-12-02 19:20:19,673 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] wal.WALFactory: Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.DefaultWALProvider
2016-12-02 19:20:19,730 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] wal.FSHLog: WAL configuration: blocksize=128 MB, rollsize=121.60 MB, prefix=hscale-dev1-dn1%2C16020%2C1480686617338.null0, suffix=, logDir=hdfs://hscale-dev1-nn:8020/apps/hbase/data/WALs/hscale-dev1-dn1,16020,1480686617338, archiveDir=hdfs://hscale-dev1-nn:8020/apps/hbase/data/oldWALs
2016-12-02 19:20:19,934 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] wal.FSHLog: Slow sync cost: 152 ms, current pipeline: []
2016-12-02 19:20:19,935 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] wal.FSHLog: New WAL /apps/hbase/data/WALs/hscale-dev1-dn1,16020,1480686617338/hscale-dev1-dn1%2C16020%2C1480686617338.null0.1480686619730
2016-12-02 19:20:19,948 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] wal.WALFactory: Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.DefaultWALProvider
2016-12-02 19:20:19,952 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] wal.FSHLog: WAL configuration: blocksize=128 MB, rollsize=121.60 MB, prefix=hscale-dev1-dn1%2C16020%2C1480686617338.null1, suffix=, logDir=hdfs://hscale-dev1-nn:8020/apps/hbase/data/WALs/hscale-dev1-dn1,16020,1480686617338, archiveDir=hdfs://hscale-dev1-nn:8020/apps/hbase/data/oldWALs
2016-12-02 19:20:20,004 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] wal.FSHLog: Slow sync cost: 33 ms, current pipeline: []
2016-12-02 19:20:20,004 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] wal.FSHLog: New WAL /apps/hbase/data/WALs/hscale-dev1-dn1,16020,1480686617338/hscale-dev1-dn1%2C16020%2C1480686617338.null1.1480686619952
2016-12-02 19:20:20,005 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] wal.WALFactory: Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.DefaultWALProvider
2016-12-02 19:20:20,010 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] wal.FSHLog: WAL configuration: blocksize=128 MB, rollsize=121.60 MB, prefix=hscale-dev1-dn1%2C16020%2C1480686617338.null2, suffix=, logDir=hdfs://hscale-dev1-nn:8020/apps/hbase/data/WALs/hscale-dev1-dn1,16020,1480686617338, archiveDir=hdfs://hscale-dev1-nn:8020/apps/hbase/data/oldWALs
2016-12-02 19:20:20,058 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] wal.FSHLog: Slow sync cost: 32 ms, current pipeline: []
2016-12-02 19:20:20,058 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] wal.FSHLog: New WAL /apps/hbase/data/WALs/hscale-dev1-dn1,16020,1480686617338/hscale-dev1-dn1%2C16020%2C1480686617338.null2.1480686620010
2016-12-02 19:20:20,059 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] wal.WALFactory: Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.DefaultWALProvider
2016-12-02 19:20:20,064 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] wal.FSHLog: WAL configuration: blocksize=128 MB, rollsize=121.60 MB, prefix=hscale-dev1-dn1%2C16020%2C1480686617338.null3, suffix=, logDir=hdfs://hscale-dev1-nn:8020/apps/hbase/data/WALs/hscale-dev1-dn1,16020,1480686617338, archiveDir=hdfs://hscale-dev1-nn:8020/apps/hbase/data/oldWALs
2016-12-02 19:20:20,111 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] wal.FSHLog: Slow sync cost: 32 ms, current pipeline: []
2016-12-02 19:20:20,111 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] wal.FSHLog: New WAL /apps/hbase/data/WALs/hscale-dev1-dn1,16020,1480686617338/hscale-dev1-dn1%2C16020%2C1480686617338.null3.1480686620065
2016-12-02 19:20:20,115 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] wal.WALFactory: Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.DefaultWALProvider
2016-12-02 19:20:20,120 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] wal.FSHLog: WAL configuration: blocksize=128 MB, rollsize=121.60 MB, prefix=hscale-dev1-dn1%2C16020%2C1480686617338.null4, suffix=, logDir=hdfs://hscale-dev1-nn:8020/apps/hbase/data/WALs/hscale-dev1-dn1,16020,1480686617338, archiveDir=hdfs://hscale-dev1-nn:8020/apps/hbase/data/oldWALs
2016-12-02 19:20:20,175 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] wal.FSHLog: Slow sync cost: 41 ms, current pipeline: []
2016-12-02 19:20:20,175 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] wal.FSHLog: New WAL /apps/hbase/data/WALs/hscale-dev1-dn1,16020,1480686617338/hscale-dev1-dn1%2C16020%2C1480686617338.null4.1480686620120
2016-12-02 19:20:20,178 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] wal.WALFactory: Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.DefaultWALProvider
2016-12-02 19:20:20,187 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] wal.FSHLog: WAL configuration: blocksize=128 MB, rollsize=121.60 MB, prefix=hscale-dev1-dn1%2C16020%2C1480686617338.null5, suffix=, logDir=hdfs://hscale-dev1-nn:8020/apps/hbase/data/WALs/hscale-dev1-dn1,16020,1480686617338, archiveDir=hdfs://hscale-dev1-nn:8020/apps/hbase/data/oldWALs
2016-12-02 19:20:20,237 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] wal.FSHLog: Slow sync cost: 32 ms, current pipeline: []
2016-12-02 19:20:20,237 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] wal.FSHLog: New WAL /apps/hbase/data/WALs/hscale-dev1-dn1,16020,1480686617338/hscale-dev1-dn1%2C16020%2C1480686617338.null5.1480686620187
2016-12-02 19:20:20,238 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] wal.WALFactory: Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.DefaultWALProvider
2016-12-02 19:20:20,244 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] wal.FSHLog: WAL configuration: blocksize=128 MB, rollsize=121.60 MB, prefix=hscale-dev1-dn1%2C16020%2C1480686617338.null6, suffix=, logDir=hdfs://hscale-dev1-nn:8020/apps/hbase/data/WALs/hscale-dev1-dn1,16020,1480686617338, archiveDir=hdfs://hscale-dev1-nn:8020/apps/hbase/data/oldWALs
2016-12-02 19:20:20,291 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] wal.FSHLog: Slow sync cost: 32 ms, current pipeline: []
2016-12-02 19:20:20,291 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] wal.FSHLog: New WAL /apps/hbase/data/WALs/hscale-dev1-dn1,16020,1480686617338/hscale-dev1-dn1%2C16020%2C1480686617338.null6.1480686620244
2016-12-02 19:20:20,293 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] wal.WALFactory: Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.DefaultWALProvider
2016-12-02 19:20:20,300 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] wal.FSHLog: WAL configuration: blocksize=128 MB, rollsize=121.60 MB, prefix=hscale-dev1-dn1%2C16020%2C1480686617338.null7, suffix=, logDir=hdfs://hscale-dev1-nn:8020/apps/hbase/data/WALs/hscale-dev1-dn1,16020,1480686617338, archiveDir=hdfs://hscale-dev1-nn:8020/apps/hbase/data/oldWALs
2016-12-02 19:20:20,353 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] wal.FSHLog: Slow sync cost: 40 ms, current pipeline: []
2016-12-02 19:20:20,353 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] wal.FSHLog: New WAL /apps/hbase/data/WALs/hscale-dev1-dn1,16020,1480686617338/hscale-dev1-dn1%2C16020%2C1480686617338.null7.1480686620300
2016-12-02 19:20:20,354 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] wal.WALFactory: Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.DefaultWALProvider
2016-12-02 19:20:20,359 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] wal.FSHLog: WAL configuration: blocksize=128 MB, rollsize=121.60 MB, prefix=hscale-dev1-dn1%2C16020%2C1480686617338.null8, suffix=, logDir=hdfs://hscale-dev1-nn:8020/apps/hbase/data/WALs/hscale-dev1-dn1,16020,1480686617338, archiveDir=hdfs://hscale-dev1-nn:8020/apps/hbase/data/oldWALs
2016-12-02 19:20:20,416 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] wal.FSHLog: Slow sync cost: 41 ms, current pipeline: []
2016-12-02 19:20:20,416 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] wal.FSHLog: New WAL /apps/hbase/data/WALs/hscale-dev1-dn1,16020,1480686617338/hscale-dev1-dn1%2C16020%2C1480686617338.null8.1480686620359
2016-12-02 19:20:20,417 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] wal.WALFactory: Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.DefaultWALProvider
2016-12-02 19:20:20,426 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] wal.FSHLog: WAL configuration: blocksize=128 MB, rollsize=121.60 MB, prefix=hscale-dev1-dn1%2C16020%2C1480686617338.null9, suffix=, logDir=hdfs://hscale-dev1-nn:8020/apps/hbase/data/WALs/hscale-dev1-dn1,16020,1480686617338, archiveDir=hdfs://hscale-dev1-nn:8020/apps/hbase/data/oldWALs
2016-12-02 19:20:20,486 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] wal.FSHLog: Slow sync cost: 41 ms, current pipeline: []
2016-12-02 19:20:20,486 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] wal.FSHLog: New WAL /apps/hbase/data/WALs/hscale-dev1-dn1,16020,1480686617338/hscale-dev1-dn1%2C16020%2C1480686617338.null9.1480686620426
2016-12-02 19:20:20,488 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] wal.WALFactory: Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.DefaultWALProvider
2016-12-02 19:20:20,498 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] wal.FSHLog: WAL configuration: blocksize=128 MB, rollsize=121.60 MB, prefix=hscale-dev1-dn1%2C16020%2C1480686617338.null10, suffix=, logDir=hdfs://hscale-dev1-nn:8020/apps/hbase/data/WALs/hscale-dev1-dn1,16020,1480686617338, archiveDir=hdfs://hscale-dev1-nn:8020/apps/hbase/data/oldWALs
2016-12-02 19:20:20,572 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] wal.FSHLog: Slow sync cost: 55 ms, current pipeline: []
2016-12-02 19:20:20,572 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] wal.FSHLog: New WAL /apps/hbase/data/WALs/hscale-dev1-dn1,16020,1480686617338/hscale-dev1-dn1%2C16020%2C1480686617338.null10.1480686620498
2016-12-02 19:20:20,572 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] wal.WALFactory: Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.DefaultWALProvider
2016-12-02 19:20:20,578 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] wal.FSHLog: WAL configuration: blocksize=128 MB, rollsize=121.60 MB, prefix=hscale-dev1-dn1%2C16020%2C1480686617338.null11, suffix=, logDir=hdfs://hscale-dev1-nn:8020/apps/hbase/data/WALs/hscale-dev1-dn1,16020,1480686617338, archiveDir=hdfs://hscale-dev1-nn:8020/apps/hbase/data/oldWALs
2016-12-02 19:20:20,640 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] wal.FSHLog: Slow sync cost: 40 ms, current pipeline: []
2016-12-02 19:20:20,640 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] wal.FSHLog: New WAL /apps/hbase/data/WALs/hscale-dev1-dn1,16020,1480686617338/hscale-dev1-dn1%2C16020%2C1480686617338.null11.1480686620578
2016-12-02 19:20:20,642 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] wal.WALFactory: Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.DefaultWALProvider
2016-12-02 19:20:20,653 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] wal.FSHLog: WAL configuration: blocksize=128 MB, rollsize=121.60 MB, prefix=hscale-dev1-dn1%2C16020%2C1480686617338.null12, suffix=, logDir=hdfs://hscale-dev1-nn:8020/apps/hbase/data/WALs/hscale-dev1-dn1,16020,1480686617338, archiveDir=hdfs://hscale-dev1-nn:8020/apps/hbase/data/oldWALs
2016-12-02 19:20:20,710 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] wal.FSHLog: Slow sync cost: 41 ms, current pipeline: []
2016-12-02 19:20:20,711 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] wal.FSHLog: New WAL /apps/hbase/data/WALs/hscale-dev1-dn1,16020,1480686617338/hscale-dev1-dn1%2C16020%2C1480686617338.null12.1480686620653
2016-12-02 19:20:20,712 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] wal.WALFactory: Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.DefaultWALProvider
2016-12-02 19:20:20,723 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] wal.FSHLog: WAL configuration: blocksize=128 MB, rollsize=121.60 MB, prefix=hscale-dev1-dn1%2C16020%2C1480686617338.null13, suffix=, logDir=hdfs://hscale-dev1-nn:8020/apps/hbase/data/WALs/hscale-dev1-dn1,16020,1480686617338, archiveDir=hdfs://hscale-dev1-nn:8020/apps/hbase/data/oldWALs
2016-12-02 19:20:20,782 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] wal.FSHLog: Slow sync cost: 41 ms, current pipeline: []
2016-12-02 19:20:20,782 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] wal.FSHLog: New WAL /apps/hbase/data/WALs/hscale-dev1-dn1,16020,1480686617338/hscale-dev1-dn1%2C16020%2C1480686617338.null13.1480686620723
2016-12-02 19:20:20,782 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] wal.WALFactory: Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.DefaultWALProvider
2016-12-02 19:20:20,787 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] wal.FSHLog: WAL configuration: blocksize=128 MB, rollsize=121.60 MB, prefix=hscale-dev1-dn1%2C16020%2C1480686617338.null14, suffix=, logDir=hdfs://hscale-dev1-nn:8020/apps/hbase/data/WALs/hscale-dev1-dn1,16020,1480686617338, archiveDir=hdfs://hscale-dev1-nn:8020/apps/hbase/data/oldWALs
2016-12-02 19:20:20,844 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] wal.FSHLog: Slow sync cost: 41 ms, current pipeline: []
2016-12-02 19:20:20,844 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] wal.FSHLog: New WAL /apps/hbase/data/WALs/hscale-dev1-dn1,16020,1480686617338/hscale-dev1-dn1%2C16020%2C1480686617338.null14.1480686620787
2016-12-02 19:20:20,846 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] wal.WALFactory: Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.DefaultWALProvider
2016-12-02 19:20:20,857 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] wal.FSHLog: WAL configuration: blocksize=128 MB, rollsize=121.60 MB, prefix=hscale-dev1-dn1%2C16020%2C1480686617338.null15, suffix=, logDir=hdfs://hscale-dev1-nn:8020/apps/hbase/data/WALs/hscale-dev1-dn1,16020,1480686617338, archiveDir=hdfs://hscale-dev1-nn:8020/apps/hbase/data/oldWALs
2016-12-02 19:20:20,923 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] wal.FSHLog: Slow sync cost: 48 ms, current pipeline: []
2016-12-02 19:20:20,923 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] wal.FSHLog: New WAL /apps/hbase/data/WALs/hscale-dev1-dn1,16020,1480686617338/hscale-dev1-dn1%2C16020%2C1480686617338.null15.1480686620857
2016-12-02 19:20:20,925 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] wal.BoundedRegionGroupingProvider: Configured to run with 16 delegate WAL providers.
2016-12-02 19:20:20,934 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] regionserver.MetricsRegionServerWrapperImpl: Computing regionserver metrics every 5000 milliseconds
2016-12-02 19:20:20,943 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] regionserver.ReplicationSourceManager: Current list of replicators: [hscale-dev1-dn1,16020,1480686617338] other RSs: [hscale-dev1-dn1,16020,1480686617338]
2016-12-02 19:20:20,987 INFO  [SplitLogWorker-hscale-dev1-dn1:16020] regionserver.SplitLogWorker: SplitLogWorker hscale-dev1-dn1,16020,1480686617338 starting
2016-12-02 19:20:20,988 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] regionserver.HeapMemoryManager: Starting HeapMemoryTuner chore.
2016-12-02 19:20:20,989 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] regionserver.HRegionServer: Serving as hscale-dev1-dn1,16020,1480686617338, RpcServer on hscale-dev1-dn1/10.60.70.11:16020, sessionid=0x458bf3f64d4002f
2016-12-02 19:20:21,000 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020] quotas.RegionServerQuotaManager: Quota support disabled
2016-12-02 19:20:21,020 INFO  [SplitLogWorker-hscale-dev1-dn1:16020] coordination.ZkSplitLogWorkerCoordination: worker hscale-dev1-dn1,16020,1480686617338 acquired task /hbase-unsecure/splitWAL/WALs%2Fhscale-dev1-dn3%2C16020%2C1480686232696-splitting%2Fhscale-dev1-dn3%252C16020%252C1480686232696..meta.1480686237399.meta
2016-12-02 19:20:21,087 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-0] wal.WALSplitter: Splitting wal: hdfs://hscale-dev1-nn:8020/apps/hbase/data/WALs/hscale-dev1-dn3,16020,1480686232696-splitting/hscale-dev1-dn3%2C16020%2C1480686232696..meta.1480686237399.meta, length=5443
2016-12-02 19:20:21,087 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-0] wal.WALSplitter: DistributedLogReplay = false
2016-12-02 19:20:21,120 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-0] util.FSHDFSUtils: Recovering lease on dfs file hdfs://hscale-dev1-nn:8020/apps/hbase/data/WALs/hscale-dev1-dn3,16020,1480686232696-splitting/hscale-dev1-dn3%2C16020%2C1480686232696..meta.1480686237399.meta
2016-12-02 19:20:21,152 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-0] util.FSHDFSUtils: recoverLease=true, attempt=0 on file=hdfs://hscale-dev1-nn:8020/apps/hbase/data/WALs/hscale-dev1-dn3,16020,1480686232696-splitting/hscale-dev1-dn3%2C16020%2C1480686232696..meta.1480686237399.meta after 32ms
2016-12-02 19:20:21,301 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-0-Writer-2] wal.WALSplitter: Creating writer path=hdfs://hscale-dev1-nn:8020/apps/hbase/data/data/hbase/meta/1588230740/recovered.edits/0000000000000000041.temp region=1588230740
2016-12-02 19:20:21,308 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-0] wal.WALSplitter: Split writers finished
2016-12-02 19:20:21,382 INFO  [split-log-closeStream-1] wal.WALSplitter: Rename hdfs://hscale-dev1-nn:8020/apps/hbase/data/data/hbase/meta/1588230740/recovered.edits/0000000000000000041.temp to hdfs://hscale-dev1-nn:8020/apps/hbase/data/data/hbase/meta/1588230740/recovered.edits/0000000000000000091
2016-12-02 19:20:21,382 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-0] wal.WALSplitter: Processed 26 edits across 1 regions; edits skipped=0; log file=hdfs://hscale-dev1-nn:8020/apps/hbase/data/WALs/hscale-dev1-dn3,16020,1480686232696-splitting/hscale-dev1-dn3%2C16020%2C1480686232696..meta.1480686237399.meta, length=5443, corrupted=false, progress failed=false
2016-12-02 19:20:21,415 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-0] coordination.ZkSplitLogWorkerCoordination: successfully transitioned task /hbase-unsecure/splitWAL/WALs%2Fhscale-dev1-dn3%2C16020%2C1480686232696-splitting%2Fhscale-dev1-dn3%252C16020%252C1480686232696..meta.1480686237399.meta to final state DONE hscale-dev1-dn1,16020,1480686617338
2016-12-02 19:20:21,415 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-0] handler.WALSplitterHandler: worker hscale-dev1-dn1,16020,1480686617338 done with task org.apache.hadoop.hbase.coordination.ZkSplitLogWorkerCoordination$ZkSplitTaskDetails@2cf38ed5 in 371ms
2016-12-02 19:20:34,022 WARN  [hbase-region-load-updater-0] client.ConnectionManager$HConnectionImplementation: Checking master connection
com.google.protobuf.ServiceException: org.apache.hadoop.hbase.exceptions.ConnectionClosingException: Call to hscale-dev1-nn/10.60.70.10:16000 failed on local exception: org.apache.hadoop.hbase.exceptions.ConnectionClosingException: Connection to hscale-dev1-nn/10.60.70.10:16000 is closing. Call id=30, waitTime=1
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient.callBlockingMethod(AbstractRpcClient.java:223)
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient$BlockingRpcChannelImplementation.callBlockingMethod(AbstractRpcClient.java:287)
    at org.apache.hadoop.hbase.protobuf.generated.MasterProtos$MasterService$BlockingStub.isMasterRunning(MasterProtos.java:58152)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation$MasterServiceState.isMasterRunning(ConnectionManager.java:1444)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.isKeepAliveMasterConnectedAndRunning(ConnectionManager.java:2099)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.getKeepAliveMasterService(ConnectionManager.java:1708)
    at org.apache.hadoop.hbase.client.MasterCallable.prepare(MasterCallable.java:38)
    at org.apache.hadoop.hbase.client.RpcRetryingCaller.callWithRetries(RpcRetryingCaller.java:124)
    at org.apache.hadoop.hbase.client.HBaseAdmin.executeCallable(HBaseAdmin.java:4083)
    at org.apache.hadoop.hbase.client.HBaseAdmin.getClusterStatus(HBaseAdmin.java:2765)
    at com.splicemachine.access.hbase.H10PartitionAdmin.allServers(H10PartitionAdmin.java:120)
    at com.splicemachine.hbase.HBaseRegionLoads.fetchRegionLoads(HBaseRegionLoads.java:164)
    at com.splicemachine.hbase.HBaseRegionLoads.access$000(HBaseRegionLoads.java:60)
    at com.splicemachine.hbase.HBaseRegionLoads$1.run(HBaseRegionLoads.java:81)
    at com.splicemachine.concurrent.LoggingScheduledThreadPoolExecutor$LoggingRunnable.run(LoggingScheduledThreadPoolExecutor.java:75)
    at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
    at java.util.concurrent.FutureTask.runAndReset(FutureTask.java:308)
    at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$301(ScheduledThreadPoolExecutor.java:180)
    at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:294)
    at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
    at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
    at java.lang.Thread.run(Thread.java:745)
Caused by: org.apache.hadoop.hbase.exceptions.ConnectionClosingException: Call to hscale-dev1-nn/10.60.70.10:16000 failed on local exception: org.apache.hadoop.hbase.exceptions.ConnectionClosingException: Connection to hscale-dev1-nn/10.60.70.10:16000 is closing. Call id=30, waitTime=1
    at org.apache.hadoop.hbase.ipc.RpcClientImpl.wrapException(RpcClientImpl.java:1259)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl.call(RpcClientImpl.java:1230)
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient.callBlockingMethod(AbstractRpcClient.java:213)
    ... 21 more
Caused by: org.apache.hadoop.hbase.exceptions.ConnectionClosingException: Connection to hscale-dev1-nn/10.60.70.10:16000 is closing. Call id=30, waitTime=1
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.cleanupCalls(RpcClientImpl.java:1047)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.close(RpcClientImpl.java:846)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.run(RpcClientImpl.java:574)
2016-12-02 19:20:34,232 WARN  [hbase-region-load-updater-0] client.ConnectionManager$HConnectionImplementation: Checking master connection
com.google.protobuf.ServiceException: java.net.ConnectException: Connection refused
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient.callBlockingMethod(AbstractRpcClient.java:223)
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient$BlockingRpcChannelImplementation.callBlockingMethod(AbstractRpcClient.java:287)
    at org.apache.hadoop.hbase.protobuf.generated.MasterProtos$MasterService$BlockingStub.isMasterRunning(MasterProtos.java:58152)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation$MasterServiceState.isMasterRunning(ConnectionManager.java:1444)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.isKeepAliveMasterConnectedAndRunning(ConnectionManager.java:2099)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.getKeepAliveMasterService(ConnectionManager.java:1708)
    at org.apache.hadoop.hbase.client.MasterCallable.prepare(MasterCallable.java:38)
    at org.apache.hadoop.hbase.client.RpcRetryingCaller.callWithRetries(RpcRetryingCaller.java:124)
    at org.apache.hadoop.hbase.client.HBaseAdmin.executeCallable(HBaseAdmin.java:4083)
    at org.apache.hadoop.hbase.client.HBaseAdmin.getClusterStatus(HBaseAdmin.java:2765)
    at com.splicemachine.access.hbase.H10PartitionAdmin.allServers(H10PartitionAdmin.java:120)
    at com.splicemachine.hbase.HBaseRegionLoads.fetchRegionLoads(HBaseRegionLoads.java:164)
    at com.splicemachine.hbase.HBaseRegionLoads.access$000(HBaseRegionLoads.java:60)
    at com.splicemachine.hbase.HBaseRegionLoads$1.run(HBaseRegionLoads.java:81)
    at com.splicemachine.concurrent.LoggingScheduledThreadPoolExecutor$LoggingRunnable.run(LoggingScheduledThreadPoolExecutor.java:75)
    at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
    at java.util.concurrent.FutureTask.runAndReset(FutureTask.java:308)
    at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$301(ScheduledThreadPoolExecutor.java:180)
    at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:294)
    at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
    at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
    at java.lang.Thread.run(Thread.java:745)
Caused by: java.net.ConnectException: Connection refused
    at sun.nio.ch.SocketChannelImpl.checkConnect(Native Method)
    at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:717)
    at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:206)
    at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:531)
    at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:495)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.setupConnection(RpcClientImpl.java:410)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.setupIOstreams(RpcClientImpl.java:716)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.writeRequest(RpcClientImpl.java:887)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.tracedWriteRequest(RpcClientImpl.java:856)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl.call(RpcClientImpl.java:1200)
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient.callBlockingMethod(AbstractRpcClient.java:213)
    ... 21 more
2016-12-02 19:20:34,538 WARN  [hbase-region-load-updater-0] client.ConnectionManager$HConnectionImplementation: Checking master connection
com.google.protobuf.ServiceException: org.apache.hadoop.hbase.ipc.FailedServerException: This server is in the failed servers list: hscale-dev1-nn/10.60.70.10:16000
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient.callBlockingMethod(AbstractRpcClient.java:223)
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient$BlockingRpcChannelImplementation.callBlockingMethod(AbstractRpcClient.java:287)
    at org.apache.hadoop.hbase.protobuf.generated.MasterProtos$MasterService$BlockingStub.isMasterRunning(MasterProtos.java:58152)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation$MasterServiceState.isMasterRunning(ConnectionManager.java:1444)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.isKeepAliveMasterConnectedAndRunning(ConnectionManager.java:2099)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.getKeepAliveMasterService(ConnectionManager.java:1708)
    at org.apache.hadoop.hbase.client.MasterCallable.prepare(MasterCallable.java:38)
    at org.apache.hadoop.hbase.client.RpcRetryingCaller.callWithRetries(RpcRetryingCaller.java:124)
    at org.apache.hadoop.hbase.client.HBaseAdmin.executeCallable(HBaseAdmin.java:4083)
    at org.apache.hadoop.hbase.client.HBaseAdmin.getClusterStatus(HBaseAdmin.java:2765)
    at com.splicemachine.access.hbase.H10PartitionAdmin.allServers(H10PartitionAdmin.java:120)
    at com.splicemachine.hbase.HBaseRegionLoads.fetchRegionLoads(HBaseRegionLoads.java:164)
    at com.splicemachine.hbase.HBaseRegionLoads.access$000(HBaseRegionLoads.java:60)
    at com.splicemachine.hbase.HBaseRegionLoads$1.run(HBaseRegionLoads.java:81)
    at com.splicemachine.concurrent.LoggingScheduledThreadPoolExecutor$LoggingRunnable.run(LoggingScheduledThreadPoolExecutor.java:75)
    at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
    at java.util.concurrent.FutureTask.runAndReset(FutureTask.java:308)
    at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$301(ScheduledThreadPoolExecutor.java:180)
    at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:294)
    at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
    at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
    at java.lang.Thread.run(Thread.java:745)
Caused by: org.apache.hadoop.hbase.ipc.FailedServerException: This server is in the failed servers list: hscale-dev1-nn/10.60.70.10:16000
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.setupIOstreams(RpcClientImpl.java:701)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.writeRequest(RpcClientImpl.java:887)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.tracedWriteRequest(RpcClientImpl.java:856)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl.call(RpcClientImpl.java:1200)
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient.callBlockingMethod(AbstractRpcClient.java:213)
    ... 21 more
2016-12-02 19:20:35,046 WARN  [hbase-region-load-updater-0] client.ConnectionManager$HConnectionImplementation: Checking master connection
com.google.protobuf.ServiceException: org.apache.hadoop.hbase.ipc.FailedServerException: This server is in the failed servers list: hscale-dev1-nn/10.60.70.10:16000
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient.callBlockingMethod(AbstractRpcClient.java:223)
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient$BlockingRpcChannelImplementation.callBlockingMethod(AbstractRpcClient.java:287)
    at org.apache.hadoop.hbase.protobuf.generated.MasterProtos$MasterService$BlockingStub.isMasterRunning(MasterProtos.java:58152)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation$MasterServiceState.isMasterRunning(ConnectionManager.java:1444)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.isKeepAliveMasterConnectedAndRunning(ConnectionManager.java:2099)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.getKeepAliveMasterService(ConnectionManager.java:1708)
    at org.apache.hadoop.hbase.client.MasterCallable.prepare(MasterCallable.java:38)
    at org.apache.hadoop.hbase.client.RpcRetryingCaller.callWithRetries(RpcRetryingCaller.java:124)
    at org.apache.hadoop.hbase.client.HBaseAdmin.executeCallable(HBaseAdmin.java:4083)
    at org.apache.hadoop.hbase.client.HBaseAdmin.getClusterStatus(HBaseAdmin.java:2765)
    at com.splicemachine.access.hbase.H10PartitionAdmin.allServers(H10PartitionAdmin.java:120)
    at com.splicemachine.hbase.HBaseRegionLoads.fetchRegionLoads(HBaseRegionLoads.java:164)
    at com.splicemachine.hbase.HBaseRegionLoads.access$000(HBaseRegionLoads.java:60)
    at com.splicemachine.hbase.HBaseRegionLoads$1.run(HBaseRegionLoads.java:81)
    at com.splicemachine.concurrent.LoggingScheduledThreadPoolExecutor$LoggingRunnable.run(LoggingScheduledThreadPoolExecutor.java:75)
    at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
    at java.util.concurrent.FutureTask.runAndReset(FutureTask.java:308)
    at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$301(ScheduledThreadPoolExecutor.java:180)
    at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:294)
    at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
    at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
    at java.lang.Thread.run(Thread.java:745)
Caused by: org.apache.hadoop.hbase.ipc.FailedServerException: This server is in the failed servers list: hscale-dev1-nn/10.60.70.10:16000
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.setupIOstreams(RpcClientImpl.java:701)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.writeRequest(RpcClientImpl.java:887)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.tracedWriteRequest(RpcClientImpl.java:856)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl.call(RpcClientImpl.java:1200)
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient.callBlockingMethod(AbstractRpcClient.java:213)
    ... 21 more
2016-12-02 19:20:36,057 WARN  [hbase-region-load-updater-0] client.ConnectionManager$HConnectionImplementation: Checking master connection
com.google.protobuf.ServiceException: org.apache.hadoop.hbase.ipc.FailedServerException: This server is in the failed servers list: hscale-dev1-nn/10.60.70.10:16000
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient.callBlockingMethod(AbstractRpcClient.java:223)
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient$BlockingRpcChannelImplementation.callBlockingMethod(AbstractRpcClient.java:287)
    at org.apache.hadoop.hbase.protobuf.generated.MasterProtos$MasterService$BlockingStub.isMasterRunning(MasterProtos.java:58152)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation$MasterServiceState.isMasterRunning(ConnectionManager.java:1444)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.isKeepAliveMasterConnectedAndRunning(ConnectionManager.java:2099)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.getKeepAliveMasterService(ConnectionManager.java:1708)
    at org.apache.hadoop.hbase.client.MasterCallable.prepare(MasterCallable.java:38)
    at org.apache.hadoop.hbase.client.RpcRetryingCaller.callWithRetries(RpcRetryingCaller.java:124)
    at org.apache.hadoop.hbase.client.HBaseAdmin.executeCallable(HBaseAdmin.java:4083)
    at org.apache.hadoop.hbase.client.HBaseAdmin.getClusterStatus(HBaseAdmin.java:2765)
    at com.splicemachine.access.hbase.H10PartitionAdmin.allServers(H10PartitionAdmin.java:120)
    at com.splicemachine.hbase.HBaseRegionLoads.fetchRegionLoads(HBaseRegionLoads.java:164)
    at com.splicemachine.hbase.HBaseRegionLoads.access$000(HBaseRegionLoads.java:60)
    at com.splicemachine.hbase.HBaseRegionLoads$1.run(HBaseRegionLoads.java:81)
    at com.splicemachine.concurrent.LoggingScheduledThreadPoolExecutor$LoggingRunnable.run(LoggingScheduledThreadPoolExecutor.java:75)
    at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
    at java.util.concurrent.FutureTask.runAndReset(FutureTask.java:308)
    at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$301(ScheduledThreadPoolExecutor.java:180)
    at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:294)
    at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
    at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
    at java.lang.Thread.run(Thread.java:745)
Caused by: org.apache.hadoop.hbase.ipc.FailedServerException: This server is in the failed servers list: hscale-dev1-nn/10.60.70.10:16000
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.setupIOstreams(RpcClientImpl.java:701)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.writeRequest(RpcClientImpl.java:887)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.tracedWriteRequest(RpcClientImpl.java:856)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl.call(RpcClientImpl.java:1200)
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient.callBlockingMethod(AbstractRpcClient.java:213)
    ... 21 more
2016-12-02 19:20:36,065 ERROR [hbase-region-load-updater-0] hbase.HBaseRegionLoads: Unable to fetch region load info
org.apache.hadoop.hbase.client.RetriesExhaustedException: Failed after attempts=5, exceptions:
Fri Dec 02 19:20:34 IST 2016, RpcRetryingCaller{globalStartTime=1480686634019, pause=100, retries=5}, org.apache.hadoop.hbase.MasterNotRunningException: java.io.IOException: Can't get master address from ZooKeeper; znode data == null
Fri Dec 02 19:20:34 IST 2016, RpcRetryingCaller{globalStartTime=1480686634019, pause=100, retries=5}, org.apache.hadoop.hbase.MasterNotRunningException: java.io.IOException: Can't get master address from ZooKeeper; znode data == null
Fri Dec 02 19:20:34 IST 2016, RpcRetryingCaller{globalStartTime=1480686634019, pause=100, retries=5}, org.apache.hadoop.hbase.MasterNotRunningException: java.io.IOException: Can't get master address from ZooKeeper; znode data == null
Fri Dec 02 19:20:35 IST 2016, RpcRetryingCaller{globalStartTime=1480686634019, pause=100, retries=5}, org.apache.hadoop.hbase.MasterNotRunningException: java.io.IOException: Can't get master address from ZooKeeper; znode data == null
Fri Dec 02 19:20:36 IST 2016, RpcRetryingCaller{globalStartTime=1480686634019, pause=100, retries=5}, org.apache.hadoop.hbase.MasterNotRunningException: java.io.IOException: Can't get master address from ZooKeeper; znode data == null

    at org.apache.hadoop.hbase.client.RpcRetryingCaller.callWithRetries(RpcRetryingCaller.java:147)
    at org.apache.hadoop.hbase.client.HBaseAdmin.executeCallable(HBaseAdmin.java:4083)
    at org.apache.hadoop.hbase.client.HBaseAdmin.getClusterStatus(HBaseAdmin.java:2765)
    at com.splicemachine.access.hbase.H10PartitionAdmin.allServers(H10PartitionAdmin.java:120)
    at com.splicemachine.hbase.HBaseRegionLoads.fetchRegionLoads(HBaseRegionLoads.java:164)
    at com.splicemachine.hbase.HBaseRegionLoads.access$000(HBaseRegionLoads.java:60)
    at com.splicemachine.hbase.HBaseRegionLoads$1.run(HBaseRegionLoads.java:81)
    at com.splicemachine.concurrent.LoggingScheduledThreadPoolExecutor$LoggingRunnable.run(LoggingScheduledThreadPoolExecutor.java:75)
    at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
    at java.util.concurrent.FutureTask.runAndReset(FutureTask.java:308)
    at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$301(ScheduledThreadPoolExecutor.java:180)
    at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:294)
    at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
    at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
    at java.lang.Thread.run(Thread.java:745)
Caused by: org.apache.hadoop.hbase.MasterNotRunningException: java.io.IOException: Can't get master address from ZooKeeper; znode data == null
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation$StubMaker.makeStub(ConnectionManager.java:1540)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation$MasterServiceStubMaker.makeStub(ConnectionManager.java:1560)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.getKeepAliveMasterService(ConnectionManager.java:1711)
    at org.apache.hadoop.hbase.client.MasterCallable.prepare(MasterCallable.java:38)
    at org.apache.hadoop.hbase.client.RpcRetryingCaller.callWithRetries(RpcRetryingCaller.java:124)
    ... 14 more
Caused by: java.io.IOException: Can't get master address from ZooKeeper; znode data == null
    at org.apache.hadoop.hbase.zookeeper.MasterAddressTracker.getMasterAddress(MasterAddressTracker.java:154)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation$StubMaker.makeStubNoRetries(ConnectionManager.java:1491)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation$StubMaker.makeStub(ConnectionManager.java:1531)
    ... 18 more
2016-12-02 19:20:38,023 WARN  [pool-10-thread-1] client.ConnectionManager$HConnectionImplementation: Checking master connection
com.google.protobuf.ServiceException: java.net.ConnectException: Connection refused
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient.callBlockingMethod(AbstractRpcClient.java:223)
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient$BlockingRpcChannelImplementation.callBlockingMethod(AbstractRpcClient.java:287)
    at org.apache.hadoop.hbase.protobuf.generated.MasterProtos$MasterService$BlockingStub.isMasterRunning(MasterProtos.java:58152)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation$MasterServiceState.isMasterRunning(ConnectionManager.java:1444)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.isKeepAliveMasterConnectedAndRunning(ConnectionManager.java:2099)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.getKeepAliveMasterService(ConnectionManager.java:1708)
    at org.apache.hadoop.hbase.client.MasterCallable.prepare(MasterCallable.java:38)
    at org.apache.hadoop.hbase.client.RpcRetryingCaller.callWithRetries(RpcRetryingCaller.java:124)
    at org.apache.hadoop.hbase.client.HBaseAdmin.executeCallable(HBaseAdmin.java:4083)
    at org.apache.hadoop.hbase.client.HBaseAdmin.access$700(HBaseAdmin.java:194)
    at org.apache.hadoop.hbase.client.HBaseAdmin$ProcedureFuture.getProcedureResult(HBaseAdmin.java:4383)
    at org.apache.hadoop.hbase.client.HBaseAdmin$ProcedureFuture.waitProcedureResult(HBaseAdmin.java:4335)
    at org.apache.hadoop.hbase.client.HBaseAdmin$ProcedureFuture.get(HBaseAdmin.java:4291)
    at org.apache.hadoop.hbase.client.HBaseAdmin.createTable(HBaseAdmin.java:647)
    at org.apache.hadoop.hbase.client.HBaseAdmin.createTable(HBaseAdmin.java:577)
    at com.splicemachine.lifecycle.RegionServerLifecycle.distributedStart(RegionServerLifecycle.java:66)
    at com.splicemachine.derby.lifecycle.EngineLifecycleService.start(EngineLifecycleService.java:81)
    at com.splicemachine.lifecycle.DatabaseLifecycleManager$Startup.bootServices(DatabaseLifecycleManager.java:229)
    at com.splicemachine.lifecycle.DatabaseLifecycleManager$Startup.run(DatabaseLifecycleManager.java:220)
    at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
    at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
    at java.lang.Thread.run(Thread.java:745)
Caused by: java.net.ConnectException: Connection refused
    at sun.nio.ch.SocketChannelImpl.checkConnect(Native Method)
    at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:717)
    at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:206)
    at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:531)
    at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:495)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.setupConnection(RpcClientImpl.java:410)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.setupIOstreams(RpcClientImpl.java:716)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.writeRequest(RpcClientImpl.java:887)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.tracedWriteRequest(RpcClientImpl.java:856)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl.call(RpcClientImpl.java:1200)
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient.callBlockingMethod(AbstractRpcClient.java:213)
    ... 21 more
2016-12-02 19:20:38,227 WARN  [pool-10-thread-1] client.ConnectionManager$HConnectionImplementation: Checking master connection
com.google.protobuf.ServiceException: org.apache.hadoop.hbase.ipc.FailedServerException: This server is in the failed servers list: hscale-dev1-nn/10.60.70.10:16000
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient.callBlockingMethod(AbstractRpcClient.java:223)
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient$BlockingRpcChannelImplementation.callBlockingMethod(AbstractRpcClient.java:287)
    at org.apache.hadoop.hbase.protobuf.generated.MasterProtos$MasterService$BlockingStub.isMasterRunning(MasterProtos.java:58152)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation$MasterServiceState.isMasterRunning(ConnectionManager.java:1444)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.isKeepAliveMasterConnectedAndRunning(ConnectionManager.java:2099)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.getKeepAliveMasterService(ConnectionManager.java:1708)
    at org.apache.hadoop.hbase.client.MasterCallable.prepare(MasterCallable.java:38)
    at org.apache.hadoop.hbase.client.RpcRetryingCaller.callWithRetries(RpcRetryingCaller.java:124)
    at org.apache.hadoop.hbase.client.HBaseAdmin.executeCallable(HBaseAdmin.java:4083)
    at org.apache.hadoop.hbase.client.HBaseAdmin.access$700(HBaseAdmin.java:194)
    at org.apache.hadoop.hbase.client.HBaseAdmin$ProcedureFuture.getProcedureResult(HBaseAdmin.java:4383)
    at org.apache.hadoop.hbase.client.HBaseAdmin$ProcedureFuture.waitProcedureResult(HBaseAdmin.java:4335)
    at org.apache.hadoop.hbase.client.HBaseAdmin$ProcedureFuture.get(HBaseAdmin.java:4291)
    at org.apache.hadoop.hbase.client.HBaseAdmin.createTable(HBaseAdmin.java:647)
    at org.apache.hadoop.hbase.client.HBaseAdmin.createTable(HBaseAdmin.java:577)
    at com.splicemachine.lifecycle.RegionServerLifecycle.distributedStart(RegionServerLifecycle.java:66)
    at com.splicemachine.derby.lifecycle.EngineLifecycleService.start(EngineLifecycleService.java:81)
    at com.splicemachine.lifecycle.DatabaseLifecycleManager$Startup.bootServices(DatabaseLifecycleManager.java:229)
    at com.splicemachine.lifecycle.DatabaseLifecycleManager$Startup.run(DatabaseLifecycleManager.java:220)
    at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
    at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
    at java.lang.Thread.run(Thread.java:745)
Caused by: org.apache.hadoop.hbase.ipc.FailedServerException: This server is in the failed servers list: hscale-dev1-nn/10.60.70.10:16000
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.setupIOstreams(RpcClientImpl.java:701)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.writeRequest(RpcClientImpl.java:887)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.tracedWriteRequest(RpcClientImpl.java:856)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl.call(RpcClientImpl.java:1200)
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient.callBlockingMethod(AbstractRpcClient.java:213)
    ... 21 more
2016-12-02 19:20:38,530 WARN  [pool-10-thread-1] client.ConnectionManager$HConnectionImplementation: Checking master connection
com.google.protobuf.ServiceException: org.apache.hadoop.hbase.ipc.FailedServerException: This server is in the failed servers list: hscale-dev1-nn/10.60.70.10:16000
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient.callBlockingMethod(AbstractRpcClient.java:223)
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient$BlockingRpcChannelImplementation.callBlockingMethod(AbstractRpcClient.java:287)
    at org.apache.hadoop.hbase.protobuf.generated.MasterProtos$MasterService$BlockingStub.isMasterRunning(MasterProtos.java:58152)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation$MasterServiceState.isMasterRunning(ConnectionManager.java:1444)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.isKeepAliveMasterConnectedAndRunning(ConnectionManager.java:2099)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.getKeepAliveMasterService(ConnectionManager.java:1708)
    at org.apache.hadoop.hbase.client.MasterCallable.prepare(MasterCallable.java:38)
    at org.apache.hadoop.hbase.client.RpcRetryingCaller.callWithRetries(RpcRetryingCaller.java:124)
    at org.apache.hadoop.hbase.client.HBaseAdmin.executeCallable(HBaseAdmin.java:4083)
    at org.apache.hadoop.hbase.client.HBaseAdmin.access$700(HBaseAdmin.java:194)
    at org.apache.hadoop.hbase.client.HBaseAdmin$ProcedureFuture.getProcedureResult(HBaseAdmin.java:4383)
    at org.apache.hadoop.hbase.client.HBaseAdmin$ProcedureFuture.waitProcedureResult(HBaseAdmin.java:4335)
    at org.apache.hadoop.hbase.client.HBaseAdmin$ProcedureFuture.get(HBaseAdmin.java:4291)
    at org.apache.hadoop.hbase.client.HBaseAdmin.createTable(HBaseAdmin.java:647)
    at org.apache.hadoop.hbase.client.HBaseAdmin.createTable(HBaseAdmin.java:577)
    at com.splicemachine.lifecycle.RegionServerLifecycle.distributedStart(RegionServerLifecycle.java:66)
    at com.splicemachine.derby.lifecycle.EngineLifecycleService.start(EngineLifecycleService.java:81)
    at com.splicemachine.lifecycle.DatabaseLifecycleManager$Startup.bootServices(DatabaseLifecycleManager.java:229)
    at com.splicemachine.lifecycle.DatabaseLifecycleManager$Startup.run(DatabaseLifecycleManager.java:220)
    at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
    at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
    at java.lang.Thread.run(Thread.java:745)
Caused by: org.apache.hadoop.hbase.ipc.FailedServerException: This server is in the failed servers list: hscale-dev1-nn/10.60.70.10:16000
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.setupIOstreams(RpcClientImpl.java:701)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.writeRequest(RpcClientImpl.java:887)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.tracedWriteRequest(RpcClientImpl.java:856)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl.call(RpcClientImpl.java:1200)
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient.callBlockingMethod(AbstractRpcClient.java:213)
    ... 21 more
2016-12-02 19:20:39,019 WARN  [hbase-region-load-updater-0] client.ConnectionManager$HConnectionImplementation: Checking master connection
com.google.protobuf.ServiceException: org.apache.hadoop.hbase.ipc.FailedServerException: This server is in the failed servers list: hscale-dev1-nn/10.60.70.10:16000
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient.callBlockingMethod(AbstractRpcClient.java:223)
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient$BlockingRpcChannelImplementation.callBlockingMethod(AbstractRpcClient.java:287)
    at org.apache.hadoop.hbase.protobuf.generated.MasterProtos$MasterService$BlockingStub.isMasterRunning(MasterProtos.java:58152)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation$MasterServiceState.isMasterRunning(ConnectionManager.java:1444)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.isKeepAliveMasterConnectedAndRunning(ConnectionManager.java:2099)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.getKeepAliveMasterService(ConnectionManager.java:1708)
    at org.apache.hadoop.hbase.client.MasterCallable.prepare(MasterCallable.java:38)
    at org.apache.hadoop.hbase.client.RpcRetryingCaller.callWithRetries(RpcRetryingCaller.java:124)
    at org.apache.hadoop.hbase.client.HBaseAdmin.executeCallable(HBaseAdmin.java:4083)
    at org.apache.hadoop.hbase.client.HBaseAdmin.getClusterStatus(HBaseAdmin.java:2765)
    at com.splicemachine.access.hbase.H10PartitionAdmin.allServers(H10PartitionAdmin.java:120)
    at com.splicemachine.hbase.HBaseRegionLoads.fetchRegionLoads(HBaseRegionLoads.java:164)
    at com.splicemachine.hbase.HBaseRegionLoads.access$000(HBaseRegionLoads.java:60)
    at com.splicemachine.hbase.HBaseRegionLoads$1.run(HBaseRegionLoads.java:81)
    at com.splicemachine.concurrent.LoggingScheduledThreadPoolExecutor$LoggingRunnable.run(LoggingScheduledThreadPoolExecutor.java:75)
    at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
    at java.util.concurrent.FutureTask.runAndReset(FutureTask.java:308)
    at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$301(ScheduledThreadPoolExecutor.java:180)
    at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:294)
    at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
    at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
    at java.lang.Thread.run(Thread.java:745)
Caused by: org.apache.hadoop.hbase.ipc.FailedServerException: This server is in the failed servers list: hscale-dev1-nn/10.60.70.10:16000
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.setupIOstreams(RpcClientImpl.java:701)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.writeRequest(RpcClientImpl.java:887)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.tracedWriteRequest(RpcClientImpl.java:856)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl.call(RpcClientImpl.java:1200)
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient.callBlockingMethod(AbstractRpcClient.java:213)
    ... 21 more
2016-12-02 19:20:39,038 WARN  [pool-10-thread-1] client.ConnectionManager$HConnectionImplementation: Checking master connection
com.google.protobuf.ServiceException: org.apache.hadoop.hbase.ipc.FailedServerException: This server is in the failed servers list: hscale-dev1-nn/10.60.70.10:16000
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient.callBlockingMethod(AbstractRpcClient.java:223)
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient$BlockingRpcChannelImplementation.callBlockingMethod(AbstractRpcClient.java:287)
    at org.apache.hadoop.hbase.protobuf.generated.MasterProtos$MasterService$BlockingStub.isMasterRunning(MasterProtos.java:58152)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation$MasterServiceState.isMasterRunning(ConnectionManager.java:1444)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.isKeepAliveMasterConnectedAndRunning(ConnectionManager.java:2099)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.getKeepAliveMasterService(ConnectionManager.java:1708)
    at org.apache.hadoop.hbase.client.MasterCallable.prepare(MasterCallable.java:38)
    at org.apache.hadoop.hbase.client.RpcRetryingCaller.callWithRetries(RpcRetryingCaller.java:124)
    at org.apache.hadoop.hbase.client.HBaseAdmin.executeCallable(HBaseAdmin.java:4083)
    at org.apache.hadoop.hbase.client.HBaseAdmin.access$700(HBaseAdmin.java:194)
    at org.apache.hadoop.hbase.client.HBaseAdmin$ProcedureFuture.getProcedureResult(HBaseAdmin.java:4383)
    at org.apache.hadoop.hbase.client.HBaseAdmin$ProcedureFuture.waitProcedureResult(HBaseAdmin.java:4335)
    at org.apache.hadoop.hbase.client.HBaseAdmin$ProcedureFuture.get(HBaseAdmin.java:4291)
    at org.apache.hadoop.hbase.client.HBaseAdmin.createTable(HBaseAdmin.java:647)
    at org.apache.hadoop.hbase.client.HBaseAdmin.createTable(HBaseAdmin.java:577)
    at com.splicemachine.lifecycle.RegionServerLifecycle.distributedStart(RegionServerLifecycle.java:66)
    at com.splicemachine.derby.lifecycle.EngineLifecycleService.start(EngineLifecycleService.java:81)
    at com.splicemachine.lifecycle.DatabaseLifecycleManager$Startup.bootServices(DatabaseLifecycleManager.java:229)
    at com.splicemachine.lifecycle.DatabaseLifecycleManager$Startup.run(DatabaseLifecycleManager.java:220)
    at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
    at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
    at java.lang.Thread.run(Thread.java:745)
Caused by: org.apache.hadoop.hbase.ipc.FailedServerException: This server is in the failed servers list: hscale-dev1-nn/10.60.70.10:16000
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.setupIOstreams(RpcClientImpl.java:701)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.writeRequest(RpcClientImpl.java:887)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.tracedWriteRequest(RpcClientImpl.java:856)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl.call(RpcClientImpl.java:1200)
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient.callBlockingMethod(AbstractRpcClient.java:213)
    ... 21 more
2016-12-02 19:20:39,225 WARN  [hbase-region-load-updater-0] client.ConnectionManager$HConnectionImplementation: Checking master connection
com.google.protobuf.ServiceException: org.apache.hadoop.hbase.ipc.FailedServerException: This server is in the failed servers list: hscale-dev1-nn/10.60.70.10:16000
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient.callBlockingMethod(AbstractRpcClient.java:223)
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient$BlockingRpcChannelImplementation.callBlockingMethod(AbstractRpcClient.java:287)
    at org.apache.hadoop.hbase.protobuf.generated.MasterProtos$MasterService$BlockingStub.isMasterRunning(MasterProtos.java:58152)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation$MasterServiceState.isMasterRunning(ConnectionManager.java:1444)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.isKeepAliveMasterConnectedAndRunning(ConnectionManager.java:2099)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.getKeepAliveMasterService(ConnectionManager.java:1708)
    at org.apache.hadoop.hbase.client.MasterCallable.prepare(MasterCallable.java:38)
    at org.apache.hadoop.hbase.client.RpcRetryingCaller.callWithRetries(RpcRetryingCaller.java:124)
    at org.apache.hadoop.hbase.client.HBaseAdmin.executeCallable(HBaseAdmin.java:4083)
    at org.apache.hadoop.hbase.client.HBaseAdmin.getClusterStatus(HBaseAdmin.java:2765)
    at com.splicemachine.access.hbase.H10PartitionAdmin.allServers(H10PartitionAdmin.java:120)
    at com.splicemachine.hbase.HBaseRegionLoads.fetchRegionLoads(HBaseRegionLoads.java:164)
    at com.splicemachine.hbase.HBaseRegionLoads.access$000(HBaseRegionLoads.java:60)
    at com.splicemachine.hbase.HBaseRegionLoads$1.run(HBaseRegionLoads.java:81)
    at com.splicemachine.concurrent.LoggingScheduledThreadPoolExecutor$LoggingRunnable.run(LoggingScheduledThreadPoolExecutor.java:75)
    at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
    at java.util.concurrent.FutureTask.runAndReset(FutureTask.java:308)
    at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$301(ScheduledThreadPoolExecutor.java:180)
    at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:294)
    at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
    at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
    at java.lang.Thread.run(Thread.java:745)
Caused by: org.apache.hadoop.hbase.ipc.FailedServerException: This server is in the failed servers list: hscale-dev1-nn/10.60.70.10:16000
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.setupIOstreams(RpcClientImpl.java:701)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.writeRequest(RpcClientImpl.java:887)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.tracedWriteRequest(RpcClientImpl.java:856)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl.call(RpcClientImpl.java:1200)
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient.callBlockingMethod(AbstractRpcClient.java:213)
    ... 21 more
2016-12-02 19:20:39,530 WARN  [hbase-region-load-updater-0] client.ConnectionManager$HConnectionImplementation: Checking master connection
com.google.protobuf.ServiceException: org.apache.hadoop.hbase.ipc.FailedServerException: This server is in the failed servers list: hscale-dev1-nn/10.60.70.10:16000
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient.callBlockingMethod(AbstractRpcClient.java:223)
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient$BlockingRpcChannelImplementation.callBlockingMethod(AbstractRpcClient.java:287)
    at org.apache.hadoop.hbase.protobuf.generated.MasterProtos$MasterService$BlockingStub.isMasterRunning(MasterProtos.java:58152)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation$MasterServiceState.isMasterRunning(ConnectionManager.java:1444)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.isKeepAliveMasterConnectedAndRunning(ConnectionManager.java:2099)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.getKeepAliveMasterService(ConnectionManager.java:1708)
    at org.apache.hadoop.hbase.client.MasterCallable.prepare(MasterCallable.java:38)
    at org.apache.hadoop.hbase.client.RpcRetryingCaller.callWithRetries(RpcRetryingCaller.java:124)
    at org.apache.hadoop.hbase.client.HBaseAdmin.executeCallable(HBaseAdmin.java:4083)
    at org.apache.hadoop.hbase.client.HBaseAdmin.getClusterStatus(HBaseAdmin.java:2765)
    at com.splicemachine.access.hbase.H10PartitionAdmin.allServers(H10PartitionAdmin.java:120)
    at com.splicemachine.hbase.HBaseRegionLoads.fetchRegionLoads(HBaseRegionLoads.java:164)
    at com.splicemachine.hbase.HBaseRegionLoads.access$000(HBaseRegionLoads.java:60)
    at com.splicemachine.hbase.HBaseRegionLoads$1.run(HBaseRegionLoads.java:81)
    at com.splicemachine.concurrent.LoggingScheduledThreadPoolExecutor$LoggingRunnable.run(LoggingScheduledThreadPoolExecutor.java:75)
    at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
    at java.util.concurrent.FutureTask.runAndReset(FutureTask.java:308)
    at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$301(ScheduledThreadPoolExecutor.java:180)
    at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:294)
    at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
    at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
    at java.lang.Thread.run(Thread.java:745)
Caused by: org.apache.hadoop.hbase.ipc.FailedServerException: This server is in the failed servers list: hscale-dev1-nn/10.60.70.10:16000
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.setupIOstreams(RpcClientImpl.java:701)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.writeRequest(RpcClientImpl.java:887)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.tracedWriteRequest(RpcClientImpl.java:856)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl.call(RpcClientImpl.java:1200)
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient.callBlockingMethod(AbstractRpcClient.java:213)
    ... 21 more
2016-12-02 19:20:40,037 WARN  [hbase-region-load-updater-0] client.ConnectionManager$HConnectionImplementation: Checking master connection
com.google.protobuf.ServiceException: java.net.ConnectException: Connection refused
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient.callBlockingMethod(AbstractRpcClient.java:223)
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient$BlockingRpcChannelImplementation.callBlockingMethod(AbstractRpcClient.java:287)
    at org.apache.hadoop.hbase.protobuf.generated.MasterProtos$MasterService$BlockingStub.isMasterRunning(MasterProtos.java:58152)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation$MasterServiceState.isMasterRunning(ConnectionManager.java:1444)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.isKeepAliveMasterConnectedAndRunning(ConnectionManager.java:2099)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.getKeepAliveMasterService(ConnectionManager.java:1708)
    at org.apache.hadoop.hbase.client.MasterCallable.prepare(MasterCallable.java:38)
    at org.apache.hadoop.hbase.client.RpcRetryingCaller.callWithRetries(RpcRetryingCaller.java:124)
    at org.apache.hadoop.hbase.client.HBaseAdmin.executeCallable(HBaseAdmin.java:4083)
    at org.apache.hadoop.hbase.client.HBaseAdmin.getClusterStatus(HBaseAdmin.java:2765)
    at com.splicemachine.access.hbase.H10PartitionAdmin.allServers(H10PartitionAdmin.java:120)
    at com.splicemachine.hbase.HBaseRegionLoads.fetchRegionLoads(HBaseRegionLoads.java:164)
    at com.splicemachine.hbase.HBaseRegionLoads.access$000(HBaseRegionLoads.java:60)
    at com.splicemachine.hbase.HBaseRegionLoads$1.run(HBaseRegionLoads.java:81)
    at com.splicemachine.concurrent.LoggingScheduledThreadPoolExecutor$LoggingRunnable.run(LoggingScheduledThreadPoolExecutor.java:75)
    at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
    at java.util.concurrent.FutureTask.runAndReset(FutureTask.java:308)
    at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$301(ScheduledThreadPoolExecutor.java:180)
    at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:294)
    at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
    at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
    at java.lang.Thread.run(Thread.java:745)
Caused by: java.net.ConnectException: Connection refused
    at sun.nio.ch.SocketChannelImpl.checkConnect(Native Method)
    at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:717)
    at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:206)
    at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:531)
    at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:495)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.setupConnection(RpcClientImpl.java:410)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.setupIOstreams(RpcClientImpl.java:716)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.writeRequest(RpcClientImpl.java:887)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.tracedWriteRequest(RpcClientImpl.java:856)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl.call(RpcClientImpl.java:1200)
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient.callBlockingMethod(AbstractRpcClient.java:213)
    ... 21 more
2016-12-02 19:20:40,042 WARN  [pool-10-thread-1] client.ConnectionManager$HConnectionImplementation: Checking master connection
com.google.protobuf.ServiceException: org.apache.hadoop.hbase.ipc.FailedServerException: This server is in the failed servers list: hscale-dev1-nn/10.60.70.10:16000
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient.callBlockingMethod(AbstractRpcClient.java:223)
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient$BlockingRpcChannelImplementation.callBlockingMethod(AbstractRpcClient.java:287)
    at org.apache.hadoop.hbase.protobuf.generated.MasterProtos$MasterService$BlockingStub.isMasterRunning(MasterProtos.java:58152)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation$MasterServiceState.isMasterRunning(ConnectionManager.java:1444)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.isKeepAliveMasterConnectedAndRunning(ConnectionManager.java:2099)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.getKeepAliveMasterService(ConnectionManager.java:1708)
    at org.apache.hadoop.hbase.client.MasterCallable.prepare(MasterCallable.java:38)
    at org.apache.hadoop.hbase.client.RpcRetryingCaller.callWithRetries(RpcRetryingCaller.java:124)
    at org.apache.hadoop.hbase.client.HBaseAdmin.executeCallable(HBaseAdmin.java:4083)
    at org.apache.hadoop.hbase.client.HBaseAdmin.access$700(HBaseAdmin.java:194)
    at org.apache.hadoop.hbase.client.HBaseAdmin$ProcedureFuture.getProcedureResult(HBaseAdmin.java:4383)
    at org.apache.hadoop.hbase.client.HBaseAdmin$ProcedureFuture.waitProcedureResult(HBaseAdmin.java:4335)
    at org.apache.hadoop.hbase.client.HBaseAdmin$ProcedureFuture.get(HBaseAdmin.java:4291)
    at org.apache.hadoop.hbase.client.HBaseAdmin.createTable(HBaseAdmin.java:647)
    at org.apache.hadoop.hbase.client.HBaseAdmin.createTable(HBaseAdmin.java:577)
    at com.splicemachine.lifecycle.RegionServerLifecycle.distributedStart(RegionServerLifecycle.java:66)
    at com.splicemachine.derby.lifecycle.EngineLifecycleService.start(EngineLifecycleService.java:81)
    at com.splicemachine.lifecycle.DatabaseLifecycleManager$Startup.bootServices(DatabaseLifecycleManager.java:229)
    at com.splicemachine.lifecycle.DatabaseLifecycleManager$Startup.run(DatabaseLifecycleManager.java:220)
    at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
    at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
    at java.lang.Thread.run(Thread.java:745)
Caused by: org.apache.hadoop.hbase.ipc.FailedServerException: This server is in the failed servers list: hscale-dev1-nn/10.60.70.10:16000
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.setupIOstreams(RpcClientImpl.java:701)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.writeRequest(RpcClientImpl.java:887)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.tracedWriteRequest(RpcClientImpl.java:856)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl.call(RpcClientImpl.java:1200)
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient.callBlockingMethod(AbstractRpcClient.java:213)
    ... 21 more
2016-12-02 19:20:40,046 WARN  [pool-10-thread-1] client.HBaseAdmin: failed to get the procedure result procId=6
org.apache.hadoop.hbase.client.RetriesExhaustedException: Failed after attempts=5, exceptions:
Fri Dec 02 19:20:38 IST 2016, RpcRetryingCaller{globalStartTime=1480686638022, pause=100, retries=5}, org.apache.hadoop.hbase.MasterNotRunningException: java.io.IOException: Can't get master address from ZooKeeper; znode data == null
Fri Dec 02 19:20:38 IST 2016, RpcRetryingCaller{globalStartTime=1480686638022, pause=100, retries=5}, org.apache.hadoop.hbase.MasterNotRunningException: java.io.IOException: Can't get master address from ZooKeeper; znode data == null
Fri Dec 02 19:20:38 IST 2016, RpcRetryingCaller{globalStartTime=1480686638022, pause=100, retries=5}, org.apache.hadoop.hbase.MasterNotRunningException: java.io.IOException: Can't get master address from ZooKeeper; znode data == null
Fri Dec 02 19:20:39 IST 2016, RpcRetryingCaller{globalStartTime=1480686638022, pause=100, retries=5}, org.apache.hadoop.hbase.MasterNotRunningException: java.io.IOException: Can't get master address from ZooKeeper; znode data == null
Fri Dec 02 19:20:40 IST 2016, RpcRetryingCaller{globalStartTime=1480686638022, pause=100, retries=5}, org.apache.hadoop.hbase.MasterNotRunningException: java.io.IOException: Can't get master address from ZooKeeper; znode data == null

    at org.apache.hadoop.hbase.client.RpcRetryingCaller.callWithRetries(RpcRetryingCaller.java:147)
    at org.apache.hadoop.hbase.client.HBaseAdmin.executeCallable(HBaseAdmin.java:4083)
    at org.apache.hadoop.hbase.client.HBaseAdmin.access$700(HBaseAdmin.java:194)
    at org.apache.hadoop.hbase.client.HBaseAdmin$ProcedureFuture.getProcedureResult(HBaseAdmin.java:4383)
    at org.apache.hadoop.hbase.client.HBaseAdmin$ProcedureFuture.waitProcedureResult(HBaseAdmin.java:4335)
    at org.apache.hadoop.hbase.client.HBaseAdmin$ProcedureFuture.get(HBaseAdmin.java:4291)
    at org.apache.hadoop.hbase.client.HBaseAdmin.createTable(HBaseAdmin.java:647)
    at org.apache.hadoop.hbase.client.HBaseAdmin.createTable(HBaseAdmin.java:577)
    at com.splicemachine.lifecycle.RegionServerLifecycle.distributedStart(RegionServerLifecycle.java:66)
    at com.splicemachine.derby.lifecycle.EngineLifecycleService.start(EngineLifecycleService.java:81)
    at com.splicemachine.lifecycle.DatabaseLifecycleManager$Startup.bootServices(DatabaseLifecycleManager.java:229)
    at com.splicemachine.lifecycle.DatabaseLifecycleManager$Startup.run(DatabaseLifecycleManager.java:220)
    at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
    at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
    at java.lang.Thread.run(Thread.java:745)
Caused by: org.apache.hadoop.hbase.MasterNotRunningException: java.io.IOException: Can't get master address from ZooKeeper; znode data == null
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation$StubMaker.makeStub(ConnectionManager.java:1540)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation$MasterServiceStubMaker.makeStub(ConnectionManager.java:1560)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.getKeepAliveMasterService(ConnectionManager.java:1711)
    at org.apache.hadoop.hbase.client.MasterCallable.prepare(MasterCallable.java:38)
    at org.apache.hadoop.hbase.client.RpcRetryingCaller.callWithRetries(RpcRetryingCaller.java:124)
    ... 14 more
Caused by: java.io.IOException: Can't get master address from ZooKeeper; znode data == null
    at org.apache.hadoop.hbase.zookeeper.MasterAddressTracker.getMasterAddress(MasterAddressTracker.java:154)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation$StubMaker.makeStubNoRetries(ConnectionManager.java:1491)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation$StubMaker.makeStub(ConnectionManager.java:1531)
    ... 18 more
2016-12-02 19:20:41,055 WARN  [hbase-region-load-updater-0] client.ConnectionManager$HConnectionImplementation: Checking master connection
com.google.protobuf.ServiceException: org.apache.hadoop.hbase.ipc.FailedServerException: This server is in the failed servers list: hscale-dev1-nn/10.60.70.10:16000
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient.callBlockingMethod(AbstractRpcClient.java:223)
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient$BlockingRpcChannelImplementation.callBlockingMethod(AbstractRpcClient.java:287)
    at org.apache.hadoop.hbase.protobuf.generated.MasterProtos$MasterService$BlockingStub.isMasterRunning(MasterProtos.java:58152)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation$MasterServiceState.isMasterRunning(ConnectionManager.java:1444)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.isKeepAliveMasterConnectedAndRunning(ConnectionManager.java:2099)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.getKeepAliveMasterService(ConnectionManager.java:1708)
    at org.apache.hadoop.hbase.client.MasterCallable.prepare(MasterCallable.java:38)
    at org.apache.hadoop.hbase.client.RpcRetryingCaller.callWithRetries(RpcRetryingCaller.java:124)
    at org.apache.hadoop.hbase.client.HBaseAdmin.executeCallable(HBaseAdmin.java:4083)
    at org.apache.hadoop.hbase.client.HBaseAdmin.getClusterStatus(HBaseAdmin.java:2765)
    at com.splicemachine.access.hbase.H10PartitionAdmin.allServers(H10PartitionAdmin.java:120)
    at com.splicemachine.hbase.HBaseRegionLoads.fetchRegionLoads(HBaseRegionLoads.java:164)
    at com.splicemachine.hbase.HBaseRegionLoads.access$000(HBaseRegionLoads.java:60)
    at com.splicemachine.hbase.HBaseRegionLoads$1.run(HBaseRegionLoads.java:81)
    at com.splicemachine.concurrent.LoggingScheduledThreadPoolExecutor$LoggingRunnable.run(LoggingScheduledThreadPoolExecutor.java:75)
    at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
    at java.util.concurrent.FutureTask.runAndReset(FutureTask.java:308)
    at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$301(ScheduledThreadPoolExecutor.java:180)
    at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:294)
    at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
    at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
    at java.lang.Thread.run(Thread.java:745)
Caused by: org.apache.hadoop.hbase.ipc.FailedServerException: This server is in the failed servers list: hscale-dev1-nn/10.60.70.10:16000
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.setupIOstreams(RpcClientImpl.java:701)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.writeRequest(RpcClientImpl.java:887)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.tracedWriteRequest(RpcClientImpl.java:856)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl.call(RpcClientImpl.java:1200)
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient.callBlockingMethod(AbstractRpcClient.java:213)
    ... 21 more
2016-12-02 19:20:41,059 ERROR [hbase-region-load-updater-0] hbase.HBaseRegionLoads: Unable to fetch region load info
org.apache.hadoop.hbase.client.RetriesExhaustedException: Failed after attempts=5, exceptions:
Fri Dec 02 19:20:39 IST 2016, RpcRetryingCaller{globalStartTime=1480686639019, pause=100, retries=5}, org.apache.hadoop.hbase.MasterNotRunningException: java.io.IOException: Can't get master address from ZooKeeper; znode data == null
Fri Dec 02 19:20:39 IST 2016, RpcRetryingCaller{globalStartTime=1480686639019, pause=100, retries=5}, org.apache.hadoop.hbase.MasterNotRunningException: java.io.IOException: Can't get master address from ZooKeeper; znode data == null
Fri Dec 02 19:20:39 IST 2016, RpcRetryingCaller{globalStartTime=1480686639019, pause=100, retries=5}, org.apache.hadoop.hbase.MasterNotRunningException: java.io.IOException: Can't get master address from ZooKeeper; znode data == null
Fri Dec 02 19:20:40 IST 2016, RpcRetryingCaller{globalStartTime=1480686639019, pause=100, retries=5}, org.apache.hadoop.hbase.MasterNotRunningException: java.io.IOException: Can't get master address from ZooKeeper; znode data == null
Fri Dec 02 19:20:41 IST 2016, RpcRetryingCaller{globalStartTime=1480686639019, pause=100, retries=5}, org.apache.hadoop.hbase.MasterNotRunningException: java.io.IOException: Can't get master address from ZooKeeper; znode data == null

    at org.apache.hadoop.hbase.client.RpcRetryingCaller.callWithRetries(RpcRetryingCaller.java:147)
    at org.apache.hadoop.hbase.client.HBaseAdmin.executeCallable(HBaseAdmin.java:4083)
    at org.apache.hadoop.hbase.client.HBaseAdmin.getClusterStatus(HBaseAdmin.java:2765)
    at com.splicemachine.access.hbase.H10PartitionAdmin.allServers(H10PartitionAdmin.java:120)
    at com.splicemachine.hbase.HBaseRegionLoads.fetchRegionLoads(HBaseRegionLoads.java:164)
    at com.splicemachine.hbase.HBaseRegionLoads.access$000(HBaseRegionLoads.java:60)
    at com.splicemachine.hbase.HBaseRegionLoads$1.run(HBaseRegionLoads.java:81)
    at com.splicemachine.concurrent.LoggingScheduledThreadPoolExecutor$LoggingRunnable.run(LoggingScheduledThreadPoolExecutor.java:75)
    at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
    at java.util.concurrent.FutureTask.runAndReset(FutureTask.java:308)
    at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$301(ScheduledThreadPoolExecutor.java:180)
    at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:294)
    at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
    at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
    at java.lang.Thread.run(Thread.java:745)
Caused by: org.apache.hadoop.hbase.MasterNotRunningException: java.io.IOException: Can't get master address from ZooKeeper; znode data == null
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation$StubMaker.makeStub(ConnectionManager.java:1540)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation$MasterServiceStubMaker.makeStub(ConnectionManager.java:1560)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.getKeepAliveMasterService(ConnectionManager.java:1711)
    at org.apache.hadoop.hbase.client.MasterCallable.prepare(MasterCallable.java:38)
    at org.apache.hadoop.hbase.client.RpcRetryingCaller.callWithRetries(RpcRetryingCaller.java:124)
    ... 14 more
Caused by: java.io.IOException: Can't get master address from ZooKeeper; znode data == null
    at org.apache.hadoop.hbase.zookeeper.MasterAddressTracker.getMasterAddress(MasterAddressTracker.java:154)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation$StubMaker.makeStubNoRetries(ConnectionManager.java:1491)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation$StubMaker.makeStub(ConnectionManager.java:1531)
    ... 18 more
2016-12-02 19:20:44,021 WARN  [hbase-region-load-updater-0] client.ConnectionManager$HConnectionImplementation: Checking master connection
com.google.protobuf.ServiceException: java.net.ConnectException: Connection refused
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient.callBlockingMethod(AbstractRpcClient.java:223)
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient$BlockingRpcChannelImplementation.callBlockingMethod(AbstractRpcClient.java:287)
    at org.apache.hadoop.hbase.protobuf.generated.MasterProtos$MasterService$BlockingStub.isMasterRunning(MasterProtos.java:58152)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation$MasterServiceState.isMasterRunning(ConnectionManager.java:1444)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.isKeepAliveMasterConnectedAndRunning(ConnectionManager.java:2099)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.getKeepAliveMasterService(ConnectionManager.java:1708)
    at org.apache.hadoop.hbase.client.MasterCallable.prepare(MasterCallable.java:38)
    at org.apache.hadoop.hbase.client.RpcRetryingCaller.callWithRetries(RpcRetryingCaller.java:124)
    at org.apache.hadoop.hbase.client.HBaseAdmin.executeCallable(HBaseAdmin.java:4083)
    at org.apache.hadoop.hbase.client.HBaseAdmin.getClusterStatus(HBaseAdmin.java:2765)
    at com.splicemachine.access.hbase.H10PartitionAdmin.allServers(H10PartitionAdmin.java:120)
    at com.splicemachine.hbase.HBaseRegionLoads.fetchRegionLoads(HBaseRegionLoads.java:164)
    at com.splicemachine.hbase.HBaseRegionLoads.access$000(HBaseRegionLoads.java:60)
    at com.splicemachine.hbase.HBaseRegionLoads$1.run(HBaseRegionLoads.java:81)
    at com.splicemachine.concurrent.LoggingScheduledThreadPoolExecutor$LoggingRunnable.run(LoggingScheduledThreadPoolExecutor.java:75)
    at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
    at java.util.concurrent.FutureTask.runAndReset(FutureTask.java:308)
    at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$301(ScheduledThreadPoolExecutor.java:180)
    at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:294)
    at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
    at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
    at java.lang.Thread.run(Thread.java:745)
Caused by: java.net.ConnectException: Connection refused
    at sun.nio.ch.SocketChannelImpl.checkConnect(Native Method)
    at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:717)
    at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:206)
    at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:531)
    at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:495)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.setupConnection(RpcClientImpl.java:410)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.setupIOstreams(RpcClientImpl.java:716)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.writeRequest(RpcClientImpl.java:887)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.tracedWriteRequest(RpcClientImpl.java:856)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl.call(RpcClientImpl.java:1200)
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient.callBlockingMethod(AbstractRpcClient.java:213)
    ... 21 more
2016-12-02 19:20:44,226 WARN  [hbase-region-load-updater-0] client.ConnectionManager$HConnectionImplementation: Checking master connection
com.google.protobuf.ServiceException: org.apache.hadoop.hbase.ipc.FailedServerException: This server is in the failed servers list: hscale-dev1-nn/10.60.70.10:16000
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient.callBlockingMethod(AbstractRpcClient.java:223)
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient$BlockingRpcChannelImplementation.callBlockingMethod(AbstractRpcClient.java:287)
    at org.apache.hadoop.hbase.protobuf.generated.MasterProtos$MasterService$BlockingStub.isMasterRunning(MasterProtos.java:58152)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation$MasterServiceState.isMasterRunning(ConnectionManager.java:1444)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.isKeepAliveMasterConnectedAndRunning(ConnectionManager.java:2099)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.getKeepAliveMasterService(ConnectionManager.java:1708)
    at org.apache.hadoop.hbase.client.MasterCallable.prepare(MasterCallable.java:38)
    at org.apache.hadoop.hbase.client.RpcRetryingCaller.callWithRetries(RpcRetryingCaller.java:124)
    at org.apache.hadoop.hbase.client.HBaseAdmin.executeCallable(HBaseAdmin.java:4083)
    at org.apache.hadoop.hbase.client.HBaseAdmin.getClusterStatus(HBaseAdmin.java:2765)
    at com.splicemachine.access.hbase.H10PartitionAdmin.allServers(H10PartitionAdmin.java:120)
    at com.splicemachine.hbase.HBaseRegionLoads.fetchRegionLoads(HBaseRegionLoads.java:164)
    at com.splicemachine.hbase.HBaseRegionLoads.access$000(HBaseRegionLoads.java:60)
    at com.splicemachine.hbase.HBaseRegionLoads$1.run(HBaseRegionLoads.java:81)
    at com.splicemachine.concurrent.LoggingScheduledThreadPoolExecutor$LoggingRunnable.run(LoggingScheduledThreadPoolExecutor.java:75)
    at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
    at java.util.concurrent.FutureTask.runAndReset(FutureTask.java:308)
    at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$301(ScheduledThreadPoolExecutor.java:180)
    at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:294)
    at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
    at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
    at java.lang.Thread.run(Thread.java:745)
Caused by: org.apache.hadoop.hbase.ipc.FailedServerException: This server is in the failed servers list: hscale-dev1-nn/10.60.70.10:16000
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.setupIOstreams(RpcClientImpl.java:701)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.writeRequest(RpcClientImpl.java:887)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.tracedWriteRequest(RpcClientImpl.java:856)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl.call(RpcClientImpl.java:1200)
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient.callBlockingMethod(AbstractRpcClient.java:213)
    ... 21 more
2016-12-02 19:20:44,530 WARN  [hbase-region-load-updater-0] client.ConnectionManager$HConnectionImplementation: Checking master connection
com.google.protobuf.ServiceException: org.apache.hadoop.hbase.ipc.FailedServerException: This server is in the failed servers list: hscale-dev1-nn/10.60.70.10:16000
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient.callBlockingMethod(AbstractRpcClient.java:223)
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient$BlockingRpcChannelImplementation.callBlockingMethod(AbstractRpcClient.java:287)
    at org.apache.hadoop.hbase.protobuf.generated.MasterProtos$MasterService$BlockingStub.isMasterRunning(MasterProtos.java:58152)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation$MasterServiceState.isMasterRunning(ConnectionManager.java:1444)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.isKeepAliveMasterConnectedAndRunning(ConnectionManager.java:2099)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.getKeepAliveMasterService(ConnectionManager.java:1708)
    at org.apache.hadoop.hbase.client.MasterCallable.prepare(MasterCallable.java:38)
    at org.apache.hadoop.hbase.client.RpcRetryingCaller.callWithRetries(RpcRetryingCaller.java:124)
    at org.apache.hadoop.hbase.client.HBaseAdmin.executeCallable(HBaseAdmin.java:4083)
    at org.apache.hadoop.hbase.client.HBaseAdmin.getClusterStatus(HBaseAdmin.java:2765)
    at com.splicemachine.access.hbase.H10PartitionAdmin.allServers(H10PartitionAdmin.java:120)
    at com.splicemachine.hbase.HBaseRegionLoads.fetchRegionLoads(HBaseRegionLoads.java:164)
    at com.splicemachine.hbase.HBaseRegionLoads.access$000(HBaseRegionLoads.java:60)
    at com.splicemachine.hbase.HBaseRegionLoads$1.run(HBaseRegionLoads.java:81)
    at com.splicemachine.concurrent.LoggingScheduledThreadPoolExecutor$LoggingRunnable.run(LoggingScheduledThreadPoolExecutor.java:75)
    at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
    at java.util.concurrent.FutureTask.runAndReset(FutureTask.java:308)
    at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$301(ScheduledThreadPoolExecutor.java:180)
    at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:294)
    at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
    at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
    at java.lang.Thread.run(Thread.java:745)
Caused by: org.apache.hadoop.hbase.ipc.FailedServerException: This server is in the failed servers list: hscale-dev1-nn/10.60.70.10:16000
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.setupIOstreams(RpcClientImpl.java:701)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.writeRequest(RpcClientImpl.java:887)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.tracedWriteRequest(RpcClientImpl.java:856)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl.call(RpcClientImpl.java:1200)
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient.callBlockingMethod(AbstractRpcClient.java:213)
    ... 21 more
2016-12-02 19:20:45,034 WARN  [hbase-region-load-updater-0] client.ConnectionManager$HConnectionImplementation: Checking master connection
com.google.protobuf.ServiceException: org.apache.hadoop.hbase.ipc.FailedServerException: This server is in the failed servers list: hscale-dev1-nn/10.60.70.10:16000
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient.callBlockingMethod(AbstractRpcClient.java:223)
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient$BlockingRpcChannelImplementation.callBlockingMethod(AbstractRpcClient.java:287)
    at org.apache.hadoop.hbase.protobuf.generated.MasterProtos$MasterService$BlockingStub.isMasterRunning(MasterProtos.java:58152)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation$MasterServiceState.isMasterRunning(ConnectionManager.java:1444)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.isKeepAliveMasterConnectedAndRunning(ConnectionManager.java:2099)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.getKeepAliveMasterService(ConnectionManager.java:1708)
    at org.apache.hadoop.hbase.client.MasterCallable.prepare(MasterCallable.java:38)
    at org.apache.hadoop.hbase.client.RpcRetryingCaller.callWithRetries(RpcRetryingCaller.java:124)
    at org.apache.hadoop.hbase.client.HBaseAdmin.executeCallable(HBaseAdmin.java:4083)
    at org.apache.hadoop.hbase.client.HBaseAdmin.getClusterStatus(HBaseAdmin.java:2765)
    at com.splicemachine.access.hbase.H10PartitionAdmin.allServers(H10PartitionAdmin.java:120)
    at com.splicemachine.hbase.HBaseRegionLoads.fetchRegionLoads(HBaseRegionLoads.java:164)
    at com.splicemachine.hbase.HBaseRegionLoads.access$000(HBaseRegionLoads.java:60)
    at com.splicemachine.hbase.HBaseRegionLoads$1.run(HBaseRegionLoads.java:81)
    at com.splicemachine.concurrent.LoggingScheduledThreadPoolExecutor$LoggingRunnable.run(LoggingScheduledThreadPoolExecutor.java:75)
    at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
    at java.util.concurrent.FutureTask.runAndReset(FutureTask.java:308)
    at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$301(ScheduledThreadPoolExecutor.java:180)
    at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:294)
    at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
    at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
    at java.lang.Thread.run(Thread.java:745)
Caused by: org.apache.hadoop.hbase.ipc.FailedServerException: This server is in the failed servers list: hscale-dev1-nn/10.60.70.10:16000
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.setupIOstreams(RpcClientImpl.java:701)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.writeRequest(RpcClientImpl.java:887)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.tracedWriteRequest(RpcClientImpl.java:856)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl.call(RpcClientImpl.java:1200)
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient.callBlockingMethod(AbstractRpcClient.java:213)
    ... 21 more
2016-12-02 19:20:46,045 WARN  [hbase-region-load-updater-0] client.ConnectionManager$HConnectionImplementation: Checking master connection
com.google.protobuf.ServiceException: java.net.ConnectException: Connection refused
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient.callBlockingMethod(AbstractRpcClient.java:223)
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient$BlockingRpcChannelImplementation.callBlockingMethod(AbstractRpcClient.java:287)
    at org.apache.hadoop.hbase.protobuf.generated.MasterProtos$MasterService$BlockingStub.isMasterRunning(MasterProtos.java:58152)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation$MasterServiceState.isMasterRunning(ConnectionManager.java:1444)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.isKeepAliveMasterConnectedAndRunning(ConnectionManager.java:2099)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.getKeepAliveMasterService(ConnectionManager.java:1708)
    at org.apache.hadoop.hbase.client.MasterCallable.prepare(MasterCallable.java:38)
    at org.apache.hadoop.hbase.client.RpcRetryingCaller.callWithRetries(RpcRetryingCaller.java:124)
    at org.apache.hadoop.hbase.client.HBaseAdmin.executeCallable(HBaseAdmin.java:4083)
    at org.apache.hadoop.hbase.client.HBaseAdmin.getClusterStatus(HBaseAdmin.java:2765)
    at com.splicemachine.access.hbase.H10PartitionAdmin.allServers(H10PartitionAdmin.java:120)
    at com.splicemachine.hbase.HBaseRegionLoads.fetchRegionLoads(HBaseRegionLoads.java:164)
    at com.splicemachine.hbase.HBaseRegionLoads.access$000(HBaseRegionLoads.java:60)
    at com.splicemachine.hbase.HBaseRegionLoads$1.run(HBaseRegionLoads.java:81)
    at com.splicemachine.concurrent.LoggingScheduledThreadPoolExecutor$LoggingRunnable.run(LoggingScheduledThreadPoolExecutor.java:75)
    at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
    at java.util.concurrent.FutureTask.runAndReset(FutureTask.java:308)
    at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$301(ScheduledThreadPoolExecutor.java:180)
    at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:294)
    at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
    at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
    at java.lang.Thread.run(Thread.java:745)
Caused by: java.net.ConnectException: Connection refused
    at sun.nio.ch.SocketChannelImpl.checkConnect(Native Method)
    at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:717)
    at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:206)
    at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:531)
    at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:495)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.setupConnection(RpcClientImpl.java:410)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.setupIOstreams(RpcClientImpl.java:716)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.writeRequest(RpcClientImpl.java:887)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.tracedWriteRequest(RpcClientImpl.java:856)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl.call(RpcClientImpl.java:1200)
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient.callBlockingMethod(AbstractRpcClient.java:213)
    ... 21 more
2016-12-02 19:20:46,053 ERROR [hbase-region-load-updater-0] hbase.HBaseRegionLoads: Unable to fetch region load info
org.apache.hadoop.hbase.client.RetriesExhaustedException: Failed after attempts=5, exceptions:
Fri Dec 02 19:20:44 IST 2016, RpcRetryingCaller{globalStartTime=1480686644019, pause=100, retries=5}, org.apache.hadoop.hbase.MasterNotRunningException: java.io.IOException: Can't get master address from ZooKeeper; znode data == null
Fri Dec 02 19:20:44 IST 2016, RpcRetryingCaller{globalStartTime=1480686644019, pause=100, retries=5}, org.apache.hadoop.hbase.MasterNotRunningException: java.io.IOException: Can't get master address from ZooKeeper; znode data == null
Fri Dec 02 19:20:44 IST 2016, RpcRetryingCaller{globalStartTime=1480686644019, pause=100, retries=5}, org.apache.hadoop.hbase.MasterNotRunningException: java.io.IOException: Can't get master address from ZooKeeper; znode data == null
Fri Dec 02 19:20:45 IST 2016, RpcRetryingCaller{globalStartTime=1480686644019, pause=100, retries=5}, org.apache.hadoop.hbase.MasterNotRunningException: java.io.IOException: Can't get master address from ZooKeeper; znode data == null
Fri Dec 02 19:20:46 IST 2016, RpcRetryingCaller{globalStartTime=1480686644019, pause=100, retries=5}, org.apache.hadoop.hbase.MasterNotRunningException: java.io.IOException: Can't get master address from ZooKeeper; znode data == null

    at org.apache.hadoop.hbase.client.RpcRetryingCaller.callWithRetries(RpcRetryingCaller.java:147)
    at org.apache.hadoop.hbase.client.HBaseAdmin.executeCallable(HBaseAdmin.java:4083)
    at org.apache.hadoop.hbase.client.HBaseAdmin.getClusterStatus(HBaseAdmin.java:2765)
    at com.splicemachine.access.hbase.H10PartitionAdmin.allServers(H10PartitionAdmin.java:120)
    at com.splicemachine.hbase.HBaseRegionLoads.fetchRegionLoads(HBaseRegionLoads.java:164)
    at com.splicemachine.hbase.HBaseRegionLoads.access$000(HBaseRegionLoads.java:60)
    at com.splicemachine.hbase.HBaseRegionLoads$1.run(HBaseRegionLoads.java:81)
    at com.splicemachine.concurrent.LoggingScheduledThreadPoolExecutor$LoggingRunnable.run(LoggingScheduledThreadPoolExecutor.java:75)
    at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
    at java.util.concurrent.FutureTask.runAndReset(FutureTask.java:308)
    at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$301(ScheduledThreadPoolExecutor.java:180)
    at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:294)
    at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
    at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
    at java.lang.Thread.run(Thread.java:745)
Caused by: org.apache.hadoop.hbase.MasterNotRunningException: java.io.IOException: Can't get master address from ZooKeeper; znode data == null
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation$StubMaker.makeStub(ConnectionManager.java:1540)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation$MasterServiceStubMaker.makeStub(ConnectionManager.java:1560)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.getKeepAliveMasterService(ConnectionManager.java:1711)
    at org.apache.hadoop.hbase.client.MasterCallable.prepare(MasterCallable.java:38)
    at org.apache.hadoop.hbase.client.RpcRetryingCaller.callWithRetries(RpcRetryingCaller.java:124)
    ... 14 more
Caused by: java.io.IOException: Can't get master address from ZooKeeper; znode data == null
    at org.apache.hadoop.hbase.zookeeper.MasterAddressTracker.getMasterAddress(MasterAddressTracker.java:154)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation$StubMaker.makeStubNoRetries(ConnectionManager.java:1491)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation$StubMaker.makeStub(ConnectionManager.java:1531)
    ... 18 more
2016-12-02 19:20:49,020 WARN  [hbase-region-load-updater-0] client.ConnectionManager$HConnectionImplementation: Checking master connection
com.google.protobuf.ServiceException: java.net.ConnectException: Connection refused
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient.callBlockingMethod(AbstractRpcClient.java:223)
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient$BlockingRpcChannelImplementation.callBlockingMethod(AbstractRpcClient.java:287)
    at org.apache.hadoop.hbase.protobuf.generated.MasterProtos$MasterService$BlockingStub.isMasterRunning(MasterProtos.java:58152)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation$MasterServiceState.isMasterRunning(ConnectionManager.java:1444)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.isKeepAliveMasterConnectedAndRunning(ConnectionManager.java:2099)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.getKeepAliveMasterService(ConnectionManager.java:1708)
    at org.apache.hadoop.hbase.client.MasterCallable.prepare(MasterCallable.java:38)
    at org.apache.hadoop.hbase.client.RpcRetryingCaller.callWithRetries(RpcRetryingCaller.java:124)
    at org.apache.hadoop.hbase.client.HBaseAdmin.executeCallable(HBaseAdmin.java:4083)
    at org.apache.hadoop.hbase.client.HBaseAdmin.getClusterStatus(HBaseAdmin.java:2765)
    at com.splicemachine.access.hbase.H10PartitionAdmin.allServers(H10PartitionAdmin.java:120)
    at com.splicemachine.hbase.HBaseRegionLoads.fetchRegionLoads(HBaseRegionLoads.java:164)
    at com.splicemachine.hbase.HBaseRegionLoads.access$000(HBaseRegionLoads.java:60)
    at com.splicemachine.hbase.HBaseRegionLoads$1.run(HBaseRegionLoads.java:81)
    at com.splicemachine.concurrent.LoggingScheduledThreadPoolExecutor$LoggingRunnable.run(LoggingScheduledThreadPoolExecutor.java:75)
    at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
    at java.util.concurrent.FutureTask.runAndReset(FutureTask.java:308)
    at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$301(ScheduledThreadPoolExecutor.java:180)
    at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:294)
    at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
    at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
    at java.lang.Thread.run(Thread.java:745)
Caused by: java.net.ConnectException: Connection refused
    at sun.nio.ch.SocketChannelImpl.checkConnect(Native Method)
    at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:717)
    at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:206)
    at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:531)
    at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:495)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.setupConnection(RpcClientImpl.java:410)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.setupIOstreams(RpcClientImpl.java:716)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.writeRequest(RpcClientImpl.java:887)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.tracedWriteRequest(RpcClientImpl.java:856)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl.call(RpcClientImpl.java:1200)
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient.callBlockingMethod(AbstractRpcClient.java:213)
    ... 21 more
2016-12-02 19:20:49,225 WARN  [hbase-region-load-updater-0] client.ConnectionManager$HConnectionImplementation: Checking master connection
com.google.protobuf.ServiceException: org.apache.hadoop.hbase.ipc.FailedServerException: This server is in the failed servers list: hscale-dev1-nn/10.60.70.10:16000
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient.callBlockingMethod(AbstractRpcClient.java:223)
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient$BlockingRpcChannelImplementation.callBlockingMethod(AbstractRpcClient.java:287)
    at org.apache.hadoop.hbase.protobuf.generated.MasterProtos$MasterService$BlockingStub.isMasterRunning(MasterProtos.java:58152)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation$MasterServiceState.isMasterRunning(ConnectionManager.java:1444)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.isKeepAliveMasterConnectedAndRunning(ConnectionManager.java:2099)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.getKeepAliveMasterService(ConnectionManager.java:1708)
    at org.apache.hadoop.hbase.client.MasterCallable.prepare(MasterCallable.java:38)
    at org.apache.hadoop.hbase.client.RpcRetryingCaller.callWithRetries(RpcRetryingCaller.java:124)
    at org.apache.hadoop.hbase.client.HBaseAdmin.executeCallable(HBaseAdmin.java:4083)
    at org.apache.hadoop.hbase.client.HBaseAdmin.getClusterStatus(HBaseAdmin.java:2765)
    at com.splicemachine.access.hbase.H10PartitionAdmin.allServers(H10PartitionAdmin.java:120)
    at com.splicemachine.hbase.HBaseRegionLoads.fetchRegionLoads(HBaseRegionLoads.java:164)
    at com.splicemachine.hbase.HBaseRegionLoads.access$000(HBaseRegionLoads.java:60)
    at com.splicemachine.hbase.HBaseRegionLoads$1.run(HBaseRegionLoads.java:81)
    at com.splicemachine.concurrent.LoggingScheduledThreadPoolExecutor$LoggingRunnable.run(LoggingScheduledThreadPoolExecutor.java:75)
    at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
    at java.util.concurrent.FutureTask.runAndReset(FutureTask.java:308)
    at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$301(ScheduledThreadPoolExecutor.java:180)
    at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:294)
    at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
    at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
    at java.lang.Thread.run(Thread.java:745)
Caused by: org.apache.hadoop.hbase.ipc.FailedServerException: This server is in the failed servers list: hscale-dev1-nn/10.60.70.10:16000
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.setupIOstreams(RpcClientImpl.java:701)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.writeRequest(RpcClientImpl.java:887)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.tracedWriteRequest(RpcClientImpl.java:856)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl.call(RpcClientImpl.java:1200)
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient.callBlockingMethod(AbstractRpcClient.java:213)
    ... 21 more
2016-12-02 19:20:49,530 WARN  [hbase-region-load-updater-0] client.ConnectionManager$HConnectionImplementation: Checking master connection
com.google.protobuf.ServiceException: org.apache.hadoop.hbase.ipc.FailedServerException: This server is in the failed servers list: hscale-dev1-nn/10.60.70.10:16000
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient.callBlockingMethod(AbstractRpcClient.java:223)
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient$BlockingRpcChannelImplementation.callBlockingMethod(AbstractRpcClient.java:287)
    at org.apache.hadoop.hbase.protobuf.generated.MasterProtos$MasterService$BlockingStub.isMasterRunning(MasterProtos.java:58152)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation$MasterServiceState.isMasterRunning(ConnectionManager.java:1444)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.isKeepAliveMasterConnectedAndRunning(ConnectionManager.java:2099)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.getKeepAliveMasterService(ConnectionManager.java:1708)
    at org.apache.hadoop.hbase.client.MasterCallable.prepare(MasterCallable.java:38)
    at org.apache.hadoop.hbase.client.RpcRetryingCaller.callWithRetries(RpcRetryingCaller.java:124)
    at org.apache.hadoop.hbase.client.HBaseAdmin.executeCallable(HBaseAdmin.java:4083)
    at org.apache.hadoop.hbase.client.HBaseAdmin.getClusterStatus(HBaseAdmin.java:2765)
    at com.splicemachine.access.hbase.H10PartitionAdmin.allServers(H10PartitionAdmin.java:120)
    at com.splicemachine.hbase.HBaseRegionLoads.fetchRegionLoads(HBaseRegionLoads.java:164)
    at com.splicemachine.hbase.HBaseRegionLoads.access$000(HBaseRegionLoads.java:60)
    at com.splicemachine.hbase.HBaseRegionLoads$1.run(HBaseRegionLoads.java:81)
    at com.splicemachine.concurrent.LoggingScheduledThreadPoolExecutor$LoggingRunnable.run(LoggingScheduledThreadPoolExecutor.java:75)
    at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
    at java.util.concurrent.FutureTask.runAndReset(FutureTask.java:308)
    at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$301(ScheduledThreadPoolExecutor.java:180)
    at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:294)
    at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
    at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
    at java.lang.Thread.run(Thread.java:745)
Caused by: org.apache.hadoop.hbase.ipc.FailedServerException: This server is in the failed servers list: hscale-dev1-nn/10.60.70.10:16000
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.setupIOstreams(RpcClientImpl.java:701)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.writeRequest(RpcClientImpl.java:887)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.tracedWriteRequest(RpcClientImpl.java:856)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl.call(RpcClientImpl.java:1200)
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient.callBlockingMethod(AbstractRpcClient.java:213)
    ... 21 more
2016-12-02 19:20:50,037 WARN  [hbase-region-load-updater-0] client.ConnectionManager$HConnectionImplementation: Checking master connection
com.google.protobuf.ServiceException: org.apache.hadoop.hbase.ipc.FailedServerException: This server is in the failed servers list: hscale-dev1-nn/10.60.70.10:16000
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient.callBlockingMethod(AbstractRpcClient.java:223)
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient$BlockingRpcChannelImplementation.callBlockingMethod(AbstractRpcClient.java:287)
    at org.apache.hadoop.hbase.protobuf.generated.MasterProtos$MasterService$BlockingStub.isMasterRunning(MasterProtos.java:58152)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation$MasterServiceState.isMasterRunning(ConnectionManager.java:1444)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.isKeepAliveMasterConnectedAndRunning(ConnectionManager.java:2099)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.getKeepAliveMasterService(ConnectionManager.java:1708)
    at org.apache.hadoop.hbase.client.MasterCallable.prepare(MasterCallable.java:38)
    at org.apache.hadoop.hbase.client.RpcRetryingCaller.callWithRetries(RpcRetryingCaller.java:124)
    at org.apache.hadoop.hbase.client.HBaseAdmin.executeCallable(HBaseAdmin.java:4083)
    at org.apache.hadoop.hbase.client.HBaseAdmin.getClusterStatus(HBaseAdmin.java:2765)
    at com.splicemachine.access.hbase.H10PartitionAdmin.allServers(H10PartitionAdmin.java:120)
    at com.splicemachine.hbase.HBaseRegionLoads.fetchRegionLoads(HBaseRegionLoads.java:164)
    at com.splicemachine.hbase.HBaseRegionLoads.access$000(HBaseRegionLoads.java:60)
    at com.splicemachine.hbase.HBaseRegionLoads$1.run(HBaseRegionLoads.java:81)
    at com.splicemachine.concurrent.LoggingScheduledThreadPoolExecutor$LoggingRunnable.run(LoggingScheduledThreadPoolExecutor.java:75)
    at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
    at java.util.concurrent.FutureTask.runAndReset(FutureTask.java:308)
    at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$301(ScheduledThreadPoolExecutor.java:180)
    at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:294)
    at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
    at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
    at java.lang.Thread.run(Thread.java:745)
Caused by: org.apache.hadoop.hbase.ipc.FailedServerException: This server is in the failed servers list: hscale-dev1-nn/10.60.70.10:16000
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.setupIOstreams(RpcClientImpl.java:701)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.writeRequest(RpcClientImpl.java:887)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.tracedWriteRequest(RpcClientImpl.java:856)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl.call(RpcClientImpl.java:1200)
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient.callBlockingMethod(AbstractRpcClient.java:213)
    ... 21 more
2016-12-02 19:20:50,047 WARN  [pool-10-thread-1] client.ConnectionManager$HConnectionImplementation: Checking master connection
com.google.protobuf.ServiceException: org.apache.hadoop.hbase.ipc.FailedServerException: This server is in the failed servers list: hscale-dev1-nn/10.60.70.10:16000
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient.callBlockingMethod(AbstractRpcClient.java:223)
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient$BlockingRpcChannelImplementation.callBlockingMethod(AbstractRpcClient.java:287)
    at org.apache.hadoop.hbase.protobuf.generated.MasterProtos$MasterService$BlockingStub.isMasterRunning(MasterProtos.java:58152)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation$MasterServiceState.isMasterRunning(ConnectionManager.java:1444)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.isKeepAliveMasterConnectedAndRunning(ConnectionManager.java:2099)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.getKeepAliveMasterService(ConnectionManager.java:1708)
    at org.apache.hadoop.hbase.client.MasterCallable.prepare(MasterCallable.java:38)
    at org.apache.hadoop.hbase.client.RpcRetryingCaller.callWithRetries(RpcRetryingCaller.java:124)
    at org.apache.hadoop.hbase.client.HBaseAdmin.executeCallable(HBaseAdmin.java:4083)
    at org.apache.hadoop.hbase.client.HBaseAdmin.access$700(HBaseAdmin.java:194)
    at org.apache.hadoop.hbase.client.HBaseAdmin$ProcedureFuture.getProcedureResult(HBaseAdmin.java:4383)
    at org.apache.hadoop.hbase.client.HBaseAdmin$ProcedureFuture.waitProcedureResult(HBaseAdmin.java:4335)
    at org.apache.hadoop.hbase.client.HBaseAdmin$ProcedureFuture.get(HBaseAdmin.java:4291)
    at org.apache.hadoop.hbase.client.HBaseAdmin.createTable(HBaseAdmin.java:647)
    at org.apache.hadoop.hbase.client.HBaseAdmin.createTable(HBaseAdmin.java:577)
    at com.splicemachine.lifecycle.RegionServerLifecycle.distributedStart(RegionServerLifecycle.java:66)
    at com.splicemachine.derby.lifecycle.EngineLifecycleService.start(EngineLifecycleService.java:81)
    at com.splicemachine.lifecycle.DatabaseLifecycleManager$Startup.bootServices(DatabaseLifecycleManager.java:229)
    at com.splicemachine.lifecycle.DatabaseLifecycleManager$Startup.run(DatabaseLifecycleManager.java:220)
    at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
    at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
    at java.lang.Thread.run(Thread.java:745)
Caused by: org.apache.hadoop.hbase.ipc.FailedServerException: This server is in the failed servers list: hscale-dev1-nn/10.60.70.10:16000
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.setupIOstreams(RpcClientImpl.java:701)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.writeRequest(RpcClientImpl.java:887)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.tracedWriteRequest(RpcClientImpl.java:856)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl.call(RpcClientImpl.java:1200)
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient.callBlockingMethod(AbstractRpcClient.java:213)
    ... 21 more
2016-12-02 19:20:50,252 WARN  [pool-10-thread-1] client.ConnectionManager$HConnectionImplementation: Checking master connection
com.google.protobuf.ServiceException: org.apache.hadoop.hbase.ipc.FailedServerException: This server is in the failed servers list: hscale-dev1-nn/10.60.70.10:16000
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient.callBlockingMethod(AbstractRpcClient.java:223)
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient$BlockingRpcChannelImplementation.callBlockingMethod(AbstractRpcClient.java:287)
    at org.apache.hadoop.hbase.protobuf.generated.MasterProtos$MasterService$BlockingStub.isMasterRunning(MasterProtos.java:58152)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation$MasterServiceState.isMasterRunning(ConnectionManager.java:1444)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.isKeepAliveMasterConnectedAndRunning(ConnectionManager.java:2099)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.getKeepAliveMasterService(ConnectionManager.java:1708)
    at org.apache.hadoop.hbase.client.MasterCallable.prepare(MasterCallable.java:38)
    at org.apache.hadoop.hbase.client.RpcRetryingCaller.callWithRetries(RpcRetryingCaller.java:124)
    at org.apache.hadoop.hbase.client.HBaseAdmin.executeCallable(HBaseAdmin.java:4083)
    at org.apache.hadoop.hbase.client.HBaseAdmin.access$700(HBaseAdmin.java:194)
    at org.apache.hadoop.hbase.client.HBaseAdmin$ProcedureFuture.getProcedureResult(HBaseAdmin.java:4383)
    at org.apache.hadoop.hbase.client.HBaseAdmin$ProcedureFuture.waitProcedureResult(HBaseAdmin.java:4335)
    at org.apache.hadoop.hbase.client.HBaseAdmin$ProcedureFuture.get(HBaseAdmin.java:4291)
    at org.apache.hadoop.hbase.client.HBaseAdmin.createTable(HBaseAdmin.java:647)
    at org.apache.hadoop.hbase.client.HBaseAdmin.createTable(HBaseAdmin.java:577)
    at com.splicemachine.lifecycle.RegionServerLifecycle.distributedStart(RegionServerLifecycle.java:66)
    at com.splicemachine.derby.lifecycle.EngineLifecycleService.start(EngineLifecycleService.java:81)
    at com.splicemachine.lifecycle.DatabaseLifecycleManager$Startup.bootServices(DatabaseLifecycleManager.java:229)
    at com.splicemachine.lifecycle.DatabaseLifecycleManager$Startup.run(DatabaseLifecycleManager.java:220)
    at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
    at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
    at java.lang.Thread.run(Thread.java:745)
Caused by: org.apache.hadoop.hbase.ipc.FailedServerException: This server is in the failed servers list: hscale-dev1-nn/10.60.70.10:16000
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.setupIOstreams(RpcClientImpl.java:701)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.writeRequest(RpcClientImpl.java:887)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.tracedWriteRequest(RpcClientImpl.java:856)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl.call(RpcClientImpl.java:1200)
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient.callBlockingMethod(AbstractRpcClient.java:213)
    ... 21 more
2016-12-02 19:20:50,558 WARN  [pool-10-thread-1] client.ConnectionManager$HConnectionImplementation: Checking master connection
com.google.protobuf.ServiceException: org.apache.hadoop.hbase.ipc.FailedServerException: This server is in the failed servers list: hscale-dev1-nn/10.60.70.10:16000
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient.callBlockingMethod(AbstractRpcClient.java:223)
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient$BlockingRpcChannelImplementation.callBlockingMethod(AbstractRpcClient.java:287)
    at org.apache.hadoop.hbase.protobuf.generated.MasterProtos$MasterService$BlockingStub.isMasterRunning(MasterProtos.java:58152)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation$MasterServiceState.isMasterRunning(ConnectionManager.java:1444)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.isKeepAliveMasterConnectedAndRunning(ConnectionManager.java:2099)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.getKeepAliveMasterService(ConnectionManager.java:1708)
    at org.apache.hadoop.hbase.client.MasterCallable.prepare(MasterCallable.java:38)
    at org.apache.hadoop.hbase.client.RpcRetryingCaller.callWithRetries(RpcRetryingCaller.java:124)
    at org.apache.hadoop.hbase.client.HBaseAdmin.executeCallable(HBaseAdmin.java:4083)
    at org.apache.hadoop.hbase.client.HBaseAdmin.access$700(HBaseAdmin.java:194)
    at org.apache.hadoop.hbase.client.HBaseAdmin$ProcedureFuture.getProcedureResult(HBaseAdmin.java:4383)
    at org.apache.hadoop.hbase.client.HBaseAdmin$ProcedureFuture.waitProcedureResult(HBaseAdmin.java:4335)
    at org.apache.hadoop.hbase.client.HBaseAdmin$ProcedureFuture.get(HBaseAdmin.java:4291)
    at org.apache.hadoop.hbase.client.HBaseAdmin.createTable(HBaseAdmin.java:647)
    at org.apache.hadoop.hbase.client.HBaseAdmin.createTable(HBaseAdmin.java:577)
    at com.splicemachine.lifecycle.RegionServerLifecycle.distributedStart(RegionServerLifecycle.java:66)
    at com.splicemachine.derby.lifecycle.EngineLifecycleService.start(EngineLifecycleService.java:81)
    at com.splicemachine.lifecycle.DatabaseLifecycleManager$Startup.bootServices(DatabaseLifecycleManager.java:229)
    at com.splicemachine.lifecycle.DatabaseLifecycleManager$Startup.run(DatabaseLifecycleManager.java:220)
    at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
    at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
    at java.lang.Thread.run(Thread.java:745)
Caused by: org.apache.hadoop.hbase.ipc.FailedServerException: This server is in the failed servers list: hscale-dev1-nn/10.60.70.10:16000
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.setupIOstreams(RpcClientImpl.java:701)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.writeRequest(RpcClientImpl.java:887)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.tracedWriteRequest(RpcClientImpl.java:856)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl.call(RpcClientImpl.java:1200)
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient.callBlockingMethod(AbstractRpcClient.java:213)
    ... 21 more
2016-12-02 19:20:51,050 WARN  [hbase-region-load-updater-0] client.ConnectionManager$HConnectionImplementation: Checking master connection
com.google.protobuf.ServiceException: java.net.ConnectException: Connection refused
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient.callBlockingMethod(AbstractRpcClient.java:223)
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient$BlockingRpcChannelImplementation.callBlockingMethod(AbstractRpcClient.java:287)
    at org.apache.hadoop.hbase.protobuf.generated.MasterProtos$MasterService$BlockingStub.isMasterRunning(MasterProtos.java:58152)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation$MasterServiceState.isMasterRunning(ConnectionManager.java:1444)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.isKeepAliveMasterConnectedAndRunning(ConnectionManager.java:2099)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.getKeepAliveMasterService(ConnectionManager.java:1708)
    at org.apache.hadoop.hbase.client.MasterCallable.prepare(MasterCallable.java:38)
    at org.apache.hadoop.hbase.client.RpcRetryingCaller.callWithRetries(RpcRetryingCaller.java:124)
    at org.apache.hadoop.hbase.client.HBaseAdmin.executeCallable(HBaseAdmin.java:4083)
    at org.apache.hadoop.hbase.client.HBaseAdmin.getClusterStatus(HBaseAdmin.java:2765)
    at com.splicemachine.access.hbase.H10PartitionAdmin.allServers(H10PartitionAdmin.java:120)
    at com.splicemachine.hbase.HBaseRegionLoads.fetchRegionLoads(HBaseRegionLoads.java:164)
    at com.splicemachine.hbase.HBaseRegionLoads.access$000(HBaseRegionLoads.java:60)
    at com.splicemachine.hbase.HBaseRegionLoads$1.run(HBaseRegionLoads.java:81)
    at com.splicemachine.concurrent.LoggingScheduledThreadPoolExecutor$LoggingRunnable.run(LoggingScheduledThreadPoolExecutor.java:75)
    at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
    at java.util.concurrent.FutureTask.runAndReset(FutureTask.java:308)
    at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$301(ScheduledThreadPoolExecutor.java:180)
    at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:294)
    at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
    at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
    at java.lang.Thread.run(Thread.java:745)
Caused by: java.net.ConnectException: Connection refused
    at sun.nio.ch.SocketChannelImpl.checkConnect(Native Method)
    at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:717)
    at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:206)
    at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:531)
    at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:495)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.setupConnection(RpcClientImpl.java:410)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.setupIOstreams(RpcClientImpl.java:716)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.writeRequest(RpcClientImpl.java:887)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.tracedWriteRequest(RpcClientImpl.java:856)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl.call(RpcClientImpl.java:1200)
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient.callBlockingMethod(AbstractRpcClient.java:213)
    ... 21 more
2016-12-02 19:20:51,054 ERROR [hbase-region-load-updater-0] hbase.HBaseRegionLoads: Unable to fetch region load info
org.apache.hadoop.hbase.client.RetriesExhaustedException: Failed after attempts=5, exceptions:
Fri Dec 02 19:20:49 IST 2016, RpcRetryingCaller{globalStartTime=1480686649019, pause=100, retries=5}, org.apache.hadoop.hbase.MasterNotRunningException: java.io.IOException: Can't get master address from ZooKeeper; znode data == null
Fri Dec 02 19:20:49 IST 2016, RpcRetryingCaller{globalStartTime=1480686649019, pause=100, retries=5}, org.apache.hadoop.hbase.MasterNotRunningException: java.io.IOException: Can't get master address from ZooKeeper; znode data == null
Fri Dec 02 19:20:49 IST 2016, RpcRetryingCaller{globalStartTime=1480686649019, pause=100, retries=5}, org.apache.hadoop.hbase.MasterNotRunningException: java.io.IOException: Can't get master address from ZooKeeper; znode data == null
Fri Dec 02 19:20:50 IST 2016, RpcRetryingCaller{globalStartTime=1480686649019, pause=100, retries=5}, org.apache.hadoop.hbase.MasterNotRunningException: java.io.IOException: Can't get master address from ZooKeeper; znode data == null
Fri Dec 02 19:20:51 IST 2016, RpcRetryingCaller{globalStartTime=1480686649019, pause=100, retries=5}, org.apache.hadoop.hbase.MasterNotRunningException: java.io.IOException: Can't get master address from ZooKeeper; znode data == null

    at org.apache.hadoop.hbase.client.RpcRetryingCaller.callWithRetries(RpcRetryingCaller.java:147)
    at org.apache.hadoop.hbase.client.HBaseAdmin.executeCallable(HBaseAdmin.java:4083)
    at org.apache.hadoop.hbase.client.HBaseAdmin.getClusterStatus(HBaseAdmin.java:2765)
    at com.splicemachine.access.hbase.H10PartitionAdmin.allServers(H10PartitionAdmin.java:120)
    at com.splicemachine.hbase.HBaseRegionLoads.fetchRegionLoads(HBaseRegionLoads.java:164)
    at com.splicemachine.hbase.HBaseRegionLoads.access$000(HBaseRegionLoads.java:60)
    at com.splicemachine.hbase.HBaseRegionLoads$1.run(HBaseRegionLoads.java:81)
    at com.splicemachine.concurrent.LoggingScheduledThreadPoolExecutor$LoggingRunnable.run(LoggingScheduledThreadPoolExecutor.java:75)
    at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
    at java.util.concurrent.FutureTask.runAndReset(FutureTask.java:308)
    at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$301(ScheduledThreadPoolExecutor.java:180)
    at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:294)
    at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
    at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
    at java.lang.Thread.run(Thread.java:745)
Caused by: org.apache.hadoop.hbase.MasterNotRunningException: java.io.IOException: Can't get master address from ZooKeeper; znode data == null
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation$StubMaker.makeStub(ConnectionManager.java:1540)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation$MasterServiceStubMaker.makeStub(ConnectionManager.java:1560)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.getKeepAliveMasterService(ConnectionManager.java:1711)
    at org.apache.hadoop.hbase.client.MasterCallable.prepare(MasterCallable.java:38)
    at org.apache.hadoop.hbase.client.RpcRetryingCaller.callWithRetries(RpcRetryingCaller.java:124)
    ... 14 more
Caused by: java.io.IOException: Can't get master address from ZooKeeper; znode data == null
    at org.apache.hadoop.hbase.zookeeper.MasterAddressTracker.getMasterAddress(MasterAddressTracker.java:154)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation$StubMaker.makeStubNoRetries(ConnectionManager.java:1491)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation$StubMaker.makeStub(ConnectionManager.java:1531)
    ... 18 more
2016-12-02 19:20:51,066 WARN  [pool-10-thread-1] client.ConnectionManager$HConnectionImplementation: Checking master connection
com.google.protobuf.ServiceException: org.apache.hadoop.hbase.ipc.FailedServerException: This server is in the failed servers list: hscale-dev1-nn/10.60.70.10:16000
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient.callBlockingMethod(AbstractRpcClient.java:223)
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient$BlockingRpcChannelImplementation.callBlockingMethod(AbstractRpcClient.java:287)
    at org.apache.hadoop.hbase.protobuf.generated.MasterProtos$MasterService$BlockingStub.isMasterRunning(MasterProtos.java:58152)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation$MasterServiceState.isMasterRunning(ConnectionManager.java:1444)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.isKeepAliveMasterConnectedAndRunning(ConnectionManager.java:2099)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.getKeepAliveMasterService(ConnectionManager.java:1708)
    at org.apache.hadoop.hbase.client.MasterCallable.prepare(MasterCallable.java:38)
    at org.apache.hadoop.hbase.client.RpcRetryingCaller.callWithRetries(RpcRetryingCaller.java:124)
    at org.apache.hadoop.hbase.client.HBaseAdmin.executeCallable(HBaseAdmin.java:4083)
    at org.apache.hadoop.hbase.client.HBaseAdmin.access$700(HBaseAdmin.java:194)
    at org.apache.hadoop.hbase.client.HBaseAdmin$ProcedureFuture.getProcedureResult(HBaseAdmin.java:4383)
    at org.apache.hadoop.hbase.client.HBaseAdmin$ProcedureFuture.waitProcedureResult(HBaseAdmin.java:4335)
    at org.apache.hadoop.hbase.client.HBaseAdmin$ProcedureFuture.get(HBaseAdmin.java:4291)
    at org.apache.hadoop.hbase.client.HBaseAdmin.createTable(HBaseAdmin.java:647)
    at org.apache.hadoop.hbase.client.HBaseAdmin.createTable(HBaseAdmin.java:577)
    at com.splicemachine.lifecycle.RegionServerLifecycle.distributedStart(RegionServerLifecycle.java:66)
    at com.splicemachine.derby.lifecycle.EngineLifecycleService.start(EngineLifecycleService.java:81)
    at com.splicemachine.lifecycle.DatabaseLifecycleManager$Startup.bootServices(DatabaseLifecycleManager.java:229)
    at com.splicemachine.lifecycle.DatabaseLifecycleManager$Startup.run(DatabaseLifecycleManager.java:220)
    at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
    at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
    at java.lang.Thread.run(Thread.java:745)
Caused by: org.apache.hadoop.hbase.ipc.FailedServerException: This server is in the failed servers list: hscale-dev1-nn/10.60.70.10:16000
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.setupIOstreams(RpcClientImpl.java:701)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.writeRequest(RpcClientImpl.java:887)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.tracedWriteRequest(RpcClientImpl.java:856)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl.call(RpcClientImpl.java:1200)
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient.callBlockingMethod(AbstractRpcClient.java:213)
    ... 21 more
2016-12-02 19:20:52,074 WARN  [pool-10-thread-1] client.ConnectionManager$HConnectionImplementation: Checking master connection
com.google.protobuf.ServiceException: org.apache.hadoop.hbase.ipc.FailedServerException: This server is in the failed servers list: hscale-dev1-nn/10.60.70.10:16000
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient.callBlockingMethod(AbstractRpcClient.java:223)
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient$BlockingRpcChannelImplementation.callBlockingMethod(AbstractRpcClient.java:287)
    at org.apache.hadoop.hbase.protobuf.generated.MasterProtos$MasterService$BlockingStub.isMasterRunning(MasterProtos.java:58152)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation$MasterServiceState.isMasterRunning(ConnectionManager.java:1444)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.isKeepAliveMasterConnectedAndRunning(ConnectionManager.java:2099)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.getKeepAliveMasterService(ConnectionManager.java:1708)
    at org.apache.hadoop.hbase.client.MasterCallable.prepare(MasterCallable.java:38)
    at org.apache.hadoop.hbase.client.RpcRetryingCaller.callWithRetries(RpcRetryingCaller.java:124)
    at org.apache.hadoop.hbase.client.HBaseAdmin.executeCallable(HBaseAdmin.java:4083)
    at org.apache.hadoop.hbase.client.HBaseAdmin.access$700(HBaseAdmin.java:194)
    at org.apache.hadoop.hbase.client.HBaseAdmin$ProcedureFuture.getProcedureResult(HBaseAdmin.java:4383)
    at org.apache.hadoop.hbase.client.HBaseAdmin$ProcedureFuture.waitProcedureResult(HBaseAdmin.java:4335)
    at org.apache.hadoop.hbase.client.HBaseAdmin$ProcedureFuture.get(HBaseAdmin.java:4291)
    at org.apache.hadoop.hbase.client.HBaseAdmin.createTable(HBaseAdmin.java:647)
    at org.apache.hadoop.hbase.client.HBaseAdmin.createTable(HBaseAdmin.java:577)
    at com.splicemachine.lifecycle.RegionServerLifecycle.distributedStart(RegionServerLifecycle.java:66)
    at com.splicemachine.derby.lifecycle.EngineLifecycleService.start(EngineLifecycleService.java:81)
    at com.splicemachine.lifecycle.DatabaseLifecycleManager$Startup.bootServices(DatabaseLifecycleManager.java:229)
    at com.splicemachine.lifecycle.DatabaseLifecycleManager$Startup.run(DatabaseLifecycleManager.java:220)
    at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
    at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
    at java.lang.Thread.run(Thread.java:745)
Caused by: org.apache.hadoop.hbase.ipc.FailedServerException: This server is in the failed servers list: hscale-dev1-nn/10.60.70.10:16000
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.setupIOstreams(RpcClientImpl.java:701)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.writeRequest(RpcClientImpl.java:887)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.tracedWriteRequest(RpcClientImpl.java:856)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl.call(RpcClientImpl.java:1200)
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient.callBlockingMethod(AbstractRpcClient.java:213)
    ... 21 more
2016-12-02 19:20:52,078 WARN  [pool-10-thread-1] client.HBaseAdmin: failed to get the procedure result procId=6
org.apache.hadoop.hbase.client.RetriesExhaustedException: Failed after attempts=5, exceptions:
Fri Dec 02 19:20:50 IST 2016, RpcRetryingCaller{globalStartTime=1480686650047, pause=100, retries=5}, org.apache.hadoop.hbase.MasterNotRunningException: java.io.IOException: Can't get master address from ZooKeeper; znode data == null
Fri Dec 02 19:20:50 IST 2016, RpcRetryingCaller{globalStartTime=1480686650047, pause=100, retries=5}, org.apache.hadoop.hbase.MasterNotRunningException: java.io.IOException: Can't get master address from ZooKeeper; znode data == null
Fri Dec 02 19:20:50 IST 2016, RpcRetryingCaller{globalStartTime=1480686650047, pause=100, retries=5}, org.apache.hadoop.hbase.MasterNotRunningException: java.io.IOException: Can't get master address from ZooKeeper; znode data == null
Fri Dec 02 19:20:51 IST 2016, RpcRetryingCaller{globalStartTime=1480686650047, pause=100, retries=5}, org.apache.hadoop.hbase.MasterNotRunningException: java.io.IOException: Can't get master address from ZooKeeper; znode data == null
Fri Dec 02 19:20:52 IST 2016, RpcRetryingCaller{globalStartTime=1480686650047, pause=100, retries=5}, org.apache.hadoop.hbase.MasterNotRunningException: java.io.IOException: Can't get master address from ZooKeeper; znode data == null

    at org.apache.hadoop.hbase.client.RpcRetryingCaller.callWithRetries(RpcRetryingCaller.java:147)
    at org.apache.hadoop.hbase.client.HBaseAdmin.executeCallable(HBaseAdmin.java:4083)
    at org.apache.hadoop.hbase.client.HBaseAdmin.access$700(HBaseAdmin.java:194)
    at org.apache.hadoop.hbase.client.HBaseAdmin$ProcedureFuture.getProcedureResult(HBaseAdmin.java:4383)
    at org.apache.hadoop.hbase.client.HBaseAdmin$ProcedureFuture.waitProcedureResult(HBaseAdmin.java:4335)
    at org.apache.hadoop.hbase.client.HBaseAdmin$ProcedureFuture.get(HBaseAdmin.java:4291)
    at org.apache.hadoop.hbase.client.HBaseAdmin.createTable(HBaseAdmin.java:647)
    at org.apache.hadoop.hbase.client.HBaseAdmin.createTable(HBaseAdmin.java:577)
    at com.splicemachine.lifecycle.RegionServerLifecycle.distributedStart(RegionServerLifecycle.java:66)
    at com.splicemachine.derby.lifecycle.EngineLifecycleService.start(EngineLifecycleService.java:81)
    at com.splicemachine.lifecycle.DatabaseLifecycleManager$Startup.bootServices(DatabaseLifecycleManager.java:229)
    at com.splicemachine.lifecycle.DatabaseLifecycleManager$Startup.run(DatabaseLifecycleManager.java:220)
    at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
    at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
    at java.lang.Thread.run(Thread.java:745)
Caused by: org.apache.hadoop.hbase.MasterNotRunningException: java.io.IOException: Can't get master address from ZooKeeper; znode data == null
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation$StubMaker.makeStub(ConnectionManager.java:1540)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation$MasterServiceStubMaker.makeStub(ConnectionManager.java:1560)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.getKeepAliveMasterService(ConnectionManager.java:1711)
    at org.apache.hadoop.hbase.client.MasterCallable.prepare(MasterCallable.java:38)
    at org.apache.hadoop.hbase.client.RpcRetryingCaller.callWithRetries(RpcRetryingCaller.java:124)
    ... 14 more
Caused by: java.io.IOException: Can't get master address from ZooKeeper; znode data == null
    at org.apache.hadoop.hbase.zookeeper.MasterAddressTracker.getMasterAddress(MasterAddressTracker.java:154)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation$StubMaker.makeStubNoRetries(ConnectionManager.java:1491)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation$StubMaker.makeStub(ConnectionManager.java:1531)
    ... 18 more
2016-12-02 19:20:54,020 WARN  [hbase-region-load-updater-0] client.ConnectionManager$HConnectionImplementation: Checking master connection
com.google.protobuf.ServiceException: java.net.ConnectException: Connection refused
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient.callBlockingMethod(AbstractRpcClient.java:223)
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient$BlockingRpcChannelImplementation.callBlockingMethod(AbstractRpcClient.java:287)
    at org.apache.hadoop.hbase.protobuf.generated.MasterProtos$MasterService$BlockingStub.isMasterRunning(MasterProtos.java:58152)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation$MasterServiceState.isMasterRunning(ConnectionManager.java:1444)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.isKeepAliveMasterConnectedAndRunning(ConnectionManager.java:2099)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.getKeepAliveMasterService(ConnectionManager.java:1708)
    at org.apache.hadoop.hbase.client.MasterCallable.prepare(MasterCallable.java:38)
    at org.apache.hadoop.hbase.client.RpcRetryingCaller.callWithRetries(RpcRetryingCaller.java:124)
    at org.apache.hadoop.hbase.client.HBaseAdmin.executeCallable(HBaseAdmin.java:4083)
    at org.apache.hadoop.hbase.client.HBaseAdmin.getClusterStatus(HBaseAdmin.java:2765)
    at com.splicemachine.access.hbase.H10PartitionAdmin.allServers(H10PartitionAdmin.java:120)
    at com.splicemachine.hbase.HBaseRegionLoads.fetchRegionLoads(HBaseRegionLoads.java:164)
    at com.splicemachine.hbase.HBaseRegionLoads.access$000(HBaseRegionLoads.java:60)
    at com.splicemachine.hbase.HBaseRegionLoads$1.run(HBaseRegionLoads.java:81)
    at com.splicemachine.concurrent.LoggingScheduledThreadPoolExecutor$LoggingRunnable.run(LoggingScheduledThreadPoolExecutor.java:75)
    at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
    at java.util.concurrent.FutureTask.runAndReset(FutureTask.java:308)
    at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$301(ScheduledThreadPoolExecutor.java:180)
    at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:294)
    at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
    at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
    at java.lang.Thread.run(Thread.java:745)
Caused by: java.net.ConnectException: Connection refused
    at sun.nio.ch.SocketChannelImpl.checkConnect(Native Method)
    at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:717)
    at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:206)
    at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:531)
    at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:495)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.setupConnection(RpcClientImpl.java:410)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.setupIOstreams(RpcClientImpl.java:716)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.writeRequest(RpcClientImpl.java:887)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.tracedWriteRequest(RpcClientImpl.java:856)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl.call(RpcClientImpl.java:1200)
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient.callBlockingMethod(AbstractRpcClient.java:213)
    ... 21 more
2016-12-02 19:20:54,225 WARN  [hbase-region-load-updater-0] client.ConnectionManager$HConnectionImplementation: Checking master connection
com.google.protobuf.ServiceException: org.apache.hadoop.hbase.ipc.FailedServerException: This server is in the failed servers list: hscale-dev1-nn/10.60.70.10:16000
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient.callBlockingMethod(AbstractRpcClient.java:223)
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient$BlockingRpcChannelImplementation.callBlockingMethod(AbstractRpcClient.java:287)
    at org.apache.hadoop.hbase.protobuf.generated.MasterProtos$MasterService$BlockingStub.isMasterRunning(MasterProtos.java:58152)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation$MasterServiceState.isMasterRunning(ConnectionManager.java:1444)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.isKeepAliveMasterConnectedAndRunning(ConnectionManager.java:2099)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.getKeepAliveMasterService(ConnectionManager.java:1708)
    at org.apache.hadoop.hbase.client.MasterCallable.prepare(MasterCallable.java:38)
    at org.apache.hadoop.hbase.client.RpcRetryingCaller.callWithRetries(RpcRetryingCaller.java:124)
    at org.apache.hadoop.hbase.client.HBaseAdmin.executeCallable(HBaseAdmin.java:4083)
    at org.apache.hadoop.hbase.client.HBaseAdmin.getClusterStatus(HBaseAdmin.java:2765)
    at com.splicemachine.access.hbase.H10PartitionAdmin.allServers(H10PartitionAdmin.java:120)
    at com.splicemachine.hbase.HBaseRegionLoads.fetchRegionLoads(HBaseRegionLoads.java:164)
    at com.splicemachine.hbase.HBaseRegionLoads.access$000(HBaseRegionLoads.java:60)
    at com.splicemachine.hbase.HBaseRegionLoads$1.run(HBaseRegionLoads.java:81)
    at com.splicemachine.concurrent.LoggingScheduledThreadPoolExecutor$LoggingRunnable.run(LoggingScheduledThreadPoolExecutor.java:75)
    at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
    at java.util.concurrent.FutureTask.runAndReset(FutureTask.java:308)
    at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$301(ScheduledThreadPoolExecutor.java:180)
    at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:294)
    at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
    at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
    at java.lang.Thread.run(Thread.java:745)
Caused by: org.apache.hadoop.hbase.ipc.FailedServerException: This server is in the failed servers list: hscale-dev1-nn/10.60.70.10:16000
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.setupIOstreams(RpcClientImpl.java:701)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.writeRequest(RpcClientImpl.java:887)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.tracedWriteRequest(RpcClientImpl.java:856)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl.call(RpcClientImpl.java:1200)
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient.callBlockingMethod(AbstractRpcClient.java:213)
    ... 21 more
2016-12-02 19:20:54,530 WARN  [hbase-region-load-updater-0] client.ConnectionManager$HConnectionImplementation: Checking master connection
com.google.protobuf.ServiceException: org.apache.hadoop.hbase.ipc.FailedServerException: This server is in the failed servers list: hscale-dev1-nn/10.60.70.10:16000
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient.callBlockingMethod(AbstractRpcClient.java:223)
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient$BlockingRpcChannelImplementation.callBlockingMethod(AbstractRpcClient.java:287)
    at org.apache.hadoop.hbase.protobuf.generated.MasterProtos$MasterService$BlockingStub.isMasterRunning(MasterProtos.java:58152)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation$MasterServiceState.isMasterRunning(ConnectionManager.java:1444)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.isKeepAliveMasterConnectedAndRunning(ConnectionManager.java:2099)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.getKeepAliveMasterService(ConnectionManager.java:1708)
    at org.apache.hadoop.hbase.client.MasterCallable.prepare(MasterCallable.java:38)
    at org.apache.hadoop.hbase.client.RpcRetryingCaller.callWithRetries(RpcRetryingCaller.java:124)
    at org.apache.hadoop.hbase.client.HBaseAdmin.executeCallable(HBaseAdmin.java:4083)
    at org.apache.hadoop.hbase.client.HBaseAdmin.getClusterStatus(HBaseAdmin.java:2765)
    at com.splicemachine.access.hbase.H10PartitionAdmin.allServers(H10PartitionAdmin.java:120)
    at com.splicemachine.hbase.HBaseRegionLoads.fetchRegionLoads(HBaseRegionLoads.java:164)
    at com.splicemachine.hbase.HBaseRegionLoads.access$000(HBaseRegionLoads.java:60)
    at com.splicemachine.hbase.HBaseRegionLoads$1.run(HBaseRegionLoads.java:81)
    at com.splicemachine.concurrent.LoggingScheduledThreadPoolExecutor$LoggingRunnable.run(LoggingScheduledThreadPoolExecutor.java:75)
    at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
    at java.util.concurrent.FutureTask.runAndReset(FutureTask.java:308)
    at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$301(ScheduledThreadPoolExecutor.java:180)
    at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:294)
    at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
    at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
    at java.lang.Thread.run(Thread.java:745)
Caused by: org.apache.hadoop.hbase.ipc.FailedServerException: This server is in the failed servers list: hscale-dev1-nn/10.60.70.10:16000
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.setupIOstreams(RpcClientImpl.java:701)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.writeRequest(RpcClientImpl.java:887)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.tracedWriteRequest(RpcClientImpl.java:856)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl.call(RpcClientImpl.java:1200)
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient.callBlockingMethod(AbstractRpcClient.java:213)
    ... 21 more
2016-12-02 19:20:55,037 WARN  [hbase-region-load-updater-0] client.ConnectionManager$HConnectionImplementation: Checking master connection
com.google.protobuf.ServiceException: org.apache.hadoop.hbase.ipc.FailedServerException: This server is in the failed servers list: hscale-dev1-nn/10.60.70.10:16000
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient.callBlockingMethod(AbstractRpcClient.java:223)
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient$BlockingRpcChannelImplementation.callBlockingMethod(AbstractRpcClient.java:287)
    at org.apache.hadoop.hbase.protobuf.generated.MasterProtos$MasterService$BlockingStub.isMasterRunning(MasterProtos.java:58152)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation$MasterServiceState.isMasterRunning(ConnectionManager.java:1444)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.isKeepAliveMasterConnectedAndRunning(ConnectionManager.java:2099)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.getKeepAliveMasterService(ConnectionManager.java:1708)
    at org.apache.hadoop.hbase.client.MasterCallable.prepare(MasterCallable.java:38)
    at org.apache.hadoop.hbase.client.RpcRetryingCaller.callWithRetries(RpcRetryingCaller.java:124)
    at org.apache.hadoop.hbase.client.HBaseAdmin.executeCallable(HBaseAdmin.java:4083)
    at org.apache.hadoop.hbase.client.HBaseAdmin.getClusterStatus(HBaseAdmin.java:2765)
    at com.splicemachine.access.hbase.H10PartitionAdmin.allServers(H10PartitionAdmin.java:120)
    at com.splicemachine.hbase.HBaseRegionLoads.fetchRegionLoads(HBaseRegionLoads.java:164)
    at com.splicemachine.hbase.HBaseRegionLoads.access$000(HBaseRegionLoads.java:60)
    at com.splicemachine.hbase.HBaseRegionLoads$1.run(HBaseRegionLoads.java:81)
    at com.splicemachine.concurrent.LoggingScheduledThreadPoolExecutor$LoggingRunnable.run(LoggingScheduledThreadPoolExecutor.java:75)
    at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
    at java.util.concurrent.FutureTask.runAndReset(FutureTask.java:308)
    at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$301(ScheduledThreadPoolExecutor.java:180)
    at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:294)
    at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
    at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
    at java.lang.Thread.run(Thread.java:745)
Caused by: org.apache.hadoop.hbase.ipc.FailedServerException: This server is in the failed servers list: hscale-dev1-nn/10.60.70.10:16000
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.setupIOstreams(RpcClientImpl.java:701)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.writeRequest(RpcClientImpl.java:887)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.tracedWriteRequest(RpcClientImpl.java:856)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl.call(RpcClientImpl.java:1200)
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient.callBlockingMethod(AbstractRpcClient.java:213)
    ... 21 more
2016-12-02 19:20:56,046 WARN  [hbase-region-load-updater-0] client.ConnectionManager$HConnectionImplementation: Checking master connection
com.google.protobuf.ServiceException: java.net.ConnectException: Connection refused
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient.callBlockingMethod(AbstractRpcClient.java:223)
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient$BlockingRpcChannelImplementation.callBlockingMethod(AbstractRpcClient.java:287)
    at org.apache.hadoop.hbase.protobuf.generated.MasterProtos$MasterService$BlockingStub.isMasterRunning(MasterProtos.java:58152)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation$MasterServiceState.isMasterRunning(ConnectionManager.java:1444)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.isKeepAliveMasterConnectedAndRunning(ConnectionManager.java:2099)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.getKeepAliveMasterService(ConnectionManager.java:1708)
    at org.apache.hadoop.hbase.client.MasterCallable.prepare(MasterCallable.java:38)
    at org.apache.hadoop.hbase.client.RpcRetryingCaller.callWithRetries(RpcRetryingCaller.java:124)
    at org.apache.hadoop.hbase.client.HBaseAdmin.executeCallable(HBaseAdmin.java:4083)
    at org.apache.hadoop.hbase.client.HBaseAdmin.getClusterStatus(HBaseAdmin.java:2765)
    at com.splicemachine.access.hbase.H10PartitionAdmin.allServers(H10PartitionAdmin.java:120)
    at com.splicemachine.hbase.HBaseRegionLoads.fetchRegionLoads(HBaseRegionLoads.java:164)
    at com.splicemachine.hbase.HBaseRegionLoads.access$000(HBaseRegionLoads.java:60)
    at com.splicemachine.hbase.HBaseRegionLoads$1.run(HBaseRegionLoads.java:81)
    at com.splicemachine.concurrent.LoggingScheduledThreadPoolExecutor$LoggingRunnable.run(LoggingScheduledThreadPoolExecutor.java:75)
    at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
    at java.util.concurrent.FutureTask.runAndReset(FutureTask.java:308)
    at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$301(ScheduledThreadPoolExecutor.java:180)
    at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:294)
    at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
    at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
    at java.lang.Thread.run(Thread.java:745)
Caused by: java.net.ConnectException: Connection refused
    at sun.nio.ch.SocketChannelImpl.checkConnect(Native Method)
    at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:717)
    at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:206)
    at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:531)
    at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:495)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.setupConnection(RpcClientImpl.java:410)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.setupIOstreams(RpcClientImpl.java:716)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.writeRequest(RpcClientImpl.java:887)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.tracedWriteRequest(RpcClientImpl.java:856)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl.call(RpcClientImpl.java:1200)
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient.callBlockingMethod(AbstractRpcClient.java:213)
    ... 21 more
2016-12-02 19:20:56,050 ERROR [hbase-region-load-updater-0] hbase.HBaseRegionLoads: Unable to fetch region load info
org.apache.hadoop.hbase.client.RetriesExhaustedException: Failed after attempts=5, exceptions:
Fri Dec 02 19:20:54 IST 2016, RpcRetryingCaller{globalStartTime=1480686654019, pause=100, retries=5}, org.apache.hadoop.hbase.MasterNotRunningException: java.io.IOException: Can't get master address from ZooKeeper; znode data == null
Fri Dec 02 19:20:54 IST 2016, RpcRetryingCaller{globalStartTime=1480686654019, pause=100, retries=5}, org.apache.hadoop.hbase.MasterNotRunningException: java.io.IOException: Can't get master address from ZooKeeper; znode data == null
Fri Dec 02 19:20:54 IST 2016, RpcRetryingCaller{globalStartTime=1480686654019, pause=100, retries=5}, org.apache.hadoop.hbase.MasterNotRunningException: java.io.IOException: Can't get master address from ZooKeeper; znode data == null
Fri Dec 02 19:20:55 IST 2016, RpcRetryingCaller{globalStartTime=1480686654019, pause=100, retries=5}, org.apache.hadoop.hbase.MasterNotRunningException: java.io.IOException: Can't get master address from ZooKeeper; znode data == null
Fri Dec 02 19:20:56 IST 2016, RpcRetryingCaller{globalStartTime=1480686654019, pause=100, retries=5}, org.apache.hadoop.hbase.MasterNotRunningException: java.io.IOException: Can't get master address from ZooKeeper; znode data == null

    at org.apache.hadoop.hbase.client.RpcRetryingCaller.callWithRetries(RpcRetryingCaller.java:147)
    at org.apache.hadoop.hbase.client.HBaseAdmin.executeCallable(HBaseAdmin.java:4083)
    at org.apache.hadoop.hbase.client.HBaseAdmin.getClusterStatus(HBaseAdmin.java:2765)
    at com.splicemachine.access.hbase.H10PartitionAdmin.allServers(H10PartitionAdmin.java:120)
    at com.splicemachine.hbase.HBaseRegionLoads.fetchRegionLoads(HBaseRegionLoads.java:164)
    at com.splicemachine.hbase.HBaseRegionLoads.access$000(HBaseRegionLoads.java:60)
    at com.splicemachine.hbase.HBaseRegionLoads$1.run(HBaseRegionLoads.java:81)
    at com.splicemachine.concurrent.LoggingScheduledThreadPoolExecutor$LoggingRunnable.run(LoggingScheduledThreadPoolExecutor.java:75)
    at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
    at java.util.concurrent.FutureTask.runAndReset(FutureTask.java:308)
    at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$301(ScheduledThreadPoolExecutor.java:180)
    at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:294)
    at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
    at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
    at java.lang.Thread.run(Thread.java:745)
Caused by: org.apache.hadoop.hbase.MasterNotRunningException: java.io.IOException: Can't get master address from ZooKeeper; znode data == null
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation$StubMaker.makeStub(ConnectionManager.java:1540)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation$MasterServiceStubMaker.makeStub(ConnectionManager.java:1560)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.getKeepAliveMasterService(ConnectionManager.java:1711)
    at org.apache.hadoop.hbase.client.MasterCallable.prepare(MasterCallable.java:38)
    at org.apache.hadoop.hbase.client.RpcRetryingCaller.callWithRetries(RpcRetryingCaller.java:124)
    ... 14 more
Caused by: java.io.IOException: Can't get master address from ZooKeeper; znode data == null
    at org.apache.hadoop.hbase.zookeeper.MasterAddressTracker.getMasterAddress(MasterAddressTracker.java:154)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation$StubMaker.makeStubNoRetries(ConnectionManager.java:1491)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation$StubMaker.makeStub(ConnectionManager.java:1531)
    ... 18 more
2016-12-02 19:20:59,020 WARN  [hbase-region-load-updater-0] client.ConnectionManager$HConnectionImplementation: Checking master connection
com.google.protobuf.ServiceException: java.net.ConnectException: Connection refused
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient.callBlockingMethod(AbstractRpcClient.java:223)
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient$BlockingRpcChannelImplementation.callBlockingMethod(AbstractRpcClient.java:287)
    at org.apache.hadoop.hbase.protobuf.generated.MasterProtos$MasterService$BlockingStub.isMasterRunning(MasterProtos.java:58152)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation$MasterServiceState.isMasterRunning(ConnectionManager.java:1444)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.isKeepAliveMasterConnectedAndRunning(ConnectionManager.java:2099)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.getKeepAliveMasterService(ConnectionManager.java:1708)
    at org.apache.hadoop.hbase.client.MasterCallable.prepare(MasterCallable.java:38)
    at org.apache.hadoop.hbase.client.RpcRetryingCaller.callWithRetries(RpcRetryingCaller.java:124)
    at org.apache.hadoop.hbase.client.HBaseAdmin.executeCallable(HBaseAdmin.java:4083)
    at org.apache.hadoop.hbase.client.HBaseAdmin.getClusterStatus(HBaseAdmin.java:2765)
    at com.splicemachine.access.hbase.H10PartitionAdmin.allServers(H10PartitionAdmin.java:120)
    at com.splicemachine.hbase.HBaseRegionLoads.fetchRegionLoads(HBaseRegionLoads.java:164)
    at com.splicemachine.hbase.HBaseRegionLoads.access$000(HBaseRegionLoads.java:60)
    at com.splicemachine.hbase.HBaseRegionLoads$1.run(HBaseRegionLoads.java:81)
    at com.splicemachine.concurrent.LoggingScheduledThreadPoolExecutor$LoggingRunnable.run(LoggingScheduledThreadPoolExecutor.java:75)
    at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
    at java.util.concurrent.FutureTask.runAndReset(FutureTask.java:308)
    at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$301(ScheduledThreadPoolExecutor.java:180)
    at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:294)
    at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
    at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
    at java.lang.Thread.run(Thread.java:745)
Caused by: java.net.ConnectException: Connection refused
    at sun.nio.ch.SocketChannelImpl.checkConnect(Native Method)
    at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:717)
    at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:206)
    at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:531)
    at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:495)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.setupConnection(RpcClientImpl.java:410)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.setupIOstreams(RpcClientImpl.java:716)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.writeRequest(RpcClientImpl.java:887)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.tracedWriteRequest(RpcClientImpl.java:856)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl.call(RpcClientImpl.java:1200)
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient.callBlockingMethod(AbstractRpcClient.java:213)
    ... 21 more
2016-12-02 19:20:59,225 WARN  [hbase-region-load-updater-0] client.ConnectionManager$HConnectionImplementation: Checking master connection
com.google.protobuf.ServiceException: org.apache.hadoop.hbase.ipc.FailedServerException: This server is in the failed servers list: hscale-dev1-nn/10.60.70.10:16000
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient.callBlockingMethod(AbstractRpcClient.java:223)
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient$BlockingRpcChannelImplementation.callBlockingMethod(AbstractRpcClient.java:287)
    at org.apache.hadoop.hbase.protobuf.generated.MasterProtos$MasterService$BlockingStub.isMasterRunning(MasterProtos.java:58152)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation$MasterServiceState.isMasterRunning(ConnectionManager.java:1444)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.isKeepAliveMasterConnectedAndRunning(ConnectionManager.java:2099)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.getKeepAliveMasterService(ConnectionManager.java:1708)
    at org.apache.hadoop.hbase.client.MasterCallable.prepare(MasterCallable.java:38)
    at org.apache.hadoop.hbase.client.RpcRetryingCaller.callWithRetries(RpcRetryingCaller.java:124)
    at org.apache.hadoop.hbase.client.HBaseAdmin.executeCallable(HBaseAdmin.java:4083)
    at org.apache.hadoop.hbase.client.HBaseAdmin.getClusterStatus(HBaseAdmin.java:2765)
    at com.splicemachine.access.hbase.H10PartitionAdmin.allServers(H10PartitionAdmin.java:120)
    at com.splicemachine.hbase.HBaseRegionLoads.fetchRegionLoads(HBaseRegionLoads.java:164)
    at com.splicemachine.hbase.HBaseRegionLoads.access$000(HBaseRegionLoads.java:60)
    at com.splicemachine.hbase.HBaseRegionLoads$1.run(HBaseRegionLoads.java:81)
    at com.splicemachine.concurrent.LoggingScheduledThreadPoolExecutor$LoggingRunnable.run(LoggingScheduledThreadPoolExecutor.java:75)
    at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
    at java.util.concurrent.FutureTask.runAndReset(FutureTask.java:308)
    at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$301(ScheduledThreadPoolExecutor.java:180)
    at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:294)
    at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
    at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
    at java.lang.Thread.run(Thread.java:745)
Caused by: org.apache.hadoop.hbase.ipc.FailedServerException: This server is in the failed servers list: hscale-dev1-nn/10.60.70.10:16000
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.setupIOstreams(RpcClientImpl.java:701)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.writeRequest(RpcClientImpl.java:887)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.tracedWriteRequest(RpcClientImpl.java:856)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl.call(RpcClientImpl.java:1200)
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient.callBlockingMethod(AbstractRpcClient.java:213)
    ... 21 more
2016-12-02 19:20:59,530 WARN  [hbase-region-load-updater-0] client.ConnectionManager$HConnectionImplementation: Checking master connection
com.google.protobuf.ServiceException: org.apache.hadoop.hbase.ipc.FailedServerException: This server is in the failed servers list: hscale-dev1-nn/10.60.70.10:16000
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient.callBlockingMethod(AbstractRpcClient.java:223)
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient$BlockingRpcChannelImplementation.callBlockingMethod(AbstractRpcClient.java:287)
    at org.apache.hadoop.hbase.protobuf.generated.MasterProtos$MasterService$BlockingStub.isMasterRunning(MasterProtos.java:58152)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation$MasterServiceState.isMasterRunning(ConnectionManager.java:1444)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.isKeepAliveMasterConnectedAndRunning(ConnectionManager.java:2099)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.getKeepAliveMasterService(ConnectionManager.java:1708)
    at org.apache.hadoop.hbase.client.MasterCallable.prepare(MasterCallable.java:38)
    at org.apache.hadoop.hbase.client.RpcRetryingCaller.callWithRetries(RpcRetryingCaller.java:124)
    at org.apache.hadoop.hbase.client.HBaseAdmin.executeCallable(HBaseAdmin.java:4083)
    at org.apache.hadoop.hbase.client.HBaseAdmin.getClusterStatus(HBaseAdmin.java:2765)
    at com.splicemachine.access.hbase.H10PartitionAdmin.allServers(H10PartitionAdmin.java:120)
    at com.splicemachine.hbase.HBaseRegionLoads.fetchRegionLoads(HBaseRegionLoads.java:164)
    at com.splicemachine.hbase.HBaseRegionLoads.access$000(HBaseRegionLoads.java:60)
    at com.splicemachine.hbase.HBaseRegionLoads$1.run(HBaseRegionLoads.java:81)
    at com.splicemachine.concurrent.LoggingScheduledThreadPoolExecutor$LoggingRunnable.run(LoggingScheduledThreadPoolExecutor.java:75)
    at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
    at java.util.concurrent.FutureTask.runAndReset(FutureTask.java:308)
    at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$301(ScheduledThreadPoolExecutor.java:180)
    at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:294)
    at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
    at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
    at java.lang.Thread.run(Thread.java:745)
Caused by: org.apache.hadoop.hbase.ipc.FailedServerException: This server is in the failed servers list: hscale-dev1-nn/10.60.70.10:16000
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.setupIOstreams(RpcClientImpl.java:701)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.writeRequest(RpcClientImpl.java:887)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.tracedWriteRequest(RpcClientImpl.java:856)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl.call(RpcClientImpl.java:1200)
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient.callBlockingMethod(AbstractRpcClient.java:213)
    ... 21 more
2016-12-02 19:21:00,036 WARN  [hbase-region-load-updater-0] client.ConnectionManager$HConnectionImplementation: Checking master connection
com.google.protobuf.ServiceException: org.apache.hadoop.hbase.ipc.FailedServerException: This server is in the failed servers list: hscale-dev1-nn/10.60.70.10:16000
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient.callBlockingMethod(AbstractRpcClient.java:223)
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient$BlockingRpcChannelImplementation.callBlockingMethod(AbstractRpcClient.java:287)
    at org.apache.hadoop.hbase.protobuf.generated.MasterProtos$MasterService$BlockingStub.isMasterRunning(MasterProtos.java:58152)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation$MasterServiceState.isMasterRunning(ConnectionManager.java:1444)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.isKeepAliveMasterConnectedAndRunning(ConnectionManager.java:2099)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.getKeepAliveMasterService(ConnectionManager.java:1708)
    at org.apache.hadoop.hbase.client.MasterCallable.prepare(MasterCallable.java:38)
    at org.apache.hadoop.hbase.client.RpcRetryingCaller.callWithRetries(RpcRetryingCaller.java:124)
    at org.apache.hadoop.hbase.client.HBaseAdmin.executeCallable(HBaseAdmin.java:4083)
    at org.apache.hadoop.hbase.client.HBaseAdmin.getClusterStatus(HBaseAdmin.java:2765)
    at com.splicemachine.access.hbase.H10PartitionAdmin.allServers(H10PartitionAdmin.java:120)
    at com.splicemachine.hbase.HBaseRegionLoads.fetchRegionLoads(HBaseRegionLoads.java:164)
    at com.splicemachine.hbase.HBaseRegionLoads.access$000(HBaseRegionLoads.java:60)
    at com.splicemachine.hbase.HBaseRegionLoads$1.run(HBaseRegionLoads.java:81)
    at com.splicemachine.concurrent.LoggingScheduledThreadPoolExecutor$LoggingRunnable.run(LoggingScheduledThreadPoolExecutor.java:75)
    at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
    at java.util.concurrent.FutureTask.runAndReset(FutureTask.java:308)
    at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$301(ScheduledThreadPoolExecutor.java:180)
    at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:294)
    at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
    at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
    at java.lang.Thread.run(Thread.java:745)
Caused by: org.apache.hadoop.hbase.ipc.FailedServerException: This server is in the failed servers list: hscale-dev1-nn/10.60.70.10:16000
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.setupIOstreams(RpcClientImpl.java:701)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.writeRequest(RpcClientImpl.java:887)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.tracedWriteRequest(RpcClientImpl.java:856)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl.call(RpcClientImpl.java:1200)
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient.callBlockingMethod(AbstractRpcClient.java:213)
    ... 21 more
2016-12-02 19:21:01,046 WARN  [hbase-region-load-updater-0] client.ConnectionManager$HConnectionImplementation: Checking master connection
com.google.protobuf.ServiceException: java.net.ConnectException: Connection refused
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient.callBlockingMethod(AbstractRpcClient.java:223)
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient$BlockingRpcChannelImplementation.callBlockingMethod(AbstractRpcClient.java:287)
    at org.apache.hadoop.hbase.protobuf.generated.MasterProtos$MasterService$BlockingStub.isMasterRunning(MasterProtos.java:58152)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation$MasterServiceState.isMasterRunning(ConnectionManager.java:1444)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.isKeepAliveMasterConnectedAndRunning(ConnectionManager.java:2099)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.getKeepAliveMasterService(ConnectionManager.java:1708)
    at org.apache.hadoop.hbase.client.MasterCallable.prepare(MasterCallable.java:38)
    at org.apache.hadoop.hbase.client.RpcRetryingCaller.callWithRetries(RpcRetryingCaller.java:124)
    at org.apache.hadoop.hbase.client.HBaseAdmin.executeCallable(HBaseAdmin.java:4083)
    at org.apache.hadoop.hbase.client.HBaseAdmin.getClusterStatus(HBaseAdmin.java:2765)
    at com.splicemachine.access.hbase.H10PartitionAdmin.allServers(H10PartitionAdmin.java:120)
    at com.splicemachine.hbase.HBaseRegionLoads.fetchRegionLoads(HBaseRegionLoads.java:164)
    at com.splicemachine.hbase.HBaseRegionLoads.access$000(HBaseRegionLoads.java:60)
    at com.splicemachine.hbase.HBaseRegionLoads$1.run(HBaseRegionLoads.java:81)
    at com.splicemachine.concurrent.LoggingScheduledThreadPoolExecutor$LoggingRunnable.run(LoggingScheduledThreadPoolExecutor.java:75)
    at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
    at java.util.concurrent.FutureTask.runAndReset(FutureTask.java:308)
    at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$301(ScheduledThreadPoolExecutor.java:180)
    at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:294)
    at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
    at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
    at java.lang.Thread.run(Thread.java:745)
Caused by: java.net.ConnectException: Connection refused
    at sun.nio.ch.SocketChannelImpl.checkConnect(Native Method)
    at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:717)
    at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:206)
    at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:531)
    at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:495)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.setupConnection(RpcClientImpl.java:410)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.setupIOstreams(RpcClientImpl.java:716)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.writeRequest(RpcClientImpl.java:887)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.tracedWriteRequest(RpcClientImpl.java:856)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl.call(RpcClientImpl.java:1200)
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient.callBlockingMethod(AbstractRpcClient.java:213)
    ... 21 more
2016-12-02 19:21:01,054 ERROR [hbase-region-load-updater-0] hbase.HBaseRegionLoads: Unable to fetch region load info
org.apache.hadoop.hbase.client.RetriesExhaustedException: Failed after attempts=5, exceptions:
Fri Dec 02 19:20:59 IST 2016, RpcRetryingCaller{globalStartTime=1480686659019, pause=100, retries=5}, org.apache.hadoop.hbase.MasterNotRunningException: java.io.IOException: Can't get master address from ZooKeeper; znode data == null
Fri Dec 02 19:20:59 IST 2016, RpcRetryingCaller{globalStartTime=1480686659019, pause=100, retries=5}, org.apache.hadoop.hbase.MasterNotRunningException: java.io.IOException: Can't get master address from ZooKeeper; znode data == null
Fri Dec 02 19:20:59 IST 2016, RpcRetryingCaller{globalStartTime=1480686659019, pause=100, retries=5}, org.apache.hadoop.hbase.MasterNotRunningException: java.io.IOException: Can't get master address from ZooKeeper; znode data == null
Fri Dec 02 19:21:00 IST 2016, RpcRetryingCaller{globalStartTime=1480686659019, pause=100, retries=5}, org.apache.hadoop.hbase.MasterNotRunningException: java.io.IOException: Can't get master address from ZooKeeper; znode data == null
Fri Dec 02 19:21:01 IST 2016, RpcRetryingCaller{globalStartTime=1480686659019, pause=100, retries=5}, org.apache.hadoop.hbase.MasterNotRunningException: java.io.IOException: Can't get master address from ZooKeeper; znode data == null

    at org.apache.hadoop.hbase.client.RpcRetryingCaller.callWithRetries(RpcRetryingCaller.java:147)
    at org.apache.hadoop.hbase.client.HBaseAdmin.executeCallable(HBaseAdmin.java:4083)
    at org.apache.hadoop.hbase.client.HBaseAdmin.getClusterStatus(HBaseAdmin.java:2765)
    at com.splicemachine.access.hbase.H10PartitionAdmin.allServers(H10PartitionAdmin.java:120)
    at com.splicemachine.hbase.HBaseRegionLoads.fetchRegionLoads(HBaseRegionLoads.java:164)
    at com.splicemachine.hbase.HBaseRegionLoads.access$000(HBaseRegionLoads.java:60)
    at com.splicemachine.hbase.HBaseRegionLoads$1.run(HBaseRegionLoads.java:81)
    at com.splicemachine.concurrent.LoggingScheduledThreadPoolExecutor$LoggingRunnable.run(LoggingScheduledThreadPoolExecutor.java:75)
    at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
    at java.util.concurrent.FutureTask.runAndReset(FutureTask.java:308)
    at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$301(ScheduledThreadPoolExecutor.java:180)
    at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:294)
    at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
    at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
    at java.lang.Thread.run(Thread.java:745)
Caused by: org.apache.hadoop.hbase.MasterNotRunningException: java.io.IOException: Can't get master address from ZooKeeper; znode data == null
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation$StubMaker.makeStub(ConnectionManager.java:1540)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation$MasterServiceStubMaker.makeStub(ConnectionManager.java:1560)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.getKeepAliveMasterService(ConnectionManager.java:1711)
    at org.apache.hadoop.hbase.client.MasterCallable.prepare(MasterCallable.java:38)
    at org.apache.hadoop.hbase.client.RpcRetryingCaller.callWithRetries(RpcRetryingCaller.java:124)
    ... 14 more
Caused by: java.io.IOException: Can't get master address from ZooKeeper; znode data == null
    at org.apache.hadoop.hbase.zookeeper.MasterAddressTracker.getMasterAddress(MasterAddressTracker.java:154)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation$StubMaker.makeStubNoRetries(ConnectionManager.java:1491)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation$StubMaker.makeStub(ConnectionManager.java:1531)
    ... 18 more
2016-12-02 19:21:02,079 WARN  [pool-10-thread-1] client.ConnectionManager$HConnectionImplementation: Checking master connection
com.google.protobuf.ServiceException: org.apache.hadoop.hbase.ipc.FailedServerException: This server is in the failed servers list: hscale-dev1-nn/10.60.70.10:16000
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient.callBlockingMethod(AbstractRpcClient.java:223)
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient$BlockingRpcChannelImplementation.callBlockingMethod(AbstractRpcClient.java:287)
    at org.apache.hadoop.hbase.protobuf.generated.MasterProtos$MasterService$BlockingStub.isMasterRunning(MasterProtos.java:58152)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation$MasterServiceState.isMasterRunning(ConnectionManager.java:1444)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.isKeepAliveMasterConnectedAndRunning(ConnectionManager.java:2099)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.getKeepAliveMasterService(ConnectionManager.java:1708)
    at org.apache.hadoop.hbase.client.MasterCallable.prepare(MasterCallable.java:38)
    at org.apache.hadoop.hbase.client.RpcRetryingCaller.callWithRetries(RpcRetryingCaller.java:124)
    at org.apache.hadoop.hbase.client.HBaseAdmin.executeCallable(HBaseAdmin.java:4083)
    at org.apache.hadoop.hbase.client.HBaseAdmin.access$700(HBaseAdmin.java:194)
    at org.apache.hadoop.hbase.client.HBaseAdmin$ProcedureFuture.getProcedureResult(HBaseAdmin.java:4383)
    at org.apache.hadoop.hbase.client.HBaseAdmin$ProcedureFuture.waitProcedureResult(HBaseAdmin.java:4335)
    at org.apache.hadoop.hbase.client.HBaseAdmin$ProcedureFuture.get(HBaseAdmin.java:4291)
    at org.apache.hadoop.hbase.client.HBaseAdmin.createTable(HBaseAdmin.java:647)
    at org.apache.hadoop.hbase.client.HBaseAdmin.createTable(HBaseAdmin.java:577)
    at com.splicemachine.lifecycle.RegionServerLifecycle.distributedStart(RegionServerLifecycle.java:66)
    at com.splicemachine.derby.lifecycle.EngineLifecycleService.start(EngineLifecycleService.java:81)
    at com.splicemachine.lifecycle.DatabaseLifecycleManager$Startup.bootServices(DatabaseLifecycleManager.java:229)
    at com.splicemachine.lifecycle.DatabaseLifecycleManager$Startup.run(DatabaseLifecycleManager.java:220)
    at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
    at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
    at java.lang.Thread.run(Thread.java:745)
Caused by: org.apache.hadoop.hbase.ipc.FailedServerException: This server is in the failed servers list: hscale-dev1-nn/10.60.70.10:16000
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.setupIOstreams(RpcClientImpl.java:701)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.writeRequest(RpcClientImpl.java:887)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.tracedWriteRequest(RpcClientImpl.java:856)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl.call(RpcClientImpl.java:1200)
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient.callBlockingMethod(AbstractRpcClient.java:213)
    ... 21 more
2016-12-02 19:21:02,282 WARN  [pool-10-thread-1] client.ConnectionManager$HConnectionImplementation: Checking master connection
com.google.protobuf.ServiceException: org.apache.hadoop.hbase.ipc.FailedServerException: This server is in the failed servers list: hscale-dev1-nn/10.60.70.10:16000
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient.callBlockingMethod(AbstractRpcClient.java:223)
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient$BlockingRpcChannelImplementation.callBlockingMethod(AbstractRpcClient.java:287)
    at org.apache.hadoop.hbase.protobuf.generated.MasterProtos$MasterService$BlockingStub.isMasterRunning(MasterProtos.java:58152)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation$MasterServiceState.isMasterRunning(ConnectionManager.java:1444)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.isKeepAliveMasterConnectedAndRunning(ConnectionManager.java:2099)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.getKeepAliveMasterService(ConnectionManager.java:1708)
    at org.apache.hadoop.hbase.client.MasterCallable.prepare(MasterCallable.java:38)
    at org.apache.hadoop.hbase.client.RpcRetryingCaller.callWithRetries(RpcRetryingCaller.java:124)
    at org.apache.hadoop.hbase.client.HBaseAdmin.executeCallable(HBaseAdmin.java:4083)
    at org.apache.hadoop.hbase.client.HBaseAdmin.access$700(HBaseAdmin.java:194)
    at org.apache.hadoop.hbase.client.HBaseAdmin$ProcedureFuture.getProcedureResult(HBaseAdmin.java:4383)
    at org.apache.hadoop.hbase.client.HBaseAdmin$ProcedureFuture.waitProcedureResult(HBaseAdmin.java:4335)
    at org.apache.hadoop.hbase.client.HBaseAdmin$ProcedureFuture.get(HBaseAdmin.java:4291)
    at org.apache.hadoop.hbase.client.HBaseAdmin.createTable(HBaseAdmin.java:647)
    at org.apache.hadoop.hbase.client.HBaseAdmin.createTable(HBaseAdmin.java:577)
    at com.splicemachine.lifecycle.RegionServerLifecycle.distributedStart(RegionServerLifecycle.java:66)
    at com.splicemachine.derby.lifecycle.EngineLifecycleService.start(EngineLifecycleService.java:81)
    at com.splicemachine.lifecycle.DatabaseLifecycleManager$Startup.bootServices(DatabaseLifecycleManager.java:229)
    at com.splicemachine.lifecycle.DatabaseLifecycleManager$Startup.run(DatabaseLifecycleManager.java:220)
    at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
    at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
    at java.lang.Thread.run(Thread.java:745)
Caused by: org.apache.hadoop.hbase.ipc.FailedServerException: This server is in the failed servers list: hscale-dev1-nn/10.60.70.10:16000
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.setupIOstreams(RpcClientImpl.java:701)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.writeRequest(RpcClientImpl.java:887)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.tracedWriteRequest(RpcClientImpl.java:856)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl.call(RpcClientImpl.java:1200)
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient.callBlockingMethod(AbstractRpcClient.java:213)
    ... 21 more
2016-12-02 19:21:02,585 WARN  [pool-10-thread-1] client.ConnectionManager$HConnectionImplementation: Checking master connection
com.google.protobuf.ServiceException: org.apache.hadoop.hbase.ipc.FailedServerException: This server is in the failed servers list: hscale-dev1-nn/10.60.70.10:16000
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient.callBlockingMethod(AbstractRpcClient.java:223)
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient$BlockingRpcChannelImplementation.callBlockingMethod(AbstractRpcClient.java:287)
    at org.apache.hadoop.hbase.protobuf.generated.MasterProtos$MasterService$BlockingStub.isMasterRunning(MasterProtos.java:58152)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation$MasterServiceState.isMasterRunning(ConnectionManager.java:1444)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.isKeepAliveMasterConnectedAndRunning(ConnectionManager.java:2099)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.getKeepAliveMasterService(ConnectionManager.java:1708)
    at org.apache.hadoop.hbase.client.MasterCallable.prepare(MasterCallable.java:38)
    at org.apache.hadoop.hbase.client.RpcRetryingCaller.callWithRetries(RpcRetryingCaller.java:124)
    at org.apache.hadoop.hbase.client.HBaseAdmin.executeCallable(HBaseAdmin.java:4083)
    at org.apache.hadoop.hbase.client.HBaseAdmin.access$700(HBaseAdmin.java:194)
    at org.apache.hadoop.hbase.client.HBaseAdmin$ProcedureFuture.getProcedureResult(HBaseAdmin.java:4383)
    at org.apache.hadoop.hbase.client.HBaseAdmin$ProcedureFuture.waitProcedureResult(HBaseAdmin.java:4335)
    at org.apache.hadoop.hbase.client.HBaseAdmin$ProcedureFuture.get(HBaseAdmin.java:4291)
    at org.apache.hadoop.hbase.client.HBaseAdmin.createTable(HBaseAdmin.java:647)
    at org.apache.hadoop.hbase.client.HBaseAdmin.createTable(HBaseAdmin.java:577)
    at com.splicemachine.lifecycle.RegionServerLifecycle.distributedStart(RegionServerLifecycle.java:66)
    at com.splicemachine.derby.lifecycle.EngineLifecycleService.start(EngineLifecycleService.java:81)
    at com.splicemachine.lifecycle.DatabaseLifecycleManager$Startup.bootServices(DatabaseLifecycleManager.java:229)
    at com.splicemachine.lifecycle.DatabaseLifecycleManager$Startup.run(DatabaseLifecycleManager.java:220)
    at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
    at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
    at java.lang.Thread.run(Thread.java:745)
Caused by: org.apache.hadoop.hbase.ipc.FailedServerException: This server is in the failed servers list: hscale-dev1-nn/10.60.70.10:16000
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.setupIOstreams(RpcClientImpl.java:701)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.writeRequest(RpcClientImpl.java:887)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.tracedWriteRequest(RpcClientImpl.java:856)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl.call(RpcClientImpl.java:1200)
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient.callBlockingMethod(AbstractRpcClient.java:213)
    ... 21 more
2016-12-02 19:21:03,093 WARN  [pool-10-thread-1] client.ConnectionManager$HConnectionImplementation: Checking master connection
com.google.protobuf.ServiceException: java.net.ConnectException: Connection refused
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient.callBlockingMethod(AbstractRpcClient.java:223)
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient$BlockingRpcChannelImplementation.callBlockingMethod(AbstractRpcClient.java:287)
    at org.apache.hadoop.hbase.protobuf.generated.MasterProtos$MasterService$BlockingStub.isMasterRunning(MasterProtos.java:58152)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation$MasterServiceState.isMasterRunning(ConnectionManager.java:1444)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.isKeepAliveMasterConnectedAndRunning(ConnectionManager.java:2099)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.getKeepAliveMasterService(ConnectionManager.java:1708)
    at org.apache.hadoop.hbase.client.MasterCallable.prepare(MasterCallable.java:38)
    at org.apache.hadoop.hbase.client.RpcRetryingCaller.callWithRetries(RpcRetryingCaller.java:124)
    at org.apache.hadoop.hbase.client.HBaseAdmin.executeCallable(HBaseAdmin.java:4083)
    at org.apache.hadoop.hbase.client.HBaseAdmin.access$700(HBaseAdmin.java:194)
    at org.apache.hadoop.hbase.client.HBaseAdmin$ProcedureFuture.getProcedureResult(HBaseAdmin.java:4383)
    at org.apache.hadoop.hbase.client.HBaseAdmin$ProcedureFuture.waitProcedureResult(HBaseAdmin.java:4335)
    at org.apache.hadoop.hbase.client.HBaseAdmin$ProcedureFuture.get(HBaseAdmin.java:4291)
    at org.apache.hadoop.hbase.client.HBaseAdmin.createTable(HBaseAdmin.java:647)
    at org.apache.hadoop.hbase.client.HBaseAdmin.createTable(HBaseAdmin.java:577)
    at com.splicemachine.lifecycle.RegionServerLifecycle.distributedStart(RegionServerLifecycle.java:66)
    at com.splicemachine.derby.lifecycle.EngineLifecycleService.start(EngineLifecycleService.java:81)
    at com.splicemachine.lifecycle.DatabaseLifecycleManager$Startup.bootServices(DatabaseLifecycleManager.java:229)
    at com.splicemachine.lifecycle.DatabaseLifecycleManager$Startup.run(DatabaseLifecycleManager.java:220)
    at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
    at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
    at java.lang.Thread.run(Thread.java:745)
Caused by: java.net.ConnectException: Connection refused
    at sun.nio.ch.SocketChannelImpl.checkConnect(Native Method)
    at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:717)
    at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:206)
    at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:531)
    at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:495)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.setupConnection(RpcClientImpl.java:410)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.setupIOstreams(RpcClientImpl.java:716)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.writeRequest(RpcClientImpl.java:887)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.tracedWriteRequest(RpcClientImpl.java:856)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl.call(RpcClientImpl.java:1200)
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient.callBlockingMethod(AbstractRpcClient.java:213)
    ... 21 more
2016-12-02 19:21:04,019 WARN  [hbase-region-load-updater-0] client.ConnectionManager$HConnectionImplementation: Checking master connection
com.google.protobuf.ServiceException: org.apache.hadoop.hbase.ipc.FailedServerException: This server is in the failed servers list: hscale-dev1-nn/10.60.70.10:16000
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient.callBlockingMethod(AbstractRpcClient.java:223)
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient$BlockingRpcChannelImplementation.callBlockingMethod(AbstractRpcClient.java:287)
    at org.apache.hadoop.hbase.protobuf.generated.MasterProtos$MasterService$BlockingStub.isMasterRunning(MasterProtos.java:58152)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation$MasterServiceState.isMasterRunning(ConnectionManager.java:1444)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.isKeepAliveMasterConnectedAndRunning(ConnectionManager.java:2099)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.getKeepAliveMasterService(ConnectionManager.java:1708)
    at org.apache.hadoop.hbase.client.MasterCallable.prepare(MasterCallable.java:38)
    at org.apache.hadoop.hbase.client.RpcRetryingCaller.callWithRetries(RpcRetryingCaller.java:124)
    at org.apache.hadoop.hbase.client.HBaseAdmin.executeCallable(HBaseAdmin.java:4083)
    at org.apache.hadoop.hbase.client.HBaseAdmin.getClusterStatus(HBaseAdmin.java:2765)
    at com.splicemachine.access.hbase.H10PartitionAdmin.allServers(H10PartitionAdmin.java:120)
    at com.splicemachine.hbase.HBaseRegionLoads.fetchRegionLoads(HBaseRegionLoads.java:164)
    at com.splicemachine.hbase.HBaseRegionLoads.access$000(HBaseRegionLoads.java:60)
    at com.splicemachine.hbase.HBaseRegionLoads$1.run(HBaseRegionLoads.java:81)
    at com.splicemachine.concurrent.LoggingScheduledThreadPoolExecutor$LoggingRunnable.run(LoggingScheduledThreadPoolExecutor.java:75)
    at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
    at java.util.concurrent.FutureTask.runAndReset(FutureTask.java:308)
    at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$301(ScheduledThreadPoolExecutor.java:180)
    at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:294)
    at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
    at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
    at java.lang.Thread.run(Thread.java:745)
Caused by: org.apache.hadoop.hbase.ipc.FailedServerException: This server is in the failed servers list: hscale-dev1-nn/10.60.70.10:16000
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.setupIOstreams(RpcClientImpl.java:701)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.writeRequest(RpcClientImpl.java:887)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.tracedWriteRequest(RpcClientImpl.java:856)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl.call(RpcClientImpl.java:1200)
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient.callBlockingMethod(AbstractRpcClient.java:213)
    ... 21 more
2016-12-02 19:21:04,104 WARN  [pool-10-thread-1] client.ConnectionManager$HConnectionImplementation: Checking master connection
com.google.protobuf.ServiceException: org.apache.hadoop.hbase.ipc.FailedServerException: This server is in the failed servers list: hscale-dev1-nn/10.60.70.10:16000
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient.callBlockingMethod(AbstractRpcClient.java:223)
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient$BlockingRpcChannelImplementation.callBlockingMethod(AbstractRpcClient.java:287)
    at org.apache.hadoop.hbase.protobuf.generated.MasterProtos$MasterService$BlockingStub.isMasterRunning(MasterProtos.java:58152)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation$MasterServiceState.isMasterRunning(ConnectionManager.java:1444)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.isKeepAliveMasterConnectedAndRunning(ConnectionManager.java:2099)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.getKeepAliveMasterService(ConnectionManager.java:1708)
    at org.apache.hadoop.hbase.client.MasterCallable.prepare(MasterCallable.java:38)
    at org.apache.hadoop.hbase.client.RpcRetryingCaller.callWithRetries(RpcRetryingCaller.java:124)
    at org.apache.hadoop.hbase.client.HBaseAdmin.executeCallable(HBaseAdmin.java:4083)
    at org.apache.hadoop.hbase.client.HBaseAdmin.access$700(HBaseAdmin.java:194)
    at org.apache.hadoop.hbase.client.HBaseAdmin$ProcedureFuture.getProcedureResult(HBaseAdmin.java:4383)
    at org.apache.hadoop.hbase.client.HBaseAdmin$ProcedureFuture.waitProcedureResult(HBaseAdmin.java:4335)
    at org.apache.hadoop.hbase.client.HBaseAdmin$ProcedureFuture.get(HBaseAdmin.java:4291)
    at org.apache.hadoop.hbase.client.HBaseAdmin.createTable(HBaseAdmin.java:647)
    at org.apache.hadoop.hbase.client.HBaseAdmin.createTable(HBaseAdmin.java:577)
    at com.splicemachine.lifecycle.RegionServerLifecycle.distributedStart(RegionServerLifecycle.java:66)
    at com.splicemachine.derby.lifecycle.EngineLifecycleService.start(EngineLifecycleService.java:81)
    at com.splicemachine.lifecycle.DatabaseLifecycleManager$Startup.bootServices(DatabaseLifecycleManager.java:229)
    at com.splicemachine.lifecycle.DatabaseLifecycleManager$Startup.run(DatabaseLifecycleManager.java:220)
    at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
    at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
    at java.lang.Thread.run(Thread.java:745)
Caused by: org.apache.hadoop.hbase.ipc.FailedServerException: This server is in the failed servers list: hscale-dev1-nn/10.60.70.10:16000
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.setupIOstreams(RpcClientImpl.java:701)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.writeRequest(RpcClientImpl.java:887)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.tracedWriteRequest(RpcClientImpl.java:856)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl.call(RpcClientImpl.java:1200)
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient.callBlockingMethod(AbstractRpcClient.java:213)
    ... 21 more
2016-12-02 19:21:04,108 WARN  [pool-10-thread-1] client.HBaseAdmin: failed to get the procedure result procId=6
org.apache.hadoop.hbase.client.RetriesExhaustedException: Failed after attempts=5, exceptions:
Fri Dec 02 19:21:02 IST 2016, RpcRetryingCaller{globalStartTime=1480686662079, pause=100, retries=5}, org.apache.hadoop.hbase.MasterNotRunningException: java.io.IOException: Can't get master address from ZooKeeper; znode data == null
Fri Dec 02 19:21:02 IST 2016, RpcRetryingCaller{globalStartTime=1480686662079, pause=100, retries=5}, org.apache.hadoop.hbase.MasterNotRunningException: java.io.IOException: Can't get master address from ZooKeeper; znode data == null
Fri Dec 02 19:21:02 IST 2016, RpcRetryingCaller{globalStartTime=1480686662079, pause=100, retries=5}, org.apache.hadoop.hbase.MasterNotRunningException: java.io.IOException: Can't get master address from ZooKeeper; znode data == null
Fri Dec 02 19:21:03 IST 2016, RpcRetryingCaller{globalStartTime=1480686662079, pause=100, retries=5}, org.apache.hadoop.hbase.MasterNotRunningException: java.io.IOException: Can't get master address from ZooKeeper; znode data == null
Fri Dec 02 19:21:04 IST 2016, RpcRetryingCaller{globalStartTime=1480686662079, pause=100, retries=5}, org.apache.hadoop.hbase.MasterNotRunningException: java.io.IOException: Can't get master address from ZooKeeper; znode data == null

    at org.apache.hadoop.hbase.client.RpcRetryingCaller.callWithRetries(RpcRetryingCaller.java:147)
    at org.apache.hadoop.hbase.client.HBaseAdmin.executeCallable(HBaseAdmin.java:4083)
    at org.apache.hadoop.hbase.client.HBaseAdmin.access$700(HBaseAdmin.java:194)
    at org.apache.hadoop.hbase.client.HBaseAdmin$ProcedureFuture.getProcedureResult(HBaseAdmin.java:4383)
    at org.apache.hadoop.hbase.client.HBaseAdmin$ProcedureFuture.waitProcedureResult(HBaseAdmin.java:4335)
    at org.apache.hadoop.hbase.client.HBaseAdmin$ProcedureFuture.get(HBaseAdmin.java:4291)
    at org.apache.hadoop.hbase.client.HBaseAdmin.createTable(HBaseAdmin.java:647)
    at org.apache.hadoop.hbase.client.HBaseAdmin.createTable(HBaseAdmin.java:577)
    at com.splicemachine.lifecycle.RegionServerLifecycle.distributedStart(RegionServerLifecycle.java:66)
    at com.splicemachine.derby.lifecycle.EngineLifecycleService.start(EngineLifecycleService.java:81)
    at com.splicemachine.lifecycle.DatabaseLifecycleManager$Startup.bootServices(DatabaseLifecycleManager.java:229)
    at com.splicemachine.lifecycle.DatabaseLifecycleManager$Startup.run(DatabaseLifecycleManager.java:220)
    at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
    at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
    at java.lang.Thread.run(Thread.java:745)
Caused by: org.apache.hadoop.hbase.MasterNotRunningException: java.io.IOException: Can't get master address from ZooKeeper; znode data == null
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation$StubMaker.makeStub(ConnectionManager.java:1540)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation$MasterServiceStubMaker.makeStub(ConnectionManager.java:1560)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.getKeepAliveMasterService(ConnectionManager.java:1711)
    at org.apache.hadoop.hbase.client.MasterCallable.prepare(MasterCallable.java:38)
    at org.apache.hadoop.hbase.client.RpcRetryingCaller.callWithRetries(RpcRetryingCaller.java:124)
    ... 14 more
Caused by: java.io.IOException: Can't get master address from ZooKeeper; znode data == null
    at org.apache.hadoop.hbase.zookeeper.MasterAddressTracker.getMasterAddress(MasterAddressTracker.java:154)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation$StubMaker.makeStubNoRetries(ConnectionManager.java:1491)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation$StubMaker.makeStub(ConnectionManager.java:1531)
    ... 18 more
2016-12-02 19:21:04,222 WARN  [hbase-region-load-updater-0] client.ConnectionManager$HConnectionImplementation: Checking master connection
com.google.protobuf.ServiceException: org.apache.hadoop.hbase.ipc.FailedServerException: This server is in the failed servers list: hscale-dev1-nn/10.60.70.10:16000
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient.callBlockingMethod(AbstractRpcClient.java:223)
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient$BlockingRpcChannelImplementation.callBlockingMethod(AbstractRpcClient.java:287)
    at org.apache.hadoop.hbase.protobuf.generated.MasterProtos$MasterService$BlockingStub.isMasterRunning(MasterProtos.java:58152)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation$MasterServiceState.isMasterRunning(ConnectionManager.java:1444)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.isKeepAliveMasterConnectedAndRunning(ConnectionManager.java:2099)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.getKeepAliveMasterService(ConnectionManager.java:1708)
    at org.apache.hadoop.hbase.client.MasterCallable.prepare(MasterCallable.java:38)
    at org.apache.hadoop.hbase.client.RpcRetryingCaller.callWithRetries(RpcRetryingCaller.java:124)
    at org.apache.hadoop.hbase.client.HBaseAdmin.executeCallable(HBaseAdmin.java:4083)
    at org.apache.hadoop.hbase.client.HBaseAdmin.getClusterStatus(HBaseAdmin.java:2765)
    at com.splicemachine.access.hbase.H10PartitionAdmin.allServers(H10PartitionAdmin.java:120)
    at com.splicemachine.hbase.HBaseRegionLoads.fetchRegionLoads(HBaseRegionLoads.java:164)
    at com.splicemachine.hbase.HBaseRegionLoads.access$000(HBaseRegionLoads.java:60)
    at com.splicemachine.hbase.HBaseRegionLoads$1.run(HBaseRegionLoads.java:81)
    at com.splicemachine.concurrent.LoggingScheduledThreadPoolExecutor$LoggingRunnable.run(LoggingScheduledThreadPoolExecutor.java:75)
    at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
    at java.util.concurrent.FutureTask.runAndReset(FutureTask.java:308)
    at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$301(ScheduledThreadPoolExecutor.java:180)
    at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:294)
    at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
    at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
    at java.lang.Thread.run(Thread.java:745)
Caused by: org.apache.hadoop.hbase.ipc.FailedServerException: This server is in the failed servers list: hscale-dev1-nn/10.60.70.10:16000
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.setupIOstreams(RpcClientImpl.java:701)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.writeRequest(RpcClientImpl.java:887)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.tracedWriteRequest(RpcClientImpl.java:856)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl.call(RpcClientImpl.java:1200)
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient.callBlockingMethod(AbstractRpcClient.java:213)
    ... 21 more
2016-12-02 19:21:04,525 WARN  [hbase-region-load-updater-0] client.ConnectionManager$HConnectionImplementation: Checking master connection
com.google.protobuf.ServiceException: org.apache.hadoop.hbase.ipc.FailedServerException: This server is in the failed servers list: hscale-dev1-nn/10.60.70.10:16000
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient.callBlockingMethod(AbstractRpcClient.java:223)
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient$BlockingRpcChannelImplementation.callBlockingMethod(AbstractRpcClient.java:287)
    at org.apache.hadoop.hbase.protobuf.generated.MasterProtos$MasterService$BlockingStub.isMasterRunning(MasterProtos.java:58152)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation$MasterServiceState.isMasterRunning(ConnectionManager.java:1444)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.isKeepAliveMasterConnectedAndRunning(ConnectionManager.java:2099)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.getKeepAliveMasterService(ConnectionManager.java:1708)
    at org.apache.hadoop.hbase.client.MasterCallable.prepare(MasterCallable.java:38)
    at org.apache.hadoop.hbase.client.RpcRetryingCaller.callWithRetries(RpcRetryingCaller.java:124)
    at org.apache.hadoop.hbase.client.HBaseAdmin.executeCallable(HBaseAdmin.java:4083)
    at org.apache.hadoop.hbase.client.HBaseAdmin.getClusterStatus(HBaseAdmin.java:2765)
    at com.splicemachine.access.hbase.H10PartitionAdmin.allServers(H10PartitionAdmin.java:120)
    at com.splicemachine.hbase.HBaseRegionLoads.fetchRegionLoads(HBaseRegionLoads.java:164)
    at com.splicemachine.hbase.HBaseRegionLoads.access$000(HBaseRegionLoads.java:60)
    at com.splicemachine.hbase.HBaseRegionLoads$1.run(HBaseRegionLoads.java:81)
    at com.splicemachine.concurrent.LoggingScheduledThreadPoolExecutor$LoggingRunnable.run(LoggingScheduledThreadPoolExecutor.java:75)
    at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
    at java.util.concurrent.FutureTask.runAndReset(FutureTask.java:308)
    at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$301(ScheduledThreadPoolExecutor.java:180)
    at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:294)
    at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
    at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
    at java.lang.Thread.run(Thread.java:745)
Caused by: org.apache.hadoop.hbase.ipc.FailedServerException: This server is in the failed servers list: hscale-dev1-nn/10.60.70.10:16000
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.setupIOstreams(RpcClientImpl.java:701)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.writeRequest(RpcClientImpl.java:887)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.tracedWriteRequest(RpcClientImpl.java:856)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl.call(RpcClientImpl.java:1200)
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient.callBlockingMethod(AbstractRpcClient.java:213)
    ... 21 more
2016-12-02 19:21:05,033 WARN  [hbase-region-load-updater-0] client.ConnectionManager$HConnectionImplementation: Checking master connection
com.google.protobuf.ServiceException: org.apache.hadoop.hbase.ipc.FailedServerException: This server is in the failed servers list: hscale-dev1-nn/10.60.70.10:16000
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient.callBlockingMethod(AbstractRpcClient.java:223)
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient$BlockingRpcChannelImplementation.callBlockingMethod(AbstractRpcClient.java:287)
    at org.apache.hadoop.hbase.protobuf.generated.MasterProtos$MasterService$BlockingStub.isMasterRunning(MasterProtos.java:58152)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation$MasterServiceState.isMasterRunning(ConnectionManager.java:1444)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.isKeepAliveMasterConnectedAndRunning(ConnectionManager.java:2099)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.getKeepAliveMasterService(ConnectionManager.java:1708)
    at org.apache.hadoop.hbase.client.MasterCallable.prepare(MasterCallable.java:38)
    at org.apache.hadoop.hbase.client.RpcRetryingCaller.callWithRetries(RpcRetryingCaller.java:124)
    at org.apache.hadoop.hbase.client.HBaseAdmin.executeCallable(HBaseAdmin.java:4083)
    at org.apache.hadoop.hbase.client.HBaseAdmin.getClusterStatus(HBaseAdmin.java:2765)
    at com.splicemachine.access.hbase.H10PartitionAdmin.allServers(H10PartitionAdmin.java:120)
    at com.splicemachine.hbase.HBaseRegionLoads.fetchRegionLoads(HBaseRegionLoads.java:164)
    at com.splicemachine.hbase.HBaseRegionLoads.access$000(HBaseRegionLoads.java:60)
    at com.splicemachine.hbase.HBaseRegionLoads$1.run(HBaseRegionLoads.java:81)
    at com.splicemachine.concurrent.LoggingScheduledThreadPoolExecutor$LoggingRunnable.run(LoggingScheduledThreadPoolExecutor.java:75)
    at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
    at java.util.concurrent.FutureTask.runAndReset(FutureTask.java:308)
    at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$301(ScheduledThreadPoolExecutor.java:180)
    at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:294)
    at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
    at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
    at java.lang.Thread.run(Thread.java:745)
Caused by: org.apache.hadoop.hbase.ipc.FailedServerException: This server is in the failed servers list: hscale-dev1-nn/10.60.70.10:16000
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.setupIOstreams(RpcClientImpl.java:701)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.writeRequest(RpcClientImpl.java:887)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.tracedWriteRequest(RpcClientImpl.java:856)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl.call(RpcClientImpl.java:1200)
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient.callBlockingMethod(AbstractRpcClient.java:213)
    ... 21 more
2016-12-02 19:21:06,046 WARN  [hbase-region-load-updater-0] client.ConnectionManager$HConnectionImplementation: Checking master connection
com.google.protobuf.ServiceException: java.net.ConnectException: Connection refused
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient.callBlockingMethod(AbstractRpcClient.java:223)
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient$BlockingRpcChannelImplementation.callBlockingMethod(AbstractRpcClient.java:287)
    at org.apache.hadoop.hbase.protobuf.generated.MasterProtos$MasterService$BlockingStub.isMasterRunning(MasterProtos.java:58152)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation$MasterServiceState.isMasterRunning(ConnectionManager.java:1444)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.isKeepAliveMasterConnectedAndRunning(ConnectionManager.java:2099)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.getKeepAliveMasterService(ConnectionManager.java:1708)
    at org.apache.hadoop.hbase.client.MasterCallable.prepare(MasterCallable.java:38)
    at org.apache.hadoop.hbase.client.RpcRetryingCaller.callWithRetries(RpcRetryingCaller.java:124)
    at org.apache.hadoop.hbase.client.HBaseAdmin.executeCallable(HBaseAdmin.java:4083)
    at org.apache.hadoop.hbase.client.HBaseAdmin.getClusterStatus(HBaseAdmin.java:2765)
    at com.splicemachine.access.hbase.H10PartitionAdmin.allServers(H10PartitionAdmin.java:120)
    at com.splicemachine.hbase.HBaseRegionLoads.fetchRegionLoads(HBaseRegionLoads.java:164)
    at com.splicemachine.hbase.HBaseRegionLoads.access$000(HBaseRegionLoads.java:60)
    at com.splicemachine.hbase.HBaseRegionLoads$1.run(HBaseRegionLoads.java:81)
    at com.splicemachine.concurrent.LoggingScheduledThreadPoolExecutor$LoggingRunnable.run(LoggingScheduledThreadPoolExecutor.java:75)
    at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
    at java.util.concurrent.FutureTask.runAndReset(FutureTask.java:308)
    at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$301(ScheduledThreadPoolExecutor.java:180)
    at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:294)
    at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
    at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
    at java.lang.Thread.run(Thread.java:745)
Caused by: java.net.ConnectException: Connection refused
    at sun.nio.ch.SocketChannelImpl.checkConnect(Native Method)
    at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:717)
    at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:206)
    at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:531)
    at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:495)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.setupConnection(RpcClientImpl.java:410)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.setupIOstreams(RpcClientImpl.java:716)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.writeRequest(RpcClientImpl.java:887)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.tracedWriteRequest(RpcClientImpl.java:856)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl.call(RpcClientImpl.java:1200)
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient.callBlockingMethod(AbstractRpcClient.java:213)
    ... 21 more
2016-12-02 19:21:06,050 ERROR [hbase-region-load-updater-0] hbase.HBaseRegionLoads: Unable to fetch region load info
org.apache.hadoop.hbase.client.RetriesExhaustedException: Failed after attempts=5, exceptions:
Fri Dec 02 19:21:04 IST 2016, RpcRetryingCaller{globalStartTime=1480686664019, pause=100, retries=5}, org.apache.hadoop.hbase.MasterNotRunningException: java.io.IOException: Can't get master address from ZooKeeper; znode data == null
Fri Dec 02 19:21:04 IST 2016, RpcRetryingCaller{globalStartTime=1480686664019, pause=100, retries=5}, org.apache.hadoop.hbase.MasterNotRunningException: java.io.IOException: Can't get master address from ZooKeeper; znode data == null
Fri Dec 02 19:21:04 IST 2016, RpcRetryingCaller{globalStartTime=1480686664019, pause=100, retries=5}, org.apache.hadoop.hbase.MasterNotRunningException: java.io.IOException: Can't get master address from ZooKeeper; znode data == null
Fri Dec 02 19:21:05 IST 2016, RpcRetryingCaller{globalStartTime=1480686664019, pause=100, retries=5}, org.apache.hadoop.hbase.MasterNotRunningException: java.io.IOException: Can't get master address from ZooKeeper; znode data == null
Fri Dec 02 19:21:06 IST 2016, RpcRetryingCaller{globalStartTime=1480686664019, pause=100, retries=5}, org.apache.hadoop.hbase.MasterNotRunningException: java.io.IOException: Can't get master address from ZooKeeper; znode data == null

    at org.apache.hadoop.hbase.client.RpcRetryingCaller.callWithRetries(RpcRetryingCaller.java:147)
    at org.apache.hadoop.hbase.client.HBaseAdmin.executeCallable(HBaseAdmin.java:4083)
    at org.apache.hadoop.hbase.client.HBaseAdmin.getClusterStatus(HBaseAdmin.java:2765)
    at com.splicemachine.access.hbase.H10PartitionAdmin.allServers(H10PartitionAdmin.java:120)
    at com.splicemachine.hbase.HBaseRegionLoads.fetchRegionLoads(HBaseRegionLoads.java:164)
    at com.splicemachine.hbase.HBaseRegionLoads.access$000(HBaseRegionLoads.java:60)
    at com.splicemachine.hbase.HBaseRegionLoads$1.run(HBaseRegionLoads.java:81)
    at com.splicemachine.concurrent.LoggingScheduledThreadPoolExecutor$LoggingRunnable.run(LoggingScheduledThreadPoolExecutor.java:75)
    at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
    at java.util.concurrent.FutureTask.runAndReset(FutureTask.java:308)
    at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$301(ScheduledThreadPoolExecutor.java:180)
    at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:294)
    at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
    at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
    at java.lang.Thread.run(Thread.java:745)
Caused by: org.apache.hadoop.hbase.MasterNotRunningException: java.io.IOException: Can't get master address from ZooKeeper; znode data == null
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation$StubMaker.makeStub(ConnectionManager.java:1540)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation$MasterServiceStubMaker.makeStub(ConnectionManager.java:1560)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.getKeepAliveMasterService(ConnectionManager.java:1711)
    at org.apache.hadoop.hbase.client.MasterCallable.prepare(MasterCallable.java:38)
    at org.apache.hadoop.hbase.client.RpcRetryingCaller.callWithRetries(RpcRetryingCaller.java:124)
    ... 14 more
Caused by: java.io.IOException: Can't get master address from ZooKeeper; znode data == null
    at org.apache.hadoop.hbase.zookeeper.MasterAddressTracker.getMasterAddress(MasterAddressTracker.java:154)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation$StubMaker.makeStubNoRetries(ConnectionManager.java:1491)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation$StubMaker.makeStub(ConnectionManager.java:1531)
    ... 18 more
2016-12-02 19:21:09,020 WARN  [hbase-region-load-updater-0] client.ConnectionManager$HConnectionImplementation: Checking master connection
com.google.protobuf.ServiceException: java.net.ConnectException: Connection refused
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient.callBlockingMethod(AbstractRpcClient.java:223)
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient$BlockingRpcChannelImplementation.callBlockingMethod(AbstractRpcClient.java:287)
    at org.apache.hadoop.hbase.protobuf.generated.MasterProtos$MasterService$BlockingStub.isMasterRunning(MasterProtos.java:58152)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation$MasterServiceState.isMasterRunning(ConnectionManager.java:1444)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.isKeepAliveMasterConnectedAndRunning(ConnectionManager.java:2099)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.getKeepAliveMasterService(ConnectionManager.java:1708)
    at org.apache.hadoop.hbase.client.MasterCallable.prepare(MasterCallable.java:38)
    at org.apache.hadoop.hbase.client.RpcRetryingCaller.callWithRetries(RpcRetryingCaller.java:124)
    at org.apache.hadoop.hbase.client.HBaseAdmin.executeCallable(HBaseAdmin.java:4083)
    at org.apache.hadoop.hbase.client.HBaseAdmin.getClusterStatus(HBaseAdmin.java:2765)
    at com.splicemachine.access.hbase.H10PartitionAdmin.allServers(H10PartitionAdmin.java:120)
    at com.splicemachine.hbase.HBaseRegionLoads.fetchRegionLoads(HBaseRegionLoads.java:164)
    at com.splicemachine.hbase.HBaseRegionLoads.access$000(HBaseRegionLoads.java:60)
    at com.splicemachine.hbase.HBaseRegionLoads$1.run(HBaseRegionLoads.java:81)
    at com.splicemachine.concurrent.LoggingScheduledThreadPoolExecutor$LoggingRunnable.run(LoggingScheduledThreadPoolExecutor.java:75)
    at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
    at java.util.concurrent.FutureTask.runAndReset(FutureTask.java:308)
    at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$301(ScheduledThreadPoolExecutor.java:180)
    at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:294)
    at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
    at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
    at java.lang.Thread.run(Thread.java:745)
Caused by: java.net.ConnectException: Connection refused
    at sun.nio.ch.SocketChannelImpl.checkConnect(Native Method)
    at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:717)
    at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:206)
    at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:531)
    at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:495)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.setupConnection(RpcClientImpl.java:410)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.setupIOstreams(RpcClientImpl.java:716)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.writeRequest(RpcClientImpl.java:887)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.tracedWriteRequest(RpcClientImpl.java:856)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl.call(RpcClientImpl.java:1200)
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient.callBlockingMethod(AbstractRpcClient.java:213)
    ... 21 more
2016-12-02 19:21:09,224 WARN  [hbase-region-load-updater-0] client.ConnectionManager$HConnectionImplementation: Checking master connection
com.google.protobuf.ServiceException: org.apache.hadoop.hbase.ipc.FailedServerException: This server is in the failed servers list: hscale-dev1-nn/10.60.70.10:16000
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient.callBlockingMethod(AbstractRpcClient.java:223)
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient$BlockingRpcChannelImplementation.callBlockingMethod(AbstractRpcClient.java:287)
    at org.apache.hadoop.hbase.protobuf.generated.MasterProtos$MasterService$BlockingStub.isMasterRunning(MasterProtos.java:58152)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation$MasterServiceState.isMasterRunning(ConnectionManager.java:1444)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.isKeepAliveMasterConnectedAndRunning(ConnectionManager.java:2099)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.getKeepAliveMasterService(ConnectionManager.java:1708)
    at org.apache.hadoop.hbase.client.MasterCallable.prepare(MasterCallable.java:38)
    at org.apache.hadoop.hbase.client.RpcRetryingCaller.callWithRetries(RpcRetryingCaller.java:124)
    at org.apache.hadoop.hbase.client.HBaseAdmin.executeCallable(HBaseAdmin.java:4083)
    at org.apache.hadoop.hbase.client.HBaseAdmin.getClusterStatus(HBaseAdmin.java:2765)
    at com.splicemachine.access.hbase.H10PartitionAdmin.allServers(H10PartitionAdmin.java:120)
    at com.splicemachine.hbase.HBaseRegionLoads.fetchRegionLoads(HBaseRegionLoads.java:164)
    at com.splicemachine.hbase.HBaseRegionLoads.access$000(HBaseRegionLoads.java:60)
    at com.splicemachine.hbase.HBaseRegionLoads$1.run(HBaseRegionLoads.java:81)
    at com.splicemachine.concurrent.LoggingScheduledThreadPoolExecutor$LoggingRunnable.run(LoggingScheduledThreadPoolExecutor.java:75)
    at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
    at java.util.concurrent.FutureTask.runAndReset(FutureTask.java:308)
    at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$301(ScheduledThreadPoolExecutor.java:180)
    at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:294)
    at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
    at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
    at java.lang.Thread.run(Thread.java:745)
Caused by: org.apache.hadoop.hbase.ipc.FailedServerException: This server is in the failed servers list: hscale-dev1-nn/10.60.70.10:16000
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.setupIOstreams(RpcClientImpl.java:701)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.writeRequest(RpcClientImpl.java:887)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.tracedWriteRequest(RpcClientImpl.java:856)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl.call(RpcClientImpl.java:1200)
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient.callBlockingMethod(AbstractRpcClient.java:213)
    ... 21 more
2016-12-02 19:21:09,529 WARN  [hbase-region-load-updater-0] client.ConnectionManager$HConnectionImplementation: Checking master connection
com.google.protobuf.ServiceException: org.apache.hadoop.hbase.ipc.FailedServerException: This server is in the failed servers list: hscale-dev1-nn/10.60.70.10:16000
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient.callBlockingMethod(AbstractRpcClient.java:223)
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient$BlockingRpcChannelImplementation.callBlockingMethod(AbstractRpcClient.java:287)
    at org.apache.hadoop.hbase.protobuf.generated.MasterProtos$MasterService$BlockingStub.isMasterRunning(MasterProtos.java:58152)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation$MasterServiceState.isMasterRunning(ConnectionManager.java:1444)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.isKeepAliveMasterConnectedAndRunning(ConnectionManager.java:2099)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.getKeepAliveMasterService(ConnectionManager.java:1708)
    at org.apache.hadoop.hbase.client.MasterCallable.prepare(MasterCallable.java:38)
    at org.apache.hadoop.hbase.client.RpcRetryingCaller.callWithRetries(RpcRetryingCaller.java:124)
    at org.apache.hadoop.hbase.client.HBaseAdmin.executeCallable(HBaseAdmin.java:4083)
    at org.apache.hadoop.hbase.client.HBaseAdmin.getClusterStatus(HBaseAdmin.java:2765)
    at com.splicemachine.access.hbase.H10PartitionAdmin.allServers(H10PartitionAdmin.java:120)
    at com.splicemachine.hbase.HBaseRegionLoads.fetchRegionLoads(HBaseRegionLoads.java:164)
    at com.splicemachine.hbase.HBaseRegionLoads.access$000(HBaseRegionLoads.java:60)
    at com.splicemachine.hbase.HBaseRegionLoads$1.run(HBaseRegionLoads.java:81)
    at com.splicemachine.concurrent.LoggingScheduledThreadPoolExecutor$LoggingRunnable.run(LoggingScheduledThreadPoolExecutor.java:75)
    at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
    at java.util.concurrent.FutureTask.runAndReset(FutureTask.java:308)
    at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$301(ScheduledThreadPoolExecutor.java:180)
    at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:294)
    at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
    at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
    at java.lang.Thread.run(Thread.java:745)
Caused by: org.apache.hadoop.hbase.ipc.FailedServerException: This server is in the failed servers list: hscale-dev1-nn/10.60.70.10:16000
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.setupIOstreams(RpcClientImpl.java:701)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.writeRequest(RpcClientImpl.java:887)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.tracedWriteRequest(RpcClientImpl.java:856)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl.call(RpcClientImpl.java:1200)
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient.callBlockingMethod(AbstractRpcClient.java:213)
    ... 21 more
2016-12-02 19:21:10,035 WARN  [hbase-region-load-updater-0] client.ConnectionManager$HConnectionImplementation: Checking master connection
com.google.protobuf.ServiceException: org.apache.hadoop.hbase.ipc.FailedServerException: This server is in the failed servers list: hscale-dev1-nn/10.60.70.10:16000
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient.callBlockingMethod(AbstractRpcClient.java:223)
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient$BlockingRpcChannelImplementation.callBlockingMethod(AbstractRpcClient.java:287)
    at org.apache.hadoop.hbase.protobuf.generated.MasterProtos$MasterService$BlockingStub.isMasterRunning(MasterProtos.java:58152)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation$MasterServiceState.isMasterRunning(ConnectionManager.java:1444)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.isKeepAliveMasterConnectedAndRunning(ConnectionManager.java:2099)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.getKeepAliveMasterService(ConnectionManager.java:1708)
    at org.apache.hadoop.hbase.client.MasterCallable.prepare(MasterCallable.java:38)
    at org.apache.hadoop.hbase.client.RpcRetryingCaller.callWithRetries(RpcRetryingCaller.java:124)
    at org.apache.hadoop.hbase.client.HBaseAdmin.executeCallable(HBaseAdmin.java:4083)
    at org.apache.hadoop.hbase.client.HBaseAdmin.getClusterStatus(HBaseAdmin.java:2765)
    at com.splicemachine.access.hbase.H10PartitionAdmin.allServers(H10PartitionAdmin.java:120)
    at com.splicemachine.hbase.HBaseRegionLoads.fetchRegionLoads(HBaseRegionLoads.java:164)
    at com.splicemachine.hbase.HBaseRegionLoads.access$000(HBaseRegionLoads.java:60)
    at com.splicemachine.hbase.HBaseRegionLoads$1.run(HBaseRegionLoads.java:81)
    at com.splicemachine.concurrent.LoggingScheduledThreadPoolExecutor$LoggingRunnable.run(LoggingScheduledThreadPoolExecutor.java:75)
    at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
    at java.util.concurrent.FutureTask.runAndReset(FutureTask.java:308)
    at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$301(ScheduledThreadPoolExecutor.java:180)
    at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:294)
    at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
    at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
    at java.lang.Thread.run(Thread.java:745)
Caused by: org.apache.hadoop.hbase.ipc.FailedServerException: This server is in the failed servers list: hscale-dev1-nn/10.60.70.10:16000
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.setupIOstreams(RpcClientImpl.java:701)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.writeRequest(RpcClientImpl.java:887)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.tracedWriteRequest(RpcClientImpl.java:856)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl.call(RpcClientImpl.java:1200)
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient.callBlockingMethod(AbstractRpcClient.java:213)
    ... 21 more
2016-12-02 19:21:11,837 WARN  [hbase-region-load-updater-0] client.ConnectionManager$HConnectionImplementation: Checking master connection
com.google.protobuf.ServiceException: org.apache.hadoop.hbase.ipc.RemoteWithExtrasException(org.apache.hadoop.hbase.ipc.ServerNotRunningYetException): org.apache.hadoop.hbase.ipc.ServerNotRunningYetException: Server is not running yet
    at org.apache.hadoop.hbase.master.HMaster.checkServiceStarted(HMaster.java:2317)
    at org.apache.hadoop.hbase.master.MasterRpcServices.isMasterRunning(MasterRpcServices.java:924)
    at org.apache.hadoop.hbase.protobuf.generated.MasterProtos$MasterService$2.callBlockingMethod(MasterProtos.java:55373)
    at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2114)
    at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:101)
    at org.apache.hadoop.hbase.ipc.RpcExecutor.consumerLoop(RpcExecutor.java:130)
    at org.apache.hadoop.hbase.ipc.RpcExecutor$1.run(RpcExecutor.java:107)
    at java.lang.Thread.run(Thread.java:745)

    at org.apache.hadoop.hbase.ipc.AbstractRpcClient.callBlockingMethod(AbstractRpcClient.java:223)
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient$BlockingRpcChannelImplementation.callBlockingMethod(AbstractRpcClient.java:287)
    at org.apache.hadoop.hbase.protobuf.generated.MasterProtos$MasterService$BlockingStub.isMasterRunning(MasterProtos.java:58152)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation$MasterServiceState.isMasterRunning(ConnectionManager.java:1444)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.isKeepAliveMasterConnectedAndRunning(ConnectionManager.java:2099)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.getKeepAliveMasterService(ConnectionManager.java:1708)
    at org.apache.hadoop.hbase.client.MasterCallable.prepare(MasterCallable.java:38)
    at org.apache.hadoop.hbase.client.RpcRetryingCaller.callWithRetries(RpcRetryingCaller.java:124)
    at org.apache.hadoop.hbase.client.HBaseAdmin.executeCallable(HBaseAdmin.java:4083)
    at org.apache.hadoop.hbase.client.HBaseAdmin.getClusterStatus(HBaseAdmin.java:2765)
    at com.splicemachine.access.hbase.H10PartitionAdmin.allServers(H10PartitionAdmin.java:120)
    at com.splicemachine.hbase.HBaseRegionLoads.fetchRegionLoads(HBaseRegionLoads.java:164)
    at com.splicemachine.hbase.HBaseRegionLoads.access$000(HBaseRegionLoads.java:60)
    at com.splicemachine.hbase.HBaseRegionLoads$1.run(HBaseRegionLoads.java:81)
    at com.splicemachine.concurrent.LoggingScheduledThreadPoolExecutor$LoggingRunnable.run(LoggingScheduledThreadPoolExecutor.java:75)
    at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
    at java.util.concurrent.FutureTask.runAndReset(FutureTask.java:308)
    at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$301(ScheduledThreadPoolExecutor.java:180)
    at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:294)
    at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
    at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
    at java.lang.Thread.run(Thread.java:745)
Caused by: org.apache.hadoop.hbase.ipc.RemoteWithExtrasException(org.apache.hadoop.hbase.ipc.ServerNotRunningYetException): org.apache.hadoop.hbase.ipc.ServerNotRunningYetException: Server is not running yet
    at org.apache.hadoop.hbase.master.HMaster.checkServiceStarted(HMaster.java:2317)
    at org.apache.hadoop.hbase.master.MasterRpcServices.isMasterRunning(MasterRpcServices.java:924)
    at org.apache.hadoop.hbase.protobuf.generated.MasterProtos$MasterService$2.callBlockingMethod(MasterProtos.java:55373)
    at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2114)
    at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:101)
    at org.apache.hadoop.hbase.ipc.RpcExecutor.consumerLoop(RpcExecutor.java:130)
    at org.apache.hadoop.hbase.ipc.RpcExecutor$1.run(RpcExecutor.java:107)
    at java.lang.Thread.run(Thread.java:745)

    at org.apache.hadoop.hbase.ipc.RpcClientImpl.call(RpcClientImpl.java:1226)
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient.callBlockingMethod(AbstractRpcClient.java:213)
    ... 21 more
2016-12-02 19:21:11,841 ERROR [hbase-region-load-updater-0] hbase.HBaseRegionLoads: Unable to fetch region load info
org.apache.hadoop.hbase.client.RetriesExhaustedException: Failed after attempts=5, exceptions:
Fri Dec 02 19:21:09 IST 2016, RpcRetryingCaller{globalStartTime=1480686669019, pause=100, retries=5}, org.apache.hadoop.hbase.MasterNotRunningException: java.io.IOException: Can't get master address from ZooKeeper; znode data == null
Fri Dec 02 19:21:09 IST 2016, RpcRetryingCaller{globalStartTime=1480686669019, pause=100, retries=5}, org.apache.hadoop.hbase.MasterNotRunningException: java.io.IOException: Can't get master address from ZooKeeper; znode data == null
Fri Dec 02 19:21:09 IST 2016, RpcRetryingCaller{globalStartTime=1480686669019, pause=100, retries=5}, org.apache.hadoop.hbase.MasterNotRunningException: java.io.IOException: Can't get master address from ZooKeeper; znode data == null
Fri Dec 02 19:21:10 IST 2016, RpcRetryingCaller{globalStartTime=1480686669019, pause=100, retries=5}, org.apache.hadoop.hbase.MasterNotRunningException: java.io.IOException: Can't get master address from ZooKeeper; znode data == null
Fri Dec 02 19:21:11 IST 2016, RpcRetryingCaller{globalStartTime=1480686669019, pause=100, retries=5}, org.apache.hadoop.hbase.MasterNotRunningException: java.io.IOException: Can't get master address from ZooKeeper; znode data == null

    at org.apache.hadoop.hbase.client.RpcRetryingCaller.callWithRetries(RpcRetryingCaller.java:147)
    at org.apache.hadoop.hbase.client.HBaseAdmin.executeCallable(HBaseAdmin.java:4083)
    at org.apache.hadoop.hbase.client.HBaseAdmin.getClusterStatus(HBaseAdmin.java:2765)
    at com.splicemachine.access.hbase.H10PartitionAdmin.allServers(H10PartitionAdmin.java:120)
    at com.splicemachine.hbase.HBaseRegionLoads.fetchRegionLoads(HBaseRegionLoads.java:164)
    at com.splicemachine.hbase.HBaseRegionLoads.access$000(HBaseRegionLoads.java:60)
    at com.splicemachine.hbase.HBaseRegionLoads$1.run(HBaseRegionLoads.java:81)
    at com.splicemachine.concurrent.LoggingScheduledThreadPoolExecutor$LoggingRunnable.run(LoggingScheduledThreadPoolExecutor.java:75)
    at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
    at java.util.concurrent.FutureTask.runAndReset(FutureTask.java:308)
    at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$301(ScheduledThreadPoolExecutor.java:180)
    at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:294)
    at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
    at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
    at java.lang.Thread.run(Thread.java:745)
Caused by: org.apache.hadoop.hbase.MasterNotRunningException: java.io.IOException: Can't get master address from ZooKeeper; znode data == null
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation$StubMaker.makeStub(ConnectionManager.java:1540)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation$MasterServiceStubMaker.makeStub(ConnectionManager.java:1560)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.getKeepAliveMasterService(ConnectionManager.java:1711)
    at org.apache.hadoop.hbase.client.MasterCallable.prepare(MasterCallable.java:38)
    at org.apache.hadoop.hbase.client.RpcRetryingCaller.callWithRetries(RpcRetryingCaller.java:124)
    ... 14 more
Caused by: java.io.IOException: Can't get master address from ZooKeeper; znode data == null
    at org.apache.hadoop.hbase.zookeeper.MasterAddressTracker.getMasterAddress(MasterAddressTracker.java:154)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation$StubMaker.makeStubNoRetries(ConnectionManager.java:1491)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation$StubMaker.makeStub(ConnectionManager.java:1531)
    ... 18 more
2016-12-02 19:21:14,020 WARN  [hbase-region-load-updater-0] client.ConnectionManager$HConnectionImplementation: Checking master connection
com.google.protobuf.ServiceException: org.apache.hadoop.hbase.exceptions.ConnectionClosingException: Call to hscale-dev1-nn/10.60.70.10:16000 failed on local exception: org.apache.hadoop.hbase.exceptions.ConnectionClosingException: Connection to hscale-dev1-nn/10.60.70.10:16000 is closing. Call id=85, waitTime=0
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient.callBlockingMethod(AbstractRpcClient.java:223)
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient$BlockingRpcChannelImplementation.callBlockingMethod(AbstractRpcClient.java:287)
    at org.apache.hadoop.hbase.protobuf.generated.MasterProtos$MasterService$BlockingStub.isMasterRunning(MasterProtos.java:58152)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation$MasterServiceState.isMasterRunning(ConnectionManager.java:1444)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.isKeepAliveMasterConnectedAndRunning(ConnectionManager.java:2099)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.getKeepAliveMasterService(ConnectionManager.java:1708)
    at org.apache.hadoop.hbase.client.MasterCallable.prepare(MasterCallable.java:38)
    at org.apache.hadoop.hbase.client.RpcRetryingCaller.callWithRetries(RpcRetryingCaller.java:124)
    at org.apache.hadoop.hbase.client.HBaseAdmin.executeCallable(HBaseAdmin.java:4083)
    at org.apache.hadoop.hbase.client.HBaseAdmin.getClusterStatus(HBaseAdmin.java:2765)
    at com.splicemachine.access.hbase.H10PartitionAdmin.allServers(H10PartitionAdmin.java:120)
    at com.splicemachine.hbase.HBaseRegionLoads.fetchRegionLoads(HBaseRegionLoads.java:164)
    at com.splicemachine.hbase.HBaseRegionLoads.access$000(HBaseRegionLoads.java:60)
    at com.splicemachine.hbase.HBaseRegionLoads$1.run(HBaseRegionLoads.java:81)
    at com.splicemachine.concurrent.LoggingScheduledThreadPoolExecutor$LoggingRunnable.run(LoggingScheduledThreadPoolExecutor.java:75)
    at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
    at java.util.concurrent.FutureTask.runAndReset(FutureTask.java:308)
    at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$301(ScheduledThreadPoolExecutor.java:180)
    at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:294)
    at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
    at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
    at java.lang.Thread.run(Thread.java:745)
Caused by: org.apache.hadoop.hbase.exceptions.ConnectionClosingException: Call to hscale-dev1-nn/10.60.70.10:16000 failed on local exception: org.apache.hadoop.hbase.exceptions.ConnectionClosingException: Connection to hscale-dev1-nn/10.60.70.10:16000 is closing. Call id=85, waitTime=0
    at org.apache.hadoop.hbase.ipc.RpcClientImpl.wrapException(RpcClientImpl.java:1259)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl.call(RpcClientImpl.java:1230)
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient.callBlockingMethod(AbstractRpcClient.java:213)
    ... 21 more
Caused by: org.apache.hadoop.hbase.exceptions.ConnectionClosingException: Connection to hscale-dev1-nn/10.60.70.10:16000 is closing. Call id=85, waitTime=0
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.cleanupCalls(RpcClientImpl.java:1047)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.close(RpcClientImpl.java:846)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.run(RpcClientImpl.java:574)
2016-12-02 19:21:14,109 WARN  [pool-10-thread-1] client.ConnectionManager$HConnectionImplementation: Checking master connection
com.google.protobuf.ServiceException: org.apache.hadoop.hbase.exceptions.ConnectionClosingException: Call to hscale-dev1-nn/10.60.70.10:16000 failed on local exception: org.apache.hadoop.hbase.exceptions.ConnectionClosingException: Connection to hscale-dev1-nn/10.60.70.10:16000 is closing. Call id=87, waitTime=0
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient.callBlockingMethod(AbstractRpcClient.java:223)
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient$BlockingRpcChannelImplementation.callBlockingMethod(AbstractRpcClient.java:287)
    at org.apache.hadoop.hbase.protobuf.generated.MasterProtos$MasterService$BlockingStub.isMasterRunning(MasterProtos.java:58152)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation$MasterServiceState.isMasterRunning(ConnectionManager.java:1444)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.isKeepAliveMasterConnectedAndRunning(ConnectionManager.java:2099)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.getKeepAliveMasterService(ConnectionManager.java:1708)
    at org.apache.hadoop.hbase.client.MasterCallable.prepare(MasterCallable.java:38)
    at org.apache.hadoop.hbase.client.RpcRetryingCaller.callWithRetries(RpcRetryingCaller.java:124)
    at org.apache.hadoop.hbase.client.HBaseAdmin.executeCallable(HBaseAdmin.java:4083)
    at org.apache.hadoop.hbase.client.HBaseAdmin.access$700(HBaseAdmin.java:194)
    at org.apache.hadoop.hbase.client.HBaseAdmin$ProcedureFuture.getProcedureResult(HBaseAdmin.java:4383)
    at org.apache.hadoop.hbase.client.HBaseAdmin$ProcedureFuture.waitProcedureResult(HBaseAdmin.java:4335)
    at org.apache.hadoop.hbase.client.HBaseAdmin$ProcedureFuture.get(HBaseAdmin.java:4291)
    at org.apache.hadoop.hbase.client.HBaseAdmin.createTable(HBaseAdmin.java:647)
    at org.apache.hadoop.hbase.client.HBaseAdmin.createTable(HBaseAdmin.java:577)
    at com.splicemachine.lifecycle.RegionServerLifecycle.distributedStart(RegionServerLifecycle.java:66)
    at com.splicemachine.derby.lifecycle.EngineLifecycleService.start(EngineLifecycleService.java:81)
    at com.splicemachine.lifecycle.DatabaseLifecycleManager$Startup.bootServices(DatabaseLifecycleManager.java:229)
    at com.splicemachine.lifecycle.DatabaseLifecycleManager$Startup.run(DatabaseLifecycleManager.java:220)
    at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
    at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
    at java.lang.Thread.run(Thread.java:745)
Caused by: org.apache.hadoop.hbase.exceptions.ConnectionClosingException: Call to hscale-dev1-nn/10.60.70.10:16000 failed on local exception: org.apache.hadoop.hbase.exceptions.ConnectionClosingException: Connection to hscale-dev1-nn/10.60.70.10:16000 is closing. Call id=87, waitTime=0
    at org.apache.hadoop.hbase.ipc.RpcClientImpl.wrapException(RpcClientImpl.java:1259)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl.call(RpcClientImpl.java:1230)
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient.callBlockingMethod(AbstractRpcClient.java:213)
    ... 21 more
Caused by: org.apache.hadoop.hbase.exceptions.ConnectionClosingException: Connection to hscale-dev1-nn/10.60.70.10:16000 is closing. Call id=87, waitTime=0
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.cleanupCalls(RpcClientImpl.java:1047)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.close(RpcClientImpl.java:846)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.run(RpcClientImpl.java:574)
2016-12-02 19:21:14,234 WARN  [hbase-region-load-updater-0] client.ConnectionManager$HConnectionImplementation: Checking master connection
com.google.protobuf.ServiceException: org.apache.hadoop.hbase.exceptions.ConnectionClosingException: Call to hscale-dev1-nn/10.60.70.10:16000 failed on local exception: org.apache.hadoop.hbase.exceptions.ConnectionClosingException: Connection to hscale-dev1-nn/10.60.70.10:16000 is closing. Call id=89, waitTime=1
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient.callBlockingMethod(AbstractRpcClient.java:223)
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient$BlockingRpcChannelImplementation.callBlockingMethod(AbstractRpcClient.java:287)
    at org.apache.hadoop.hbase.protobuf.generated.MasterProtos$MasterService$BlockingStub.isMasterRunning(MasterProtos.java:58152)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation$MasterServiceState.isMasterRunning(ConnectionManager.java:1444)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.isKeepAliveMasterConnectedAndRunning(ConnectionManager.java:2099)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.getKeepAliveMasterService(ConnectionManager.java:1708)
    at org.apache.hadoop.hbase.client.MasterCallable.prepare(MasterCallable.java:38)
    at org.apache.hadoop.hbase.client.RpcRetryingCaller.callWithRetries(RpcRetryingCaller.java:124)
    at org.apache.hadoop.hbase.client.HBaseAdmin.executeCallable(HBaseAdmin.java:4083)
    at org.apache.hadoop.hbase.client.HBaseAdmin.getClusterStatus(HBaseAdmin.java:2765)
    at com.splicemachine.access.hbase.H10PartitionAdmin.allServers(H10PartitionAdmin.java:120)
    at com.splicemachine.hbase.HBaseRegionLoads.fetchRegionLoads(HBaseRegionLoads.java:164)
    at com.splicemachine.hbase.HBaseRegionLoads.access$000(HBaseRegionLoads.java:60)
    at com.splicemachine.hbase.HBaseRegionLoads$1.run(HBaseRegionLoads.java:81)
    at com.splicemachine.concurrent.LoggingScheduledThreadPoolExecutor$LoggingRunnable.run(LoggingScheduledThreadPoolExecutor.java:75)
    at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
    at java.util.concurrent.FutureTask.runAndReset(FutureTask.java:308)
    at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$301(ScheduledThreadPoolExecutor.java:180)
    at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:294)
    at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
    at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
    at java.lang.Thread.run(Thread.java:745)
Caused by: org.apache.hadoop.hbase.exceptions.ConnectionClosingException: Call to hscale-dev1-nn/10.60.70.10:16000 failed on local exception: org.apache.hadoop.hbase.exceptions.ConnectionClosingException: Connection to hscale-dev1-nn/10.60.70.10:16000 is closing. Call id=89, waitTime=1
    at org.apache.hadoop.hbase.ipc.RpcClientImpl.wrapException(RpcClientImpl.java:1259)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl.call(RpcClientImpl.java:1230)
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient.callBlockingMethod(AbstractRpcClient.java:213)
    ... 21 more
Caused by: org.apache.hadoop.hbase.exceptions.ConnectionClosingException: Connection to hscale-dev1-nn/10.60.70.10:16000 is closing. Call id=89, waitTime=1
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.cleanupCalls(RpcClientImpl.java:1047)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.close(RpcClientImpl.java:846)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.run(RpcClientImpl.java:574)
2016-12-02 19:21:14,319 WARN  [pool-10-thread-1] client.ConnectionManager$HConnectionImplementation: Checking master connection
com.google.protobuf.ServiceException: org.apache.hadoop.hbase.exceptions.ConnectionClosingException: Call to hscale-dev1-nn/10.60.70.10:16000 failed on local exception: org.apache.hadoop.hbase.exceptions.ConnectionClosingException: Connection to hscale-dev1-nn/10.60.70.10:16000 is closing. Call id=91, waitTime=0
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient.callBlockingMethod(AbstractRpcClient.java:223)
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient$BlockingRpcChannelImplementation.callBlockingMethod(AbstractRpcClient.java:287)
    at org.apache.hadoop.hbase.protobuf.generated.MasterProtos$MasterService$BlockingStub.isMasterRunning(MasterProtos.java:58152)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation$MasterServiceState.isMasterRunning(ConnectionManager.java:1444)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.isKeepAliveMasterConnectedAndRunning(ConnectionManager.java:2099)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.getKeepAliveMasterService(ConnectionManager.java:1708)
    at org.apache.hadoop.hbase.client.MasterCallable.prepare(MasterCallable.java:38)
    at org.apache.hadoop.hbase.client.RpcRetryingCaller.callWithRetries(RpcRetryingCaller.java:124)
    at org.apache.hadoop.hbase.client.HBaseAdmin.executeCallable(HBaseAdmin.java:4083)
    at org.apache.hadoop.hbase.client.HBaseAdmin.access$700(HBaseAdmin.java:194)
    at org.apache.hadoop.hbase.client.HBaseAdmin$ProcedureFuture.getProcedureResult(HBaseAdmin.java:4383)
    at org.apache.hadoop.hbase.client.HBaseAdmin$ProcedureFuture.waitProcedureResult(HBaseAdmin.java:4335)
    at org.apache.hadoop.hbase.client.HBaseAdmin$ProcedureFuture.get(HBaseAdmin.java:4291)
    at org.apache.hadoop.hbase.client.HBaseAdmin.createTable(HBaseAdmin.java:647)
    at org.apache.hadoop.hbase.client.HBaseAdmin.createTable(HBaseAdmin.java:577)
    at com.splicemachine.lifecycle.RegionServerLifecycle.distributedStart(RegionServerLifecycle.java:66)
    at com.splicemachine.derby.lifecycle.EngineLifecycleService.start(EngineLifecycleService.java:81)
    at com.splicemachine.lifecycle.DatabaseLifecycleManager$Startup.bootServices(DatabaseLifecycleManager.java:229)
    at com.splicemachine.lifecycle.DatabaseLifecycleManager$Startup.run(DatabaseLifecycleManager.java:220)
    at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
    at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
    at java.lang.Thread.run(Thread.java:745)
Caused by: org.apache.hadoop.hbase.exceptions.ConnectionClosingException: Call to hscale-dev1-nn/10.60.70.10:16000 failed on local exception: org.apache.hadoop.hbase.exceptions.ConnectionClosingException: Connection to hscale-dev1-nn/10.60.70.10:16000 is closing. Call id=91, waitTime=0
    at org.apache.hadoop.hbase.ipc.RpcClientImpl.wrapException(RpcClientImpl.java:1259)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl.call(RpcClientImpl.java:1230)
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient.callBlockingMethod(AbstractRpcClient.java:213)
    ... 21 more
Caused by: org.apache.hadoop.hbase.exceptions.ConnectionClosingException: Connection to hscale-dev1-nn/10.60.70.10:16000 is closing. Call id=91, waitTime=0
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.cleanupCalls(RpcClientImpl.java:1047)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.close(RpcClientImpl.java:846)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.run(RpcClientImpl.java:574)
2016-12-02 19:21:16,083 ERROR [hbase-region-load-updater-0] hbase.HBaseRegionLoads: Unable to fetch region load info
org.apache.hadoop.hbase.client.RetriesExhaustedException: Failed after attempts=5, exceptions:
Fri Dec 02 19:21:14 IST 2016, RpcRetryingCaller{globalStartTime=1480686674019, pause=100, retries=5}, org.apache.hadoop.hbase.ipc.RemoteWithExtrasException(org.apache.hadoop.hbase.PleaseHoldException): org.apache.hadoop.hbase.PleaseHoldException: Master is initializing
    at org.apache.hadoop.hbase.master.HMaster.checkInitialized(HMaster.java:2324)
    at org.apache.hadoop.hbase.master.MasterRpcServices.getClusterStatus(MasterRpcServices.java:770)
    at org.apache.hadoop.hbase.protobuf.generated.MasterProtos$MasterService$2.callBlockingMethod(MasterProtos.java:55371)
    at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2114)
    at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:101)
    at org.apache.hadoop.hbase.ipc.RpcExecutor.consumerLoop(RpcExecutor.java:130)
    at org.apache.hadoop.hbase.ipc.RpcExecutor$1.run(RpcExecutor.java:107)
    at java.lang.Thread.run(Thread.java:745)

Fri Dec 02 19:21:14 IST 2016, RpcRetryingCaller{globalStartTime=1480686674019, pause=100, retries=5}, org.apache.hadoop.hbase.ipc.RemoteWithExtrasException(org.apache.hadoop.hbase.PleaseHoldException): org.apache.hadoop.hbase.PleaseHoldException: Master is initializing
    at org.apache.hadoop.hbase.master.HMaster.checkInitialized(HMaster.java:2324)
    at org.apache.hadoop.hbase.master.MasterRpcServices.getClusterStatus(MasterRpcServices.java:770)
    at org.apache.hadoop.hbase.protobuf.generated.MasterProtos$MasterService$2.callBlockingMethod(MasterProtos.java:55371)
    at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2114)
    at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:101)
    at org.apache.hadoop.hbase.ipc.RpcExecutor.consumerLoop(RpcExecutor.java:130)
    at org.apache.hadoop.hbase.ipc.RpcExecutor$1.run(RpcExecutor.java:107)
    at java.lang.Thread.run(Thread.java:745)

Fri Dec 02 19:21:14 IST 2016, RpcRetryingCaller{globalStartTime=1480686674019, pause=100, retries=5}, org.apache.hadoop.hbase.ipc.RemoteWithExtrasException(org.apache.hadoop.hbase.PleaseHoldException): org.apache.hadoop.hbase.PleaseHoldException: Master is initializing
    at org.apache.hadoop.hbase.master.HMaster.checkInitialized(HMaster.java:2324)
    at org.apache.hadoop.hbase.master.MasterRpcServices.getClusterStatus(MasterRpcServices.java:770)
    at org.apache.hadoop.hbase.protobuf.generated.MasterProtos$MasterService$2.callBlockingMethod(MasterProtos.java:55371)
    at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2114)
    at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:101)
    at org.apache.hadoop.hbase.ipc.RpcExecutor.consumerLoop(RpcExecutor.java:130)
    at org.apache.hadoop.hbase.ipc.RpcExecutor$1.run(RpcExecutor.java:107)
    at java.lang.Thread.run(Thread.java:745)

Fri Dec 02 19:21:15 IST 2016, RpcRetryingCaller{globalStartTime=1480686674019, pause=100, retries=5}, org.apache.hadoop.hbase.exceptions.ConnectionClosingException: Call to hscale-dev1-nn/10.60.70.10:16000 failed on local exception: org.apache.hadoop.hbase.exceptions.ConnectionClosingException: Connection to hscale-dev1-nn/10.60.70.10:16000 is closing. Call id=98, waitTime=1
Fri Dec 02 19:21:16 IST 2016, RpcRetryingCaller{globalStartTime=1480686674019, pause=100, retries=5}, org.apache.hadoop.hbase.exceptions.ConnectionClosingException: Call to hscale-dev1-nn/10.60.70.10:16000 failed on local exception: org.apache.hadoop.hbase.exceptions.ConnectionClosingException: Connection to hscale-dev1-nn/10.60.70.10:16000 is closing. Call id=102, waitTime=4

    at org.apache.hadoop.hbase.client.RpcRetryingCaller.callWithRetries(RpcRetryingCaller.java:147)
    at org.apache.hadoop.hbase.client.HBaseAdmin.executeCallable(HBaseAdmin.java:4083)
    at org.apache.hadoop.hbase.client.HBaseAdmin.getClusterStatus(HBaseAdmin.java:2765)
    at com.splicemachine.access.hbase.H10PartitionAdmin.allServers(H10PartitionAdmin.java:120)
    at com.splicemachine.hbase.HBaseRegionLoads.fetchRegionLoads(HBaseRegionLoads.java:164)
    at com.splicemachine.hbase.HBaseRegionLoads.access$000(HBaseRegionLoads.java:60)
    at com.splicemachine.hbase.HBaseRegionLoads$1.run(HBaseRegionLoads.java:81)
    at com.splicemachine.concurrent.LoggingScheduledThreadPoolExecutor$LoggingRunnable.run(LoggingScheduledThreadPoolExecutor.java:75)
    at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
    at java.util.concurrent.FutureTask.runAndReset(FutureTask.java:308)
    at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$301(ScheduledThreadPoolExecutor.java:180)
    at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:294)
    at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
    at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
    at java.lang.Thread.run(Thread.java:745)
Caused by: org.apache.hadoop.hbase.exceptions.ConnectionClosingException: Call to hscale-dev1-nn/10.60.70.10:16000 failed on local exception: org.apache.hadoop.hbase.exceptions.ConnectionClosingException: Connection to hscale-dev1-nn/10.60.70.10:16000 is closing. Call id=102, waitTime=4
    at org.apache.hadoop.hbase.ipc.RpcClientImpl.wrapException(RpcClientImpl.java:1259)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl.call(RpcClientImpl.java:1230)
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient.callBlockingMethod(AbstractRpcClient.java:213)
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient$BlockingRpcChannelImplementation.callBlockingMethod(AbstractRpcClient.java:287)
    at org.apache.hadoop.hbase.protobuf.generated.MasterProtos$MasterService$BlockingStub.getClusterStatus(MasterProtos.java:58140)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation$4.getClusterStatus(ConnectionManager.java:2036)
    at org.apache.hadoop.hbase.client.HBaseAdmin$33.call(HBaseAdmin.java:2769)
    at org.apache.hadoop.hbase.client.HBaseAdmin$33.call(HBaseAdmin.java:2765)
    at org.apache.hadoop.hbase.client.RpcRetryingCaller.callWithRetries(RpcRetryingCaller.java:126)
    ... 14 more
Caused by: org.apache.hadoop.hbase.exceptions.ConnectionClosingException: Connection to hscale-dev1-nn/10.60.70.10:16000 is closing. Call id=102, waitTime=4
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.cleanupCalls(RpcClientImpl.java:1047)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.close(RpcClientImpl.java:846)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.run(RpcClientImpl.java:574)
2016-12-02 19:21:16,165 WARN  [pool-10-thread-1] client.HBaseAdmin: failed to get the procedure result procId=6
org.apache.hadoop.hbase.client.RetriesExhaustedException: Failed after attempts=5, exceptions:
Fri Dec 02 19:21:14 IST 2016, RpcRetryingCaller{globalStartTime=1480686674108, pause=100, retries=5}, org.apache.hadoop.hbase.ipc.RemoteWithExtrasException(org.apache.hadoop.hbase.PleaseHoldException): org.apache.hadoop.hbase.PleaseHoldException: Master is initializing
    at org.apache.hadoop.hbase.master.HMaster.checkInitialized(HMaster.java:2324)
    at org.apache.hadoop.hbase.master.MasterRpcServices.getProcedureResult(MasterRpcServices.java:1023)
    at org.apache.hadoop.hbase.protobuf.generated.MasterProtos$MasterService$2.callBlockingMethod(MasterProtos.java:55469)
    at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2114)
    at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:101)
    at org.apache.hadoop.hbase.ipc.RpcExecutor.consumerLoop(RpcExecutor.java:130)
    at org.apache.hadoop.hbase.ipc.RpcExecutor$1.run(RpcExecutor.java:107)
    at java.lang.Thread.run(Thread.java:745)

Fri Dec 02 19:21:14 IST 2016, RpcRetryingCaller{globalStartTime=1480686674108, pause=100, retries=5}, org.apache.hadoop.hbase.ipc.RemoteWithExtrasException(org.apache.hadoop.hbase.PleaseHoldException): org.apache.hadoop.hbase.PleaseHoldException: Master is initializing
    at org.apache.hadoop.hbase.master.HMaster.checkInitialized(HMaster.java:2324)
    at org.apache.hadoop.hbase.master.MasterRpcServices.getProcedureResult(MasterRpcServices.java:1023)
    at org.apache.hadoop.hbase.protobuf.generated.MasterProtos$MasterService$2.callBlockingMethod(MasterProtos.java:55469)
    at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2114)
    at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:101)
    at org.apache.hadoop.hbase.ipc.RpcExecutor.consumerLoop(RpcExecutor.java:130)
    at org.apache.hadoop.hbase.ipc.RpcExecutor$1.run(RpcExecutor.java:107)
    at java.lang.Thread.run(Thread.java:745)

Fri Dec 02 19:21:14 IST 2016, RpcRetryingCaller{globalStartTime=1480686674108, pause=100, retries=5}, org.apache.hadoop.hbase.ipc.RemoteWithExtrasException(org.apache.hadoop.hbase.PleaseHoldException): org.apache.hadoop.hbase.PleaseHoldException: Master is initializing
    at org.apache.hadoop.hbase.master.HMaster.checkInitialized(HMaster.java:2324)
    at org.apache.hadoop.hbase.master.MasterRpcServices.getProcedureResult(MasterRpcServices.java:1023)
    at org.apache.hadoop.hbase.protobuf.generated.MasterProtos$MasterService$2.callBlockingMethod(MasterProtos.java:55469)
    at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2114)
    at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:101)
    at org.apache.hadoop.hbase.ipc.RpcExecutor.consumerLoop(RpcExecutor.java:130)
    at org.apache.hadoop.hbase.ipc.RpcExecutor$1.run(RpcExecutor.java:107)
    at java.lang.Thread.run(Thread.java:745)

Fri Dec 02 19:21:15 IST 2016, RpcRetryingCaller{globalStartTime=1480686674108, pause=100, retries=5}, org.apache.hadoop.hbase.exceptions.ConnectionClosingException: Call to hscale-dev1-nn/10.60.70.10:16000 failed on local exception: org.apache.hadoop.hbase.exceptions.ConnectionClosingException: Connection to hscale-dev1-nn/10.60.70.10:16000 is closing. Call id=100, waitTime=0
Fri Dec 02 19:21:16 IST 2016, RpcRetryingCaller{globalStartTime=1480686674108, pause=100, retries=5}, org.apache.hadoop.hbase.ipc.RemoteWithExtrasException(org.apache.hadoop.hbase.PleaseHoldException): org.apache.hadoop.hbase.PleaseHoldException: Master is initializing
    at org.apache.hadoop.hbase.master.HMaster.checkInitialized(HMaster.java:2324)
    at org.apache.hadoop.hbase.master.MasterRpcServices.getProcedureResult(MasterRpcServices.java:1023)
    at org.apache.hadoop.hbase.protobuf.generated.MasterProtos$MasterService$2.callBlockingMethod(MasterProtos.java:55469)
    at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2114)
    at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:101)
    at org.apache.hadoop.hbase.ipc.RpcExecutor.consumerLoop(RpcExecutor.java:130)
    at org.apache.hadoop.hbase.ipc.RpcExecutor$1.run(RpcExecutor.java:107)
    at java.lang.Thread.run(Thread.java:745)


    at org.apache.hadoop.hbase.client.RpcRetryingCaller.callWithRetries(RpcRetryingCaller.java:147)
    at org.apache.hadoop.hbase.client.HBaseAdmin.executeCallable(HBaseAdmin.java:4083)
    at org.apache.hadoop.hbase.client.HBaseAdmin.access$700(HBaseAdmin.java:194)
    at org.apache.hadoop.hbase.client.HBaseAdmin$ProcedureFuture.getProcedureResult(HBaseAdmin.java:4383)
    at org.apache.hadoop.hbase.client.HBaseAdmin$ProcedureFuture.waitProcedureResult(HBaseAdmin.java:4335)
    at org.apache.hadoop.hbase.client.HBaseAdmin$ProcedureFuture.get(HBaseAdmin.java:4291)
    at org.apache.hadoop.hbase.client.HBaseAdmin.createTable(HBaseAdmin.java:647)
    at org.apache.hadoop.hbase.client.HBaseAdmin.createTable(HBaseAdmin.java:577)
    at com.splicemachine.lifecycle.RegionServerLifecycle.distributedStart(RegionServerLifecycle.java:66)
    at com.splicemachine.derby.lifecycle.EngineLifecycleService.start(EngineLifecycleService.java:81)
    at com.splicemachine.lifecycle.DatabaseLifecycleManager$Startup.bootServices(DatabaseLifecycleManager.java:229)
    at com.splicemachine.lifecycle.DatabaseLifecycleManager$Startup.run(DatabaseLifecycleManager.java:220)
    at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
    at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
    at java.lang.Thread.run(Thread.java:745)
Caused by: org.apache.hadoop.hbase.ipc.RemoteWithExtrasException(org.apache.hadoop.hbase.PleaseHoldException): org.apache.hadoop.hbase.PleaseHoldException: Master is initializing
    at org.apache.hadoop.hbase.master.HMaster.checkInitialized(HMaster.java:2324)
    at org.apache.hadoop.hbase.master.MasterRpcServices.getProcedureResult(MasterRpcServices.java:1023)
    at org.apache.hadoop.hbase.protobuf.generated.MasterProtos$MasterService$2.callBlockingMethod(MasterProtos.java:55469)
    at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2114)
    at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:101)
    at org.apache.hadoop.hbase.ipc.RpcExecutor.consumerLoop(RpcExecutor.java:130)
    at org.apache.hadoop.hbase.ipc.RpcExecutor$1.run(RpcExecutor.java:107)
    at java.lang.Thread.run(Thread.java:745)

    at org.apache.hadoop.hbase.ipc.RpcClientImpl.call(RpcClientImpl.java:1226)
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient.callBlockingMethod(AbstractRpcClient.java:213)
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient$BlockingRpcChannelImplementation.callBlockingMethod(AbstractRpcClient.java:287)
    at org.apache.hadoop.hbase.protobuf.generated.MasterProtos$MasterService$BlockingStub.getProcedureResult(MasterProtos.java:58728)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation$4.getProcedureResult(ConnectionManager.java:1951)
    at org.apache.hadoop.hbase.client.HBaseAdmin$ProcedureFuture$2.call(HBaseAdmin.java:4387)
    at org.apache.hadoop.hbase.client.HBaseAdmin$ProcedureFuture$2.call(HBaseAdmin.java:4384)
    at org.apache.hadoop.hbase.client.RpcRetryingCaller.callWithRetries(RpcRetryingCaller.java:126)
    ... 14 more
2016-12-02 19:21:18,662 INFO  [PriorityRpcServer.handler=1,queue=1,port=16020] regionserver.RSRpcServices: Open hbase:meta,,1.1588230740
2016-12-02 19:21:18,701 INFO  [RS_OPEN_META-hscale-dev1-dn1:16020-0] wal.WALFactory: Instantiating WALProvider of type class org.apache.hadoop.hbase.wal.DefaultWALProvider
2016-12-02 19:21:18,706 INFO  [RS_OPEN_META-hscale-dev1-dn1:16020-0] wal.FSHLog: WAL configuration: blocksize=128 MB, rollsize=121.60 MB, prefix=hscale-dev1-dn1%2C16020%2C1480686617338..meta, suffix=.meta, logDir=hdfs://hscale-dev1-nn:8020/apps/hbase/data/WALs/hscale-dev1-dn1,16020,1480686617338, archiveDir=hdfs://hscale-dev1-nn:8020/apps/hbase/data/oldWALs
2016-12-02 19:21:18,764 INFO  [RS_OPEN_META-hscale-dev1-dn1:16020-0] wal.FSHLog: Slow sync cost: 41 ms, current pipeline: []
2016-12-02 19:21:18,764 INFO  [RS_OPEN_META-hscale-dev1-dn1:16020-0] wal.FSHLog: New WAL /apps/hbase/data/WALs/hscale-dev1-dn1,16020,1480686617338/hscale-dev1-dn1%2C16020%2C1480686617338..meta.1480686678706.meta
2016-12-02 19:21:18,824 INFO  [RS_OPEN_META-hscale-dev1-dn1:16020-0] coprocessor.CoprocessorHost: System coprocessor org.apache.hadoop.hbase.security.access.SecureBulkLoadEndpoint was loaded successfully with priority (536870911).
2016-12-02 19:21:18,838 INFO  [RS_OPEN_META-hscale-dev1-dn1:16020-0] coprocessor.CoprocessorHost: System coprocessor com.splicemachine.hbase.MemstoreAwareObserver was loaded successfully with priority (536870912).
2016-12-02 19:21:18,846 INFO  [RS_OPEN_META-hscale-dev1-dn1:16020-0] coprocessor.CoprocessorHost: System coprocessor com.splicemachine.derby.hbase.SpliceIndexObserver was loaded successfully with priority (536870913).
2016-12-02 19:21:18,855 INFO  [RS_OPEN_META-hscale-dev1-dn1:16020-0] coprocessor.CoprocessorHost: System coprocessor com.splicemachine.derby.hbase.SpliceIndexEndpoint was loaded successfully with priority (536870914).
2016-12-02 19:21:18,860 INFO  [RS_OPEN_META-hscale-dev1-dn1:16020-0] coprocessor.CoprocessorHost: System coprocessor com.splicemachine.hbase.RegionSizeEndpoint was loaded successfully with priority (536870915).
2016-12-02 19:21:18,873 INFO  [RS_OPEN_META-hscale-dev1-dn1:16020-0] coprocessor.CoprocessorHost: System coprocessor com.splicemachine.si.data.hbase.coprocessor.TxnLifecycleEndpoint was loaded successfully with priority (536870916).
2016-12-02 19:21:18,877 INFO  [RS_OPEN_META-hscale-dev1-dn1:16020-0] coprocessor.CoprocessorHost: System coprocessor com.splicemachine.si.data.hbase.coprocessor.SIObserver was loaded successfully with priority (536870917).
2016-12-02 19:21:18,881 INFO  [RS_OPEN_META-hscale-dev1-dn1:16020-0] coprocessor.CoprocessorHost: System coprocessor com.splicemachine.hbase.BackupEndpointObserver was loaded successfully with priority (536870918).
2016-12-02 19:21:18,898 INFO  [RS_OPEN_META-hscale-dev1-dn1:16020-0] regionserver.RegionCoprocessorHost: Loaded coprocessor org.apache.hadoop.hbase.coprocessor.MultiRowMutationEndpoint from HTD of hbase:meta successfully.
2016-12-02 19:21:18,939 INFO  [StoreOpener-1588230740-1] hfile.CacheConfig: blockCache=LruBlockCache{blockCount=0, currentSize=827112, freeSize=804479256, maxSize=805306368, heapSize=827112, minSize=765041024, minFactor=0.95, multiSize=382520512, multiFactor=0.5, singleSize=191260256, singleFactor=0.25}, cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=true, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false
2016-12-02 19:21:18,947 INFO  [StoreOpener-1588230740-1] compactions.CompactionConfiguration: size [16777216, 260046848); files [5, 10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000
2016-12-02 19:21:19,163 INFO  [RS_OPEN_META-hscale-dev1-dn1:16020-0] regionserver.HRegion: Replaying edits from hdfs://hscale-dev1-nn:8020/apps/hbase/data/data/hbase/meta/1588230740/recovered.edits/0000000000000000091
2016-12-02 19:21:19,190 INFO  [RS_OPEN_META-hscale-dev1-dn1:16020-0] regionserver.HRegion: Started memstore flush for hbase:meta,,1.1588230740, current region memstore size 20.01 KB, and 1/1 column families' memstores are being flushed.; wal is null, using passed sequenceid=91
2016-12-02 19:21:19,385 INFO  [RS_OPEN_META-hscale-dev1-dn1:16020-0] regionserver.DefaultStoreFlusher: Flushed, sequenceid=91, memsize=20.0 K, hasBloomFilter=false, into tmp file hdfs://hscale-dev1-nn:8020/apps/hbase/data/data/hbase/meta/1588230740/.tmp/2164ac9fc3c24ceebfaf979dff8767dc
2016-12-02 19:21:19,439 INFO  [RS_OPEN_META-hscale-dev1-dn1:16020-0] regionserver.HStore: Added hdfs://hscale-dev1-nn:8020/apps/hbase/data/data/hbase/meta/1588230740/info/2164ac9fc3c24ceebfaf979dff8767dc, entries=80, sequenceid=91, filesize=14.8 K
2016-12-02 19:21:19,439 INFO  [RS_OPEN_META-hscale-dev1-dn1:16020-0] regionserver.HRegion: Finished memstore flush of ~20.01 KB/20488, currentsize=0 B/0 for region hbase:meta,,1.1588230740 in 249ms, sequenceid=91, compaction requested=true; wal=null
2016-12-02 19:21:19,513 INFO  [RS_OPEN_META-hscale-dev1-dn1:16020-0] regionserver.HRegion: Onlined 1588230740; next sequenceid=92
2016-12-02 19:21:19,526 INFO  [PostOpenDeployTasks:1588230740] regionserver.HRegionServer: Post open deploy tasks for hbase:meta,,1.1588230740
2016-12-02 19:21:19,527 INFO  [PostOpenDeployTasks:1588230740] zookeeper.MetaTableLocator: Setting hbase:meta region location in ZooKeeper as hscale-dev1-dn1,16020,1480686617338
2016-12-02 19:21:19,546 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020-shortCompactions-1480686679527] regionserver.HRegion: Starting compaction on info in region hbase:meta,,1.1588230740
2016-12-02 19:21:19,546 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020-shortCompactions-1480686679527] regionserver.HStore: Starting compaction of 5 file(s) in info of hbase:meta,,1.1588230740 into tmpdir=hdfs://hscale-dev1-nn:8020/apps/hbase/data/data/hbase/meta/1588230740/.tmp, totalSize=35.2 K
2016-12-02 19:21:19,565 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020-shortCompactions-1480686679527] hfile.CacheConfig: blockCache=LruBlockCache{blockCount=0, currentSize=827112, freeSize=804479256, maxSize=805306368, heapSize=827112, minSize=765041024, minFactor=0.95, multiSize=382520512, multiFactor=0.5, singleSize=191260256, singleFactor=0.25}, cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=true, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false
2016-12-02 19:21:19,863 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020-shortCompactions-1480686679527] regionserver.HStore: Completed compaction of 5 (all) file(s) in info of hbase:meta,,1.1588230740 into 3b9b91eb7b8641749511413ca3b958ec(size=16.0 K), total size for store is 16.0 K. This selection was in queue for 0sec, and took 0sec to execute.
2016-12-02 19:21:19,864 INFO  [regionserver/hscale-dev1-dn1/10.60.70.11:16020-shortCompactions-1480686679527] regionserver.CompactSplitThread: Completed compaction: Request = regionName=hbase:meta,,1.1588230740, storeName=info, fileCount=5, fileSize=35.2 K, priority=15, time=5433160075837372; duration=0sec
2016-12-02 19:21:19,901 INFO  [SplitLogWorker-hscale-dev1-dn1:16020] coordination.ZkSplitLogWorkerCoordination: worker hscale-dev1-dn1,16020,1480686617338 acquired task /hbase-unsecure/splitWAL/WALs%2Fhscale-dev1-dn2%2C16020%2C1480686240889-splitting%2Fhscale-dev1-dn2%252C16020%252C1480686240889.null0.1480686243207
2016-12-02 19:21:19,923 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-1] wal.WALSplitter: Splitting wal: hdfs://hscale-dev1-nn:8020/apps/hbase/data/WALs/hscale-dev1-dn2,16020,1480686240889-splitting/hscale-dev1-dn2%2C16020%2C1480686240889.null0.1480686243207, length=380
2016-12-02 19:21:19,923 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-1] wal.WALSplitter: DistributedLogReplay = false
2016-12-02 19:21:19,944 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-1] util.FSHDFSUtils: Recovering lease on dfs file hdfs://hscale-dev1-nn:8020/apps/hbase/data/WALs/hscale-dev1-dn2,16020,1480686240889-splitting/hscale-dev1-dn2%2C16020%2C1480686240889.null0.1480686243207
2016-12-02 19:21:19,946 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-1] util.FSHDFSUtils: recoverLease=true, attempt=0 on file=hdfs://hscale-dev1-nn:8020/apps/hbase/data/WALs/hscale-dev1-dn2,16020,1480686240889-splitting/hscale-dev1-dn2%2C16020%2C1480686240889.null0.1480686243207 after 1ms
2016-12-02 19:21:20,005 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-1-Writer-0] wal.WALSplitter: Creating writer path=hdfs://hscale-dev1-nn:8020/apps/hbase/data/data/splice/SPLICE_TXN/5147876c22787fb8040fbe21bfff47ad/recovered.edits/0000000000000000003.temp region=5147876c22787fb8040fbe21bfff47ad
2016-12-02 19:21:20,006 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-1] wal.WALSplitter: Split writers finished
2016-12-02 19:21:20,049 WARN  [hbase-region-load-updater-0] client.ConnectionManager$HConnectionImplementation: Checking master connection
com.google.protobuf.ServiceException: org.apache.hadoop.hbase.exceptions.ConnectionClosingException: Call to hscale-dev1-nn/10.60.70.10:16000 failed on local exception: org.apache.hadoop.hbase.exceptions.ConnectionClosingException: Connection to hscale-dev1-nn/10.60.70.10:16000 is closing. Call id=111, waitTime=0
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient.callBlockingMethod(AbstractRpcClient.java:223)
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient$BlockingRpcChannelImplementation.callBlockingMethod(AbstractRpcClient.java:287)
    at org.apache.hadoop.hbase.protobuf.generated.MasterProtos$MasterService$BlockingStub.isMasterRunning(MasterProtos.java:58152)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation$MasterServiceState.isMasterRunning(ConnectionManager.java:1444)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.isKeepAliveMasterConnectedAndRunning(ConnectionManager.java:2099)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation.getKeepAliveMasterService(ConnectionManager.java:1708)
    at org.apache.hadoop.hbase.client.MasterCallable.prepare(MasterCallable.java:38)
    at org.apache.hadoop.hbase.client.RpcRetryingCaller.callWithRetries(RpcRetryingCaller.java:124)
    at org.apache.hadoop.hbase.client.HBaseAdmin.executeCallable(HBaseAdmin.java:4083)
    at org.apache.hadoop.hbase.client.HBaseAdmin.getClusterStatus(HBaseAdmin.java:2765)
    at com.splicemachine.access.hbase.H10PartitionAdmin.allServers(H10PartitionAdmin.java:120)
    at com.splicemachine.hbase.HBaseRegionLoads.fetchRegionLoads(HBaseRegionLoads.java:164)
    at com.splicemachine.hbase.HBaseRegionLoads.access$000(HBaseRegionLoads.java:60)
    at com.splicemachine.hbase.HBaseRegionLoads$1.run(HBaseRegionLoads.java:81)
    at com.splicemachine.concurrent.LoggingScheduledThreadPoolExecutor$LoggingRunnable.run(LoggingScheduledThreadPoolExecutor.java:75)
    at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
    at java.util.concurrent.FutureTask.runAndReset(FutureTask.java:308)
    at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$301(ScheduledThreadPoolExecutor.java:180)
    at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:294)
    at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
    at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
    at java.lang.Thread.run(Thread.java:745)
Caused by: org.apache.hadoop.hbase.exceptions.ConnectionClosingException: Call to hscale-dev1-nn/10.60.70.10:16000 failed on local exception: org.apache.hadoop.hbase.exceptions.ConnectionClosingException: Connection to hscale-dev1-nn/10.60.70.10:16000 is closing. Call id=111, waitTime=0
    at org.apache.hadoop.hbase.ipc.RpcClientImpl.wrapException(RpcClientImpl.java:1259)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl.call(RpcClientImpl.java:1230)
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient.callBlockingMethod(AbstractRpcClient.java:213)
    ... 21 more
Caused by: org.apache.hadoop.hbase.exceptions.ConnectionClosingException: Connection to hscale-dev1-nn/10.60.70.10:16000 is closing. Call id=111, waitTime=0
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.cleanupCalls(RpcClientImpl.java:1047)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.close(RpcClientImpl.java:846)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.run(RpcClientImpl.java:574)
2016-12-02 19:21:20,053 INFO  [split-log-closeStream-1] wal.WALSplitter: Rename hdfs://hscale-dev1-nn:8020/apps/hbase/data/data/splice/SPLICE_TXN/5147876c22787fb8040fbe21bfff47ad/recovered.edits/0000000000000000003.temp to hdfs://hscale-dev1-nn:8020/apps/hbase/data/data/splice/SPLICE_TXN/5147876c22787fb8040fbe21bfff47ad/recovered.edits/0000000000000000003
2016-12-02 19:21:20,053 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-1] wal.WALSplitter: Processed 1 edits across 1 regions; edits skipped=0; log file=hdfs://hscale-dev1-nn:8020/apps/hbase/data/WALs/hscale-dev1-dn2,16020,1480686240889-splitting/hscale-dev1-dn2%2C16020%2C1480686240889.null0.1480686243207, length=380, corrupted=false, progress failed=false
2016-12-02 19:21:20,092 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-1] coordination.ZkSplitLogWorkerCoordination: successfully transitioned task /hbase-unsecure/splitWAL/WALs%2Fhscale-dev1-dn2%2C16020%2C1480686240889-splitting%2Fhscale-dev1-dn2%252C16020%252C1480686240889.null0.1480686243207 to final state DONE hscale-dev1-dn1,16020,1480686617338
2016-12-02 19:21:20,093 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-1] handler.WALSplitterHandler: worker hscale-dev1-dn1,16020,1480686617338 done with task org.apache.hadoop.hbase.coordination.ZkSplitLogWorkerCoordination$ZkSplitTaskDetails@59fbbfc2 in 191ms
2016-12-02 19:21:20,604 INFO  [SplitLogWorker-hscale-dev1-dn1:16020] coordination.ZkSplitLogWorkerCoordination: worker hscale-dev1-dn1,16020,1480686617338 acquired task /hbase-unsecure/splitWAL/WALs%2Fhscale-dev1-dn4%2C16020%2C1480686241892-splitting%2Fhscale-dev1-dn4%252C16020%252C1480686241892.null4.1480686244615
2016-12-02 19:21:20,624 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-0] wal.WALSplitter: Splitting wal: hdfs://hscale-dev1-nn:8020/apps/hbase/data/WALs/hscale-dev1-dn4,16020,1480686241892-splitting/hscale-dev1-dn4%2C16020%2C1480686241892.null4.1480686244615, length=91
2016-12-02 19:21:20,624 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-0] wal.WALSplitter: DistributedLogReplay = false
2016-12-02 19:21:20,642 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-0] util.FSHDFSUtils: Recovering lease on dfs file hdfs://hscale-dev1-nn:8020/apps/hbase/data/WALs/hscale-dev1-dn4,16020,1480686241892-splitting/hscale-dev1-dn4%2C16020%2C1480686241892.null4.1480686244615
2016-12-02 19:21:20,643 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-0] util.FSHDFSUtils: recoverLease=true, attempt=0 on file=hdfs://hscale-dev1-nn:8020/apps/hbase/data/WALs/hscale-dev1-dn4,16020,1480686241892-splitting/hscale-dev1-dn4%2C16020%2C1480686241892.null4.1480686244615 after 1ms
2016-12-02 19:21:20,665 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-0] wal.WALSplitter: Split writers finished
2016-12-02 19:21:20,665 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-0] wal.WALSplitter: Processed 0 edits across 0 regions; edits skipped=0; log file=hdfs://hscale-dev1-nn:8020/apps/hbase/data/WALs/hscale-dev1-dn4,16020,1480686241892-splitting/hscale-dev1-dn4%2C16020%2C1480686241892.null4.1480686244615, length=91, corrupted=false, progress failed=false
2016-12-02 19:21:20,681 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-0] coordination.ZkSplitLogWorkerCoordination: successfully transitioned task /hbase-unsecure/splitWAL/WALs%2Fhscale-dev1-dn4%2C16020%2C1480686241892-splitting%2Fhscale-dev1-dn4%252C16020%252C1480686241892.null4.1480686244615 to final state DONE hscale-dev1-dn1,16020,1480686617338
2016-12-02 19:21:20,681 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-0] handler.WALSplitterHandler: worker hscale-dev1-dn1,16020,1480686617338 done with task org.apache.hadoop.hbase.coordination.ZkSplitLogWorkerCoordination$ZkSplitTaskDetails@67d1352d in 77ms
2016-12-02 19:21:21,097 ERROR [hbase-region-load-updater-0] hbase.HBaseRegionLoads: Unable to fetch region load info
org.apache.hadoop.hbase.client.RetriesExhaustedException: Failed after attempts=5, exceptions:
Fri Dec 02 19:21:19 IST 2016, RpcRetryingCaller{globalStartTime=1480686679018, pause=100, retries=5}, org.apache.hadoop.hbase.ipc.RemoteWithExtrasException(org.apache.hadoop.hbase.PleaseHoldException): org.apache.hadoop.hbase.PleaseHoldException: Master is initializing
    at org.apache.hadoop.hbase.master.HMaster.checkInitialized(HMaster.java:2324)
    at org.apache.hadoop.hbase.master.MasterRpcServices.getClusterStatus(MasterRpcServices.java:770)
    at org.apache.hadoop.hbase.protobuf.generated.MasterProtos$MasterService$2.callBlockingMethod(MasterProtos.java:55371)
    at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2114)
    at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:101)
    at org.apache.hadoop.hbase.ipc.RpcExecutor.consumerLoop(RpcExecutor.java:130)
    at org.apache.hadoop.hbase.ipc.RpcExecutor$1.run(RpcExecutor.java:107)
    at java.lang.Thread.run(Thread.java:745)

Fri Dec 02 19:21:19 IST 2016, RpcRetryingCaller{globalStartTime=1480686679018, pause=100, retries=5}, org.apache.hadoop.hbase.ipc.RemoteWithExtrasException(org.apache.hadoop.hbase.PleaseHoldException): org.apache.hadoop.hbase.PleaseHoldException: Master is initializing
    at org.apache.hadoop.hbase.master.HMaster.checkInitialized(HMaster.java:2324)
    at org.apache.hadoop.hbase.master.MasterRpcServices.getClusterStatus(MasterRpcServices.java:770)
    at org.apache.hadoop.hbase.protobuf.generated.MasterProtos$MasterService$2.callBlockingMethod(MasterProtos.java:55371)
    at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2114)
    at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:101)
    at org.apache.hadoop.hbase.ipc.RpcExecutor.consumerLoop(RpcExecutor.java:130)
    at org.apache.hadoop.hbase.ipc.RpcExecutor$1.run(RpcExecutor.java:107)
    at java.lang.Thread.run(Thread.java:745)

Fri Dec 02 19:21:19 IST 2016, RpcRetryingCaller{globalStartTime=1480686679018, pause=100, retries=5}, org.apache.hadoop.hbase.ipc.RemoteWithExtrasException(org.apache.hadoop.hbase.PleaseHoldException): org.apache.hadoop.hbase.PleaseHoldException: Master is initializing
    at org.apache.hadoop.hbase.master.HMaster.checkInitialized(HMaster.java:2324)
    at org.apache.hadoop.hbase.master.MasterRpcServices.getClusterStatus(MasterRpcServices.java:770)
    at org.apache.hadoop.hbase.protobuf.generated.MasterProtos$MasterService$2.callBlockingMethod(MasterProtos.java:55371)
    at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2114)
    at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:101)
    at org.apache.hadoop.hbase.ipc.RpcExecutor.consumerLoop(RpcExecutor.java:130)
    at org.apache.hadoop.hbase.ipc.RpcExecutor$1.run(RpcExecutor.java:107)
    at java.lang.Thread.run(Thread.java:745)

Fri Dec 02 19:21:20 IST 2016, RpcRetryingCaller{globalStartTime=1480686679018, pause=100, retries=5}, org.apache.hadoop.hbase.ipc.RemoteWithExtrasException(org.apache.hadoop.hbase.PleaseHoldException): org.apache.hadoop.hbase.PleaseHoldException: Master is initializing
    at org.apache.hadoop.hbase.master.HMaster.checkInitialized(HMaster.java:2324)
    at org.apache.hadoop.hbase.master.MasterRpcServices.getClusterStatus(MasterRpcServices.java:770)
    at org.apache.hadoop.hbase.protobuf.generated.MasterProtos$MasterService$2.callBlockingMethod(MasterProtos.java:55371)
    at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2114)
    at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:101)
    at org.apache.hadoop.hbase.ipc.RpcExecutor.consumerLoop(RpcExecutor.java:130)
    at org.apache.hadoop.hbase.ipc.RpcExecutor$1.run(RpcExecutor.java:107)
    at java.lang.Thread.run(Thread.java:745)

Fri Dec 02 19:21:21 IST 2016, RpcRetryingCaller{globalStartTime=1480686679018, pause=100, retries=5}, org.apache.hadoop.hbase.ipc.RemoteWithExtrasException(org.apache.hadoop.hbase.PleaseHoldException): org.apache.hadoop.hbase.PleaseHoldException: Master is initializing
    at org.apache.hadoop.hbase.master.HMaster.checkInitialized(HMaster.java:2324)
    at org.apache.hadoop.hbase.master.MasterRpcServices.getClusterStatus(MasterRpcServices.java:770)
    at org.apache.hadoop.hbase.protobuf.generated.MasterProtos$MasterService$2.callBlockingMethod(MasterProtos.java:55371)
    at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2114)
    at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:101)
    at org.apache.hadoop.hbase.ipc.RpcExecutor.consumerLoop(RpcExecutor.java:130)
    at org.apache.hadoop.hbase.ipc.RpcExecutor$1.run(RpcExecutor.java:107)
    at java.lang.Thread.run(Thread.java:745)


    at org.apache.hadoop.hbase.client.RpcRetryingCaller.callWithRetries(RpcRetryingCaller.java:147)
    at org.apache.hadoop.hbase.client.HBaseAdmin.executeCallable(HBaseAdmin.java:4083)
    at org.apache.hadoop.hbase.client.HBaseAdmin.getClusterStatus(HBaseAdmin.java:2765)
    at com.splicemachine.access.hbase.H10PartitionAdmin.allServers(H10PartitionAdmin.java:120)
    at com.splicemachine.hbase.HBaseRegionLoads.fetchRegionLoads(HBaseRegionLoads.java:164)
    at com.splicemachine.hbase.HBaseRegionLoads.access$000(HBaseRegionLoads.java:60)
    at com.splicemachine.hbase.HBaseRegionLoads$1.run(HBaseRegionLoads.java:81)
    at com.splicemachine.concurrent.LoggingScheduledThreadPoolExecutor$LoggingRunnable.run(LoggingScheduledThreadPoolExecutor.java:75)
    at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
    at java.util.concurrent.FutureTask.runAndReset(FutureTask.java:308)
    at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$301(ScheduledThreadPoolExecutor.java:180)
    at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:294)
    at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
    at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
    at java.lang.Thread.run(Thread.java:745)
Caused by: org.apache.hadoop.hbase.ipc.RemoteWithExtrasException(org.apache.hadoop.hbase.PleaseHoldException): org.apache.hadoop.hbase.PleaseHoldException: Master is initializing
    at org.apache.hadoop.hbase.master.HMaster.checkInitialized(HMaster.java:2324)
    at org.apache.hadoop.hbase.master.MasterRpcServices.getClusterStatus(MasterRpcServices.java:770)
    at org.apache.hadoop.hbase.protobuf.generated.MasterProtos$MasterService$2.callBlockingMethod(MasterProtos.java:55371)
    at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2114)
    at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:101)
    at org.apache.hadoop.hbase.ipc.RpcExecutor.consumerLoop(RpcExecutor.java:130)
    at org.apache.hadoop.hbase.ipc.RpcExecutor$1.run(RpcExecutor.java:107)
    at java.lang.Thread.run(Thread.java:745)

    at org.apache.hadoop.hbase.ipc.RpcClientImpl.call(RpcClientImpl.java:1226)
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient.callBlockingMethod(AbstractRpcClient.java:213)
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient$BlockingRpcChannelImplementation.callBlockingMethod(AbstractRpcClient.java:287)
    at org.apache.hadoop.hbase.protobuf.generated.MasterProtos$MasterService$BlockingStub.getClusterStatus(MasterProtos.java:58140)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation$4.getClusterStatus(ConnectionManager.java:2036)
    at org.apache.hadoop.hbase.client.HBaseAdmin$33.call(HBaseAdmin.java:2769)
    at org.apache.hadoop.hbase.client.HBaseAdmin$33.call(HBaseAdmin.java:2765)
    at org.apache.hadoop.hbase.client.RpcRetryingCaller.callWithRetries(RpcRetryingCaller.java:126)
    ... 14 more
2016-12-02 19:21:21,521 INFO  [SplitLogWorker-hscale-dev1-dn1:16020] coordination.ZkSplitLogWorkerCoordination: worker hscale-dev1-dn1,16020,1480686617338 acquired task /hbase-unsecure/splitWAL/WALs%2Fhscale-dev1-dn1%2C16020%2C1480686239118-splitting%2Fhscale-dev1-dn1%252C16020%252C1480686239118.null6.1480686242056
2016-12-02 19:21:21,541 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-1] wal.WALSplitter: Splitting wal: hdfs://hscale-dev1-nn:8020/apps/hbase/data/WALs/hscale-dev1-dn1,16020,1480686239118-splitting/hscale-dev1-dn1%2C16020%2C1480686239118.null6.1480686242056, length=380
2016-12-02 19:21:21,541 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-1] wal.WALSplitter: DistributedLogReplay = false
2016-12-02 19:21:21,575 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-1] util.FSHDFSUtils: Recovering lease on dfs file hdfs://hscale-dev1-nn:8020/apps/hbase/data/WALs/hscale-dev1-dn1,16020,1480686239118-splitting/hscale-dev1-dn1%2C16020%2C1480686239118.null6.1480686242056
2016-12-02 19:21:21,576 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-1] util.FSHDFSUtils: recoverLease=true, attempt=0 on file=hdfs://hscale-dev1-nn:8020/apps/hbase/data/WALs/hscale-dev1-dn1,16020,1480686239118-splitting/hscale-dev1-dn1%2C16020%2C1480686239118.null6.1480686242056 after 1ms
2016-12-02 19:21:21,627 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-1-Writer-2] wal.WALSplitter: Creating writer path=hdfs://hscale-dev1-nn:8020/apps/hbase/data/data/splice/TENTATIVE_DDL/7ba2eae2b5a4d7d5a163e2c50aa33a6b/recovered.edits/0000000000000000003.temp region=7ba2eae2b5a4d7d5a163e2c50aa33a6b
2016-12-02 19:21:21,627 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-1] wal.WALSplitter: Split writers finished
2016-12-02 19:21:21,684 INFO  [split-log-closeStream-1] wal.WALSplitter: Rename hdfs://hscale-dev1-nn:8020/apps/hbase/data/data/splice/TENTATIVE_DDL/7ba2eae2b5a4d7d5a163e2c50aa33a6b/recovered.edits/0000000000000000003.temp to hdfs://hscale-dev1-nn:8020/apps/hbase/data/data/splice/TENTATIVE_DDL/7ba2eae2b5a4d7d5a163e2c50aa33a6b/recovered.edits/0000000000000000003
2016-12-02 19:21:21,684 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-1] wal.WALSplitter: Processed 1 edits across 1 regions; edits skipped=0; log file=hdfs://hscale-dev1-nn:8020/apps/hbase/data/WALs/hscale-dev1-dn1,16020,1480686239118-splitting/hscale-dev1-dn1%2C16020%2C1480686239118.null6.1480686242056, length=380, corrupted=false, progress failed=false
2016-12-02 19:21:21,722 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-1] coordination.ZkSplitLogWorkerCoordination: successfully transitioned task /hbase-unsecure/splitWAL/WALs%2Fhscale-dev1-dn1%2C16020%2C1480686239118-splitting%2Fhscale-dev1-dn1%252C16020%252C1480686239118.null6.1480686242056 to final state DONE hscale-dev1-dn1,16020,1480686617338
2016-12-02 19:21:21,722 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-1] handler.WALSplitterHandler: worker hscale-dev1-dn1,16020,1480686617338 done with task org.apache.hadoop.hbase.coordination.ZkSplitLogWorkerCoordination$ZkSplitTaskDetails@6adaa437 in 201ms
2016-12-02 19:21:22,185 INFO  [SplitLogWorker-hscale-dev1-dn1:16020] coordination.ZkSplitLogWorkerCoordination: worker hscale-dev1-dn1,16020,1480686617338 acquired task /hbase-unsecure/splitWAL/WALs%2Fhscale-dev1-dn3%2C16020%2C1480686232696-splitting%2Fhscale-dev1-dn3%252C16020%252C1480686232696.null5.1480686235460
2016-12-02 19:21:22,205 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-0] wal.WALSplitter: Splitting wal: hdfs://hscale-dev1-nn:8020/apps/hbase/data/WALs/hscale-dev1-dn3,16020,1480686232696-splitting/hscale-dev1-dn3%2C16020%2C1480686232696.null5.1480686235460, length=380
2016-12-02 19:21:22,205 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-0] wal.WALSplitter: DistributedLogReplay = false
2016-12-02 19:21:22,222 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-0] util.FSHDFSUtils: Recovering lease on dfs file hdfs://hscale-dev1-nn:8020/apps/hbase/data/WALs/hscale-dev1-dn3,16020,1480686232696-splitting/hscale-dev1-dn3%2C16020%2C1480686232696.null5.1480686235460
2016-12-02 19:21:22,223 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-0] util.FSHDFSUtils: recoverLease=true, attempt=0 on file=hdfs://hscale-dev1-nn:8020/apps/hbase/data/WALs/hscale-dev1-dn3,16020,1480686232696-splitting/hscale-dev1-dn3%2C16020%2C1480686232696.null5.1480686235460 after 1ms
2016-12-02 19:21:22,272 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-0-Writer-2] wal.WALSplitter: Creating writer path=hdfs://hscale-dev1-nn:8020/apps/hbase/data/data/splice/SPLICE_TXN/7347e789234a06dd5d95df6cbbbbe149/recovered.edits/0000000000000000003.temp region=7347e789234a06dd5d95df6cbbbbe149
2016-12-02 19:21:22,273 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-0] wal.WALSplitter: Split writers finished
2016-12-02 19:21:22,328 INFO  [split-log-closeStream-1] wal.WALSplitter: Rename hdfs://hscale-dev1-nn:8020/apps/hbase/data/data/splice/SPLICE_TXN/7347e789234a06dd5d95df6cbbbbe149/recovered.edits/0000000000000000003.temp to hdfs://hscale-dev1-nn:8020/apps/hbase/data/data/splice/SPLICE_TXN/7347e789234a06dd5d95df6cbbbbe149/recovered.edits/0000000000000000003
2016-12-02 19:21:22,329 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-0] wal.WALSplitter: Processed 1 edits across 1 regions; edits skipped=0; log file=hdfs://hscale-dev1-nn:8020/apps/hbase/data/WALs/hscale-dev1-dn3,16020,1480686232696-splitting/hscale-dev1-dn3%2C16020%2C1480686232696.null5.1480686235460, length=380, corrupted=false, progress failed=false
2016-12-02 19:21:22,346 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-0] coordination.ZkSplitLogWorkerCoordination: successfully transitioned task /hbase-unsecure/splitWAL/WALs%2Fhscale-dev1-dn3%2C16020%2C1480686232696-splitting%2Fhscale-dev1-dn3%252C16020%252C1480686232696.null5.1480686235460 to final state DONE hscale-dev1-dn1,16020,1480686617338
2016-12-02 19:21:22,346 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-0] handler.WALSplitterHandler: worker hscale-dev1-dn1,16020,1480686617338 done with task org.apache.hadoop.hbase.coordination.ZkSplitLogWorkerCoordination$ZkSplitTaskDetails@56462ee3 in 161ms
2016-12-02 19:21:23,079 INFO  [SplitLogWorker-hscale-dev1-dn1:16020] coordination.ZkSplitLogWorkerCoordination: worker hscale-dev1-dn1,16020,1480686617338 acquired task /hbase-unsecure/splitWAL/WALs%2Fhscale-dev1-dn3%2C16020%2C1480686232696-splitting%2Fhscale-dev1-dn3%252C16020%252C1480686232696.null4.1480686235390
2016-12-02 19:21:23,098 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-1] wal.WALSplitter: Splitting wal: hdfs://hscale-dev1-nn:8020/apps/hbase/data/WALs/hscale-dev1-dn3,16020,1480686232696-splitting/hscale-dev1-dn3%2C16020%2C1480686232696.null4.1480686235390, length=380
2016-12-02 19:21:23,098 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-1] wal.WALSplitter: DistributedLogReplay = false
2016-12-02 19:21:23,130 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-1] util.FSHDFSUtils: Recovering lease on dfs file hdfs://hscale-dev1-nn:8020/apps/hbase/data/WALs/hscale-dev1-dn3,16020,1480686232696-splitting/hscale-dev1-dn3%2C16020%2C1480686232696.null4.1480686235390
2016-12-02 19:21:23,131 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-1] util.FSHDFSUtils: recoverLease=true, attempt=0 on file=hdfs://hscale-dev1-nn:8020/apps/hbase/data/WALs/hscale-dev1-dn3,16020,1480686232696-splitting/hscale-dev1-dn3%2C16020%2C1480686232696.null4.1480686235390 after 1ms
2016-12-02 19:21:23,158 WARN  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-1] regionserver.HRegionServer: Unable to connect to the master to check the last flushed sequence id
com.google.protobuf.ServiceException: org.apache.hadoop.hbase.exceptions.ConnectionClosingException: Call to hscale-dev1-nn/10.60.70.10:16000 failed on local exception: org.apache.hadoop.hbase.exceptions.ConnectionClosingException: Connection to hscale-dev1-nn/10.60.70.10:16000 is closing. Call id=11, waitTime=0
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient.callBlockingMethod(AbstractRpcClient.java:223)
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient$BlockingRpcChannelImplementation.callBlockingMethod(AbstractRpcClient.java:287)
    at org.apache.hadoop.hbase.protobuf.generated.RegionServerStatusProtos$RegionServerStatusService$BlockingStub.getLastFlushedSequenceId(RegionServerStatusProtos.java:9018)
    at org.apache.hadoop.hbase.regionserver.HRegionServer.getLastSequenceId(HRegionServer.java:2303)
    at org.apache.hadoop.hbase.wal.WALSplitter.splitLogFile(WALSplitter.java:338)
    at org.apache.hadoop.hbase.wal.WALSplitter.splitLogFile(WALSplitter.java:235)
    at org.apache.hadoop.hbase.regionserver.SplitLogWorker$1.exec(SplitLogWorker.java:104)
    at org.apache.hadoop.hbase.regionserver.handler.WALSplitterHandler.process(WALSplitterHandler.java:72)
    at org.apache.hadoop.hbase.executor.EventHandler.run(EventHandler.java:128)
    at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
    at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
    at java.lang.Thread.run(Thread.java:745)
Caused by: org.apache.hadoop.hbase.exceptions.ConnectionClosingException: Call to hscale-dev1-nn/10.60.70.10:16000 failed on local exception: org.apache.hadoop.hbase.exceptions.ConnectionClosingException: Connection to hscale-dev1-nn/10.60.70.10:16000 is closing. Call id=11, waitTime=0
    at org.apache.hadoop.hbase.ipc.RpcClientImpl.wrapException(RpcClientImpl.java:1259)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl.call(RpcClientImpl.java:1230)
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient.callBlockingMethod(AbstractRpcClient.java:213)
    ... 11 more
Caused by: org.apache.hadoop.hbase.exceptions.ConnectionClosingException: Connection to hscale-dev1-nn/10.60.70.10:16000 is closing. Call id=11, waitTime=0
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.cleanupCalls(RpcClientImpl.java:1047)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.close(RpcClientImpl.java:846)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.run(RpcClientImpl.java:574)
2016-12-02 19:21:23,182 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-1-Writer-1] wal.WALSplitter: Creating writer path=hdfs://hscale-dev1-nn:8020/apps/hbase/data/data/splice/SPLICE_TXN/9bbbc596e99d588223974ef41420724d/recovered.edits/0000000000000000003.temp region=9bbbc596e99d588223974ef41420724d
2016-12-02 19:21:23,182 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-1] wal.WALSplitter: Split writers finished
2016-12-02 19:21:23,238 INFO  [split-log-closeStream-1] wal.WALSplitter: Rename hdfs://hscale-dev1-nn:8020/apps/hbase/data/data/splice/SPLICE_TXN/9bbbc596e99d588223974ef41420724d/recovered.edits/0000000000000000003.temp to hdfs://hscale-dev1-nn:8020/apps/hbase/data/data/splice/SPLICE_TXN/9bbbc596e99d588223974ef41420724d/recovered.edits/0000000000000000003
2016-12-02 19:21:23,238 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-1] wal.WALSplitter: Processed 1 edits across 1 regions; edits skipped=0; log file=hdfs://hscale-dev1-nn:8020/apps/hbase/data/WALs/hscale-dev1-dn3,16020,1480686232696-splitting/hscale-dev1-dn3%2C16020%2C1480686232696.null4.1480686235390, length=380, corrupted=false, progress failed=false
2016-12-02 19:21:23,255 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-1] coordination.ZkSplitLogWorkerCoordination: successfully transitioned task /hbase-unsecure/splitWAL/WALs%2Fhscale-dev1-dn3%2C16020%2C1480686232696-splitting%2Fhscale-dev1-dn3%252C16020%252C1480686232696.null4.1480686235390 to final state DONE hscale-dev1-dn1,16020,1480686617338
2016-12-02 19:21:23,255 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-1] handler.WALSplitterHandler: worker hscale-dev1-dn1,16020,1480686617338 done with task org.apache.hadoop.hbase.coordination.ZkSplitLogWorkerCoordination$ZkSplitTaskDetails@25956d2c in 175ms
2016-12-02 19:21:23,791 INFO  [SplitLogWorker-hscale-dev1-dn1:16020] coordination.ZkSplitLogWorkerCoordination: worker hscale-dev1-dn1,16020,1480686617338 acquired task /hbase-unsecure/splitWAL/WALs%2Fhscale-dev1-dn2%2C16020%2C1480686240889-splitting%2Fhscale-dev1-dn2%252C16020%252C1480686240889.null15.1480686244514
2016-12-02 19:21:23,810 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-0] wal.WALSplitter: Splitting wal: hdfs://hscale-dev1-nn:8020/apps/hbase/data/WALs/hscale-dev1-dn2,16020,1480686240889-splitting/hscale-dev1-dn2%2C16020%2C1480686240889.null15.1480686244514, length=91
2016-12-02 19:21:23,810 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-0] wal.WALSplitter: DistributedLogReplay = false
2016-12-02 19:21:23,826 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-0] util.FSHDFSUtils: Recovering lease on dfs file hdfs://hscale-dev1-nn:8020/apps/hbase/data/WALs/hscale-dev1-dn2,16020,1480686240889-splitting/hscale-dev1-dn2%2C16020%2C1480686240889.null15.1480686244514
2016-12-02 19:21:23,827 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-0] util.FSHDFSUtils: recoverLease=true, attempt=0 on file=hdfs://hscale-dev1-nn:8020/apps/hbase/data/WALs/hscale-dev1-dn2,16020,1480686240889-splitting/hscale-dev1-dn2%2C16020%2C1480686240889.null15.1480686244514 after 1ms
2016-12-02 19:21:23,849 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-0] wal.WALSplitter: Split writers finished
2016-12-02 19:21:23,849 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-0] wal.WALSplitter: Processed 0 edits across 0 regions; edits skipped=0; log file=hdfs://hscale-dev1-nn:8020/apps/hbase/data/WALs/hscale-dev1-dn2,16020,1480686240889-splitting/hscale-dev1-dn2%2C16020%2C1480686240889.null15.1480686244514, length=91, corrupted=false, progress failed=false
2016-12-02 19:21:23,865 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-0] coordination.ZkSplitLogWorkerCoordination: successfully transitioned task /hbase-unsecure/splitWAL/WALs%2Fhscale-dev1-dn2%2C16020%2C1480686240889-splitting%2Fhscale-dev1-dn2%252C16020%252C1480686240889.null15.1480686244514 to final state DONE hscale-dev1-dn1,16020,1480686617338
2016-12-02 19:21:23,865 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-0] handler.WALSplitterHandler: worker hscale-dev1-dn1,16020,1480686617338 done with task org.apache.hadoop.hbase.coordination.ZkSplitLogWorkerCoordination$ZkSplitTaskDetails@511d7865 in 74ms
2016-12-02 19:21:24,572 INFO  [SplitLogWorker-hscale-dev1-dn1:16020] coordination.ZkSplitLogWorkerCoordination: worker hscale-dev1-dn1,16020,1480686617338 acquired task /hbase-unsecure/splitWAL/WALs%2Fhscale-dev1-dn1%2C16020%2C1480686239118-splitting%2Fhscale-dev1-dn1%252C16020%252C1480686239118.null3.1480686241879
2016-12-02 19:21:24,593 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-1] wal.WALSplitter: Splitting wal: hdfs://hscale-dev1-nn:8020/apps/hbase/data/WALs/hscale-dev1-dn1,16020,1480686239118-splitting/hscale-dev1-dn1%2C16020%2C1480686239118.null3.1480686241879, length=380
2016-12-02 19:21:24,593 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-1] wal.WALSplitter: DistributedLogReplay = false
2016-12-02 19:21:24,606 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-1] util.FSHDFSUtils: Recovering lease on dfs file hdfs://hscale-dev1-nn:8020/apps/hbase/data/WALs/hscale-dev1-dn1,16020,1480686239118-splitting/hscale-dev1-dn1%2C16020%2C1480686239118.null3.1480686241879
2016-12-02 19:21:24,608 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-1] util.FSHDFSUtils: recoverLease=true, attempt=0 on file=hdfs://hscale-dev1-nn:8020/apps/hbase/data/WALs/hscale-dev1-dn1,16020,1480686239118-splitting/hscale-dev1-dn1%2C16020%2C1480686239118.null3.1480686241879 after 2ms
2016-12-02 19:21:24,632 WARN  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-1] regionserver.HRegionServer: Unable to connect to the master to check the last flushed sequence id
com.google.protobuf.ServiceException: org.apache.hadoop.hbase.exceptions.ConnectionClosingException: Call to hscale-dev1-nn/10.60.70.10:16000 failed on local exception: org.apache.hadoop.hbase.exceptions.ConnectionClosingException: Connection to hscale-dev1-nn/10.60.70.10:16000 is closing. Call id=13, waitTime=1
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient.callBlockingMethod(AbstractRpcClient.java:223)
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient$BlockingRpcChannelImplementation.callBlockingMethod(AbstractRpcClient.java:287)
    at org.apache.hadoop.hbase.protobuf.generated.RegionServerStatusProtos$RegionServerStatusService$BlockingStub.getLastFlushedSequenceId(RegionServerStatusProtos.java:9018)
    at org.apache.hadoop.hbase.regionserver.HRegionServer.getLastSequenceId(HRegionServer.java:2303)
    at org.apache.hadoop.hbase.wal.WALSplitter.splitLogFile(WALSplitter.java:338)
    at org.apache.hadoop.hbase.wal.WALSplitter.splitLogFile(WALSplitter.java:235)
    at org.apache.hadoop.hbase.regionserver.SplitLogWorker$1.exec(SplitLogWorker.java:104)
    at org.apache.hadoop.hbase.regionserver.handler.WALSplitterHandler.process(WALSplitterHandler.java:72)
    at org.apache.hadoop.hbase.executor.EventHandler.run(EventHandler.java:128)
    at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
    at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
    at java.lang.Thread.run(Thread.java:745)
Caused by: org.apache.hadoop.hbase.exceptions.ConnectionClosingException: Call to hscale-dev1-nn/10.60.70.10:16000 failed on local exception: org.apache.hadoop.hbase.exceptions.ConnectionClosingException: Connection to hscale-dev1-nn/10.60.70.10:16000 is closing. Call id=13, waitTime=1
    at org.apache.hadoop.hbase.ipc.RpcClientImpl.wrapException(RpcClientImpl.java:1259)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl.call(RpcClientImpl.java:1230)
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient.callBlockingMethod(AbstractRpcClient.java:213)
    ... 11 more
Caused by: org.apache.hadoop.hbase.exceptions.ConnectionClosingException: Connection to hscale-dev1-nn/10.60.70.10:16000 is closing. Call id=13, waitTime=1
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.cleanupCalls(RpcClientImpl.java:1047)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.close(RpcClientImpl.java:846)
    at org.apache.hadoop.hbase.ipc.RpcClientImpl$Connection.run(RpcClientImpl.java:574)
2016-12-02 19:21:24,662 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-1-Writer-2] wal.WALSplitter: Creating writer path=hdfs://hscale-dev1-nn:8020/apps/hbase/data/data/splice/SPLICE_TXN/91b7a8c3c77f908c0cff6940ccfd6d7d/recovered.edits/0000000000000000003.temp region=91b7a8c3c77f908c0cff6940ccfd6d7d
2016-12-02 19:21:24,662 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-1] wal.WALSplitter: Split writers finished
2016-12-02 19:21:24,748 INFO  [split-log-closeStream-1] wal.WALSplitter: Rename hdfs://hscale-dev1-nn:8020/apps/hbase/data/data/splice/SPLICE_TXN/91b7a8c3c77f908c0cff6940ccfd6d7d/recovered.edits/0000000000000000003.temp to hdfs://hscale-dev1-nn:8020/apps/hbase/data/data/splice/SPLICE_TXN/91b7a8c3c77f908c0cff6940ccfd6d7d/recovered.edits/0000000000000000003
2016-12-02 19:21:24,748 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-1] wal.WALSplitter: Processed 1 edits across 1 regions; edits skipped=0; log file=hdfs://hscale-dev1-nn:8020/apps/hbase/data/WALs/hscale-dev1-dn1,16020,1480686239118-splitting/hscale-dev1-dn1%2C16020%2C1480686239118.null3.1480686241879, length=380, corrupted=false, progress failed=false
2016-12-02 19:21:24,765 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-1] coordination.ZkSplitLogWorkerCoordination: successfully transitioned task /hbase-unsecure/splitWAL/WALs%2Fhscale-dev1-dn1%2C16020%2C1480686239118-splitting%2Fhscale-dev1-dn1%252C16020%252C1480686239118.null3.1480686241879 to final state DONE hscale-dev1-dn1,16020,1480686617338
2016-12-02 19:21:24,765 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-1] handler.WALSplitterHandler: worker hscale-dev1-dn1,16020,1480686617338 done with task org.apache.hadoop.hbase.coordination.ZkSplitLogWorkerCoordination$ZkSplitTaskDetails@5755659e in 193ms
2016-12-02 19:21:25,161 INFO  [SplitLogWorker-hscale-dev1-dn1:16020] coordination.ZkSplitLogWorkerCoordination: worker hscale-dev1-dn1,16020,1480686617338 acquired task /hbase-unsecure/splitWAL/WALs%2Fhscale-dev1-dn4%2C16020%2C1480686241892-splitting%2Fhscale-dev1-dn4%252C16020%252C1480686241892.null8.1480686244925
2016-12-02 19:21:25,186 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-0] wal.WALSplitter: Splitting wal: hdfs://hscale-dev1-nn:8020/apps/hbase/data/WALs/hscale-dev1-dn4,16020,1480686241892-splitting/hscale-dev1-dn4%2C16020%2C1480686241892.null8.1480686244925, length=91
2016-12-02 19:21:25,186 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-0] wal.WALSplitter: DistributedLogReplay = false
2016-12-02 19:21:25,202 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-0] util.FSHDFSUtils: Recovering lease on dfs file hdfs://hscale-dev1-nn:8020/apps/hbase/data/WALs/hscale-dev1-dn4,16020,1480686241892-splitting/hscale-dev1-dn4%2C16020%2C1480686241892.null8.1480686244925
2016-12-02 19:21:25,204 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-0] util.FSHDFSUtils: recoverLease=true, attempt=0 on file=hdfs://hscale-dev1-nn:8020/apps/hbase/data/WALs/hscale-dev1-dn4,16020,1480686241892-splitting/hscale-dev1-dn4%2C16020%2C1480686241892.null8.1480686244925 after 2ms
2016-12-02 19:21:25,233 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-0] wal.WALSplitter: Split writers finished
2016-12-02 19:21:25,233 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-0] wal.WALSplitter: Processed 0 edits across 0 regions; edits skipped=0; log file=hdfs://hscale-dev1-nn:8020/apps/hbase/data/WALs/hscale-dev1-dn4,16020,1480686241892-splitting/hscale-dev1-dn4%2C16020%2C1480686241892.null8.1480686244925, length=91, corrupted=false, progress failed=false
2016-12-02 19:21:25,249 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-0] coordination.ZkSplitLogWorkerCoordination: successfully transitioned task /hbase-unsecure/splitWAL/WALs%2Fhscale-dev1-dn4%2C16020%2C1480686241892-splitting%2Fhscale-dev1-dn4%252C16020%252C1480686241892.null8.1480686244925 to final state DONE hscale-dev1-dn1,16020,1480686617338
2016-12-02 19:21:25,249 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-0] handler.WALSplitterHandler: worker hscale-dev1-dn1,16020,1480686617338 done with task org.apache.hadoop.hbase.coordination.ZkSplitLogWorkerCoordination$ZkSplitTaskDetails@2d3a1fe5 in 88ms
2016-12-02 19:21:25,729 INFO  [SplitLogWorker-hscale-dev1-dn1:16020] coordination.ZkSplitLogWorkerCoordination: worker hscale-dev1-dn1,16020,1480686617338 acquired task /hbase-unsecure/splitWAL/WALs%2Fhscale-dev1-dn4%2C16020%2C1480686241892-splitting%2Fhscale-dev1-dn4%252C16020%252C1480686241892.null9.1480686245010
2016-12-02 19:21:25,756 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-1] wal.WALSplitter: Splitting wal: hdfs://hscale-dev1-nn:8020/apps/hbase/data/WALs/hscale-dev1-dn4,16020,1480686241892-splitting/hscale-dev1-dn4%2C16020%2C1480686241892.null9.1480686245010, length=91
2016-12-02 19:21:25,756 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-1] wal.WALSplitter: DistributedLogReplay = false
2016-12-02 19:21:25,775 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-1] util.FSHDFSUtils: Recovering lease on dfs file hdfs://hscale-dev1-nn:8020/apps/hbase/data/WALs/hscale-dev1-dn4,16020,1480686241892-splitting/hscale-dev1-dn4%2C16020%2C1480686241892.null9.1480686245010
2016-12-02 19:21:25,777 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-1] util.FSHDFSUtils: recoverLease=true, attempt=0 on file=hdfs://hscale-dev1-nn:8020/apps/hbase/data/WALs/hscale-dev1-dn4,16020,1480686241892-splitting/hscale-dev1-dn4%2C16020%2C1480686241892.null9.1480686245010 after 2ms
2016-12-02 19:21:25,809 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-1] wal.WALSplitter: Split writers finished
2016-12-02 19:21:25,810 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-1] wal.WALSplitter: Processed 0 edits across 0 regions; edits skipped=0; log file=hdfs://hscale-dev1-nn:8020/apps/hbase/data/WALs/hscale-dev1-dn4,16020,1480686241892-splitting/hscale-dev1-dn4%2C16020%2C1480686241892.null9.1480686245010, length=91, corrupted=false, progress failed=false
2016-12-02 19:21:25,829 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-1] coordination.ZkSplitLogWorkerCoordination: successfully transitioned task /hbase-unsecure/splitWAL/WALs%2Fhscale-dev1-dn4%2C16020%2C1480686241892-splitting%2Fhscale-dev1-dn4%252C16020%252C1480686241892.null9.1480686245010 to final state DONE hscale-dev1-dn1,16020,1480686617338
2016-12-02 19:21:25,829 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-1] handler.WALSplitterHandler: worker hscale-dev1-dn1,16020,1480686617338 done with task org.apache.hadoop.hbase.coordination.ZkSplitLogWorkerCoordination$ZkSplitTaskDetails@66a732ee in 100ms
2016-12-02 19:21:26,057 ERROR [hbase-region-load-updater-0] hbase.HBaseRegionLoads: Unable to fetch region load info
org.apache.hadoop.hbase.client.RetriesExhaustedException: Failed after attempts=5, exceptions:
Fri Dec 02 19:21:24 IST 2016, RpcRetryingCaller{globalStartTime=1480686684018, pause=100, retries=5}, org.apache.hadoop.hbase.ipc.RemoteWithExtrasException(org.apache.hadoop.hbase.PleaseHoldException): org.apache.hadoop.hbase.PleaseHoldException: Master is initializing
    at org.apache.hadoop.hbase.master.HMaster.checkInitialized(HMaster.java:2324)
    at org.apache.hadoop.hbase.master.MasterRpcServices.getClusterStatus(MasterRpcServices.java:770)
    at org.apache.hadoop.hbase.protobuf.generated.MasterProtos$MasterService$2.callBlockingMethod(MasterProtos.java:55371)
    at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2114)
    at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:101)
    at org.apache.hadoop.hbase.ipc.RpcExecutor.consumerLoop(RpcExecutor.java:130)
    at org.apache.hadoop.hbase.ipc.RpcExecutor$1.run(RpcExecutor.java:107)
    at java.lang.Thread.run(Thread.java:745)

Fri Dec 02 19:21:24 IST 2016, RpcRetryingCaller{globalStartTime=1480686684018, pause=100, retries=5}, org.apache.hadoop.hbase.ipc.RemoteWithExtrasException(org.apache.hadoop.hbase.PleaseHoldException): org.apache.hadoop.hbase.PleaseHoldException: Master is initializing
    at org.apache.hadoop.hbase.master.HMaster.checkInitialized(HMaster.java:2324)
    at org.apache.hadoop.hbase.master.MasterRpcServices.getClusterStatus(MasterRpcServices.java:770)
    at org.apache.hadoop.hbase.protobuf.generated.MasterProtos$MasterService$2.callBlockingMethod(MasterProtos.java:55371)
    at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2114)
    at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:101)
    at org.apache.hadoop.hbase.ipc.RpcExecutor.consumerLoop(RpcExecutor.java:130)
    at org.apache.hadoop.hbase.ipc.RpcExecutor$1.run(RpcExecutor.java:107)
    at java.lang.Thread.run(Thread.java:745)

Fri Dec 02 19:21:24 IST 2016, RpcRetryingCaller{globalStartTime=1480686684018, pause=100, retries=5}, org.apache.hadoop.hbase.ipc.RemoteWithExtrasException(org.apache.hadoop.hbase.PleaseHoldException): org.apache.hadoop.hbase.PleaseHoldException: Master is initializing
    at org.apache.hadoop.hbase.master.HMaster.checkInitialized(HMaster.java:2324)
    at org.apache.hadoop.hbase.master.MasterRpcServices.getClusterStatus(MasterRpcServices.java:770)
    at org.apache.hadoop.hbase.protobuf.generated.MasterProtos$MasterService$2.callBlockingMethod(MasterProtos.java:55371)
    at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2114)
    at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:101)
    at org.apache.hadoop.hbase.ipc.RpcExecutor.consumerLoop(RpcExecutor.java:130)
    at org.apache.hadoop.hbase.ipc.RpcExecutor$1.run(RpcExecutor.java:107)
    at java.lang.Thread.run(Thread.java:745)

Fri Dec 02 19:21:25 IST 2016, RpcRetryingCaller{globalStartTime=1480686684018, pause=100, retries=5}, org.apache.hadoop.hbase.exceptions.ConnectionClosingException: Call to hscale-dev1-nn/10.60.70.10:16000 failed on local exception: org.apache.hadoop.hbase.exceptions.ConnectionClosingException: Connection to hscale-dev1-nn/10.60.70.10:16000 is closing. Call id=122, waitTime=1
Fri Dec 02 19:21:26 IST 2016, RpcRetryingCaller{globalStartTime=1480686684018, pause=100, retries=5}, org.apache.hadoop.hbase.ipc.RemoteWithExtrasException(org.apache.hadoop.hbase.PleaseHoldException): org.apache.hadoop.hbase.PleaseHoldException: Master is initializing
    at org.apache.hadoop.hbase.master.HMaster.checkInitialized(HMaster.java:2324)
    at org.apache.hadoop.hbase.master.MasterRpcServices.getClusterStatus(MasterRpcServices.java:770)
    at org.apache.hadoop.hbase.protobuf.generated.MasterProtos$MasterService$2.callBlockingMethod(MasterProtos.java:55371)
    at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2114)
    at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:101)
    at org.apache.hadoop.hbase.ipc.RpcExecutor.consumerLoop(RpcExecutor.java:130)
    at org.apache.hadoop.hbase.ipc.RpcExecutor$1.run(RpcExecutor.java:107)
    at java.lang.Thread.run(Thread.java:745)


    at org.apache.hadoop.hbase.client.RpcRetryingCaller.callWithRetries(RpcRetryingCaller.java:147)
    at org.apache.hadoop.hbase.client.HBaseAdmin.executeCallable(HBaseAdmin.java:4083)
    at org.apache.hadoop.hbase.client.HBaseAdmin.getClusterStatus(HBaseAdmin.java:2765)
    at com.splicemachine.access.hbase.H10PartitionAdmin.allServers(H10PartitionAdmin.java:120)
    at com.splicemachine.hbase.HBaseRegionLoads.fetchRegionLoads(HBaseRegionLoads.java:164)
    at com.splicemachine.hbase.HBaseRegionLoads.access$000(HBaseRegionLoads.java:60)
    at com.splicemachine.hbase.HBaseRegionLoads$1.run(HBaseRegionLoads.java:81)
    at com.splicemachine.concurrent.LoggingScheduledThreadPoolExecutor$LoggingRunnable.run(LoggingScheduledThreadPoolExecutor.java:75)
    at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
    at java.util.concurrent.FutureTask.runAndReset(FutureTask.java:308)
    at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$301(ScheduledThreadPoolExecutor.java:180)
    at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:294)
    at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
    at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
    at java.lang.Thread.run(Thread.java:745)
Caused by: org.apache.hadoop.hbase.ipc.RemoteWithExtrasException(org.apache.hadoop.hbase.PleaseHoldException): org.apache.hadoop.hbase.PleaseHoldException: Master is initializing
    at org.apache.hadoop.hbase.master.HMaster.checkInitialized(HMaster.java:2324)
    at org.apache.hadoop.hbase.master.MasterRpcServices.getClusterStatus(MasterRpcServices.java:770)
    at org.apache.hadoop.hbase.protobuf.generated.MasterProtos$MasterService$2.callBlockingMethod(MasterProtos.java:55371)
    at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2114)
    at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:101)
    at org.apache.hadoop.hbase.ipc.RpcExecutor.consumerLoop(RpcExecutor.java:130)
    at org.apache.hadoop.hbase.ipc.RpcExecutor$1.run(RpcExecutor.java:107)
    at java.lang.Thread.run(Thread.java:745)

    at org.apache.hadoop.hbase.ipc.RpcClientImpl.call(RpcClientImpl.java:1226)
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient.callBlockingMethod(AbstractRpcClient.java:213)
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient$BlockingRpcChannelImplementation.callBlockingMethod(AbstractRpcClient.java:287)
    at org.apache.hadoop.hbase.protobuf.generated.MasterProtos$MasterService$BlockingStub.getClusterStatus(MasterProtos.java:58140)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation$4.getClusterStatus(ConnectionManager.java:2036)
    at org.apache.hadoop.hbase.client.HBaseAdmin$33.call(HBaseAdmin.java:2769)
    at org.apache.hadoop.hbase.client.HBaseAdmin$33.call(HBaseAdmin.java:2765)
    at org.apache.hadoop.hbase.client.RpcRetryingCaller.callWithRetries(RpcRetryingCaller.java:126)
    ... 14 more
2016-12-02 19:21:26,289 INFO  [SplitLogWorker-hscale-dev1-dn1:16020] coordination.ZkSplitLogWorkerCoordination: worker hscale-dev1-dn1,16020,1480686617338 acquired task /hbase-unsecure/splitWAL/WALs%2Fhscale-dev1-dn1%2C16020%2C1480686239118-splitting%2Fhscale-dev1-dn1%252C16020%252C1480686239118.null8.1480686242237
2016-12-02 19:21:26,307 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-0] wal.WALSplitter: Splitting wal: hdfs://hscale-dev1-nn:8020/apps/hbase/data/WALs/hscale-dev1-dn1,16020,1480686239118-splitting/hscale-dev1-dn1%2C16020%2C1480686239118.null8.1480686242237, length=532
2016-12-02 19:21:26,307 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-0] wal.WALSplitter: DistributedLogReplay = false
2016-12-02 19:21:26,325 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-0] util.FSHDFSUtils: Recovering lease on dfs file hdfs://hscale-dev1-nn:8020/apps/hbase/data/WALs/hscale-dev1-dn1,16020,1480686239118-splitting/hscale-dev1-dn1%2C16020%2C1480686239118.null8.1480686242237
2016-12-02 19:21:26,326 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-0] util.FSHDFSUtils: recoverLease=true, attempt=0 on file=hdfs://hscale-dev1-nn:8020/apps/hbase/data/WALs/hscale-dev1-dn1,16020,1480686239118-splitting/hscale-dev1-dn1%2C16020%2C1480686239118.null8.1480686242237 after 1ms
2016-12-02 19:21:26,390 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-0-Writer-2] wal.WALSplitter: Creating writer path=hdfs://hscale-dev1-nn:8020/apps/hbase/data/data/splice/16/75c25e3e9e9f5281c8face7c909f0b04/recovered.edits/0000000000000000006.temp region=75c25e3e9e9f5281c8face7c909f0b04
2016-12-02 19:21:26,390 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-0] wal.WALSplitter: Split writers finished
2016-12-02 19:21:26,461 INFO  [split-log-closeStream-1] wal.WALSplitter: Rename hdfs://hscale-dev1-nn:8020/apps/hbase/data/data/splice/16/75c25e3e9e9f5281c8face7c909f0b04/recovered.edits/0000000000000000006.temp to hdfs://hscale-dev1-nn:8020/apps/hbase/data/data/splice/16/75c25e3e9e9f5281c8face7c909f0b04/recovered.edits/0000000000000000007
2016-12-02 19:21:26,462 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-0] wal.WALSplitter: Processed 2 edits across 1 regions; edits skipped=0; log file=hdfs://hscale-dev1-nn:8020/apps/hbase/data/WALs/hscale-dev1-dn1,16020,1480686239118-splitting/hscale-dev1-dn1%2C16020%2C1480686239118.null8.1480686242237, length=532, corrupted=false, progress failed=false
2016-12-02 19:21:26,480 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-0] coordination.ZkSplitLogWorkerCoordination: successfully transitioned task /hbase-unsecure/splitWAL/WALs%2Fhscale-dev1-dn1%2C16020%2C1480686239118-splitting%2Fhscale-dev1-dn1%252C16020%252C1480686239118.null8.1480686242237 to final state DONE hscale-dev1-dn1,16020,1480686617338
2016-12-02 19:21:26,481 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-0] handler.WALSplitterHandler: worker hscale-dev1-dn1,16020,1480686617338 done with task org.apache.hadoop.hbase.coordination.ZkSplitLogWorkerCoordination$ZkSplitTaskDetails@d9c6bdc in 192ms
2016-12-02 19:21:27,021 INFO  [SplitLogWorker-hscale-dev1-dn1:16020] coordination.ZkSplitLogWorkerCoordination: worker hscale-dev1-dn1,16020,1480686617338 acquired task /hbase-unsecure/splitWAL/WALs%2Fhscale-dev1-dn4%2C16020%2C1480686241892-splitting%2Fhscale-dev1-dn4%252C16020%252C1480686241892.null13.1480686245360
2016-12-02 19:21:27,046 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-1] wal.WALSplitter: Splitting wal: hdfs://hscale-dev1-nn:8020/apps/hbase/data/WALs/hscale-dev1-dn4,16020,1480686241892-splitting/hscale-dev1-dn4%2C16020%2C1480686241892.null13.1480686245360, length=91
2016-12-02 19:21:27,046 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-1] wal.WALSplitter: DistributedLogReplay = false
2016-12-02 19:21:27,070 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-1] util.FSHDFSUtils: Recovering lease on dfs file hdfs://hscale-dev1-nn:8020/apps/hbase/data/WALs/hscale-dev1-dn4,16020,1480686241892-splitting/hscale-dev1-dn4%2C16020%2C1480686241892.null13.1480686245360
2016-12-02 19:21:27,071 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-1] util.FSHDFSUtils: recoverLease=true, attempt=0 on file=hdfs://hscale-dev1-nn:8020/apps/hbase/data/WALs/hscale-dev1-dn4,16020,1480686241892-splitting/hscale-dev1-dn4%2C16020%2C1480686241892.null13.1480686245360 after 1ms
2016-12-02 19:21:27,100 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-1] wal.WALSplitter: Split writers finished
2016-12-02 19:21:27,100 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-1] wal.WALSplitter: Processed 0 edits across 0 regions; edits skipped=0; log file=hdfs://hscale-dev1-nn:8020/apps/hbase/data/WALs/hscale-dev1-dn4,16020,1480686241892-splitting/hscale-dev1-dn4%2C16020%2C1480686241892.null13.1480686245360, length=91, corrupted=false, progress failed=false
2016-12-02 19:21:27,117 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-1] coordination.ZkSplitLogWorkerCoordination: successfully transitioned task /hbase-unsecure/splitWAL/WALs%2Fhscale-dev1-dn4%2C16020%2C1480686241892-splitting%2Fhscale-dev1-dn4%252C16020%252C1480686241892.null13.1480686245360 to final state DONE hscale-dev1-dn1,16020,1480686617338
2016-12-02 19:21:27,118 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-1] handler.WALSplitterHandler: worker hscale-dev1-dn1,16020,1480686617338 done with task org.apache.hadoop.hbase.coordination.ZkSplitLogWorkerCoordination$ZkSplitTaskDetails@2878abaf in 96ms
2016-12-02 19:21:27,593 INFO  [SplitLogWorker-hscale-dev1-dn1:16020] coordination.ZkSplitLogWorkerCoordination: worker hscale-dev1-dn1,16020,1480686617338 acquired task /hbase-unsecure/splitWAL/WALs%2Fhscale-dev1-dn4%2C16020%2C1480686241892-splitting%2Fhscale-dev1-dn4%252C16020%252C1480686241892.null14.1480686245431
2016-12-02 19:21:27,615 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-0] wal.WALSplitter: Splitting wal: hdfs://hscale-dev1-nn:8020/apps/hbase/data/WALs/hscale-dev1-dn4,16020,1480686241892-splitting/hscale-dev1-dn4%2C16020%2C1480686241892.null14.1480686245431, length=91
2016-12-02 19:21:27,616 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-0] wal.WALSplitter: DistributedLogReplay = false
2016-12-02 19:21:27,631 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-0] util.FSHDFSUtils: Recovering lease on dfs file hdfs://hscale-dev1-nn:8020/apps/hbase/data/WALs/hscale-dev1-dn4,16020,1480686241892-splitting/hscale-dev1-dn4%2C16020%2C1480686241892.null14.1480686245431
2016-12-02 19:21:27,633 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-0] util.FSHDFSUtils: recoverLease=true, attempt=0 on file=hdfs://hscale-dev1-nn:8020/apps/hbase/data/WALs/hscale-dev1-dn4,16020,1480686241892-splitting/hscale-dev1-dn4%2C16020%2C1480686241892.null14.1480686245431 after 2ms
2016-12-02 19:21:27,662 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-0] wal.WALSplitter: Split writers finished
2016-12-02 19:21:27,662 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-0] wal.WALSplitter: Processed 0 edits across 0 regions; edits skipped=0; log file=hdfs://hscale-dev1-nn:8020/apps/hbase/data/WALs/hscale-dev1-dn4,16020,1480686241892-splitting/hscale-dev1-dn4%2C16020%2C1480686241892.null14.1480686245431, length=91, corrupted=false, progress failed=false
2016-12-02 19:21:27,677 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-0] coordination.ZkSplitLogWorkerCoordination: successfully transitioned task /hbase-unsecure/splitWAL/WALs%2Fhscale-dev1-dn4%2C16020%2C1480686241892-splitting%2Fhscale-dev1-dn4%252C16020%252C1480686241892.null14.1480686245431 to final state DONE hscale-dev1-dn1,16020,1480686617338
2016-12-02 19:21:27,677 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-0] handler.WALSplitterHandler: worker hscale-dev1-dn1,16020,1480686617338 done with task org.apache.hadoop.hbase.coordination.ZkSplitLogWorkerCoordination$ZkSplitTaskDetails@6871add in 84ms
2016-12-02 19:21:28,305 INFO  [SplitLogWorker-hscale-dev1-dn1:16020] coordination.ZkSplitLogWorkerCoordination: worker hscale-dev1-dn1,16020,1480686617338 acquired task /hbase-unsecure/splitWAL/WALs%2Fhscale-dev1-dn1%2C16020%2C1480686239118-splitting%2Fhscale-dev1-dn1%252C16020%252C1480686239118.null1.1480686241768
2016-12-02 19:21:28,328 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-1] wal.WALSplitter: Splitting wal: hdfs://hscale-dev1-nn:8020/apps/hbase/data/WALs/hscale-dev1-dn1,16020,1480686239118-splitting/hscale-dev1-dn1%2C16020%2C1480686239118.null1.1480686241768, length=379
2016-12-02 19:21:28,328 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-1] wal.WALSplitter: DistributedLogReplay = false
2016-12-02 19:21:28,344 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-1] util.FSHDFSUtils: Recovering lease on dfs file hdfs://hscale-dev1-nn:8020/apps/hbase/data/WALs/hscale-dev1-dn1,16020,1480686239118-splitting/hscale-dev1-dn1%2C16020%2C1480686239118.null1.1480686241768
2016-12-02 19:21:28,345 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-1] util.FSHDFSUtils: recoverLease=true, attempt=0 on file=hdfs://hscale-dev1-nn:8020/apps/hbase/data/WALs/hscale-dev1-dn1,16020,1480686239118-splitting/hscale-dev1-dn1%2C16020%2C1480686239118.null1.1480686241768 after 1ms
2016-12-02 19:21:28,447 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-1-Writer-2] wal.WALSplitter: Creating writer path=hdfs://hscale-dev1-nn:8020/apps/hbase/data/data/splice/SPLICE_TXN/befbb055ce59b866459081d27215d354/recovered.edits/0000000000000000003.temp region=befbb055ce59b866459081d27215d354
2016-12-02 19:21:28,448 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-1] wal.WALSplitter: Split writers finished
2016-12-02 19:21:28,525 INFO  [split-log-closeStream-1] wal.WALSplitter: Rename hdfs://hscale-dev1-nn:8020/apps/hbase/data/data/splice/SPLICE_TXN/befbb055ce59b866459081d27215d354/recovered.edits/0000000000000000003.temp to hdfs://hscale-dev1-nn:8020/apps/hbase/data/data/splice/SPLICE_TXN/befbb055ce59b866459081d27215d354/recovered.edits/0000000000000000003
2016-12-02 19:21:28,526 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-1] wal.WALSplitter: Processed 1 edits across 1 regions; edits skipped=0; log file=hdfs://hscale-dev1-nn:8020/apps/hbase/data/WALs/hscale-dev1-dn1,16020,1480686239118-splitting/hscale-dev1-dn1%2C16020%2C1480686239118.null1.1480686241768, length=379, corrupted=false, progress failed=false
2016-12-02 19:21:28,540 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-1] coordination.ZkSplitLogWorkerCoordination: successfully transitioned task /hbase-unsecure/splitWAL/WALs%2Fhscale-dev1-dn1%2C16020%2C1480686239118-splitting%2Fhscale-dev1-dn1%252C16020%252C1480686239118.null1.1480686241768 to final state DONE hscale-dev1-dn1,16020,1480686617338
2016-12-02 19:21:28,540 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-1] handler.WALSplitterHandler: worker hscale-dev1-dn1,16020,1480686617338 done with task org.apache.hadoop.hbase.coordination.ZkSplitLogWorkerCoordination$ZkSplitTaskDetails@548de1a8 in 234ms
2016-12-02 19:21:29,033 INFO  [SplitLogWorker-hscale-dev1-dn1:16020] coordination.ZkSplitLogWorkerCoordination: worker hscale-dev1-dn1,16020,1480686617338 acquired task /hbase-unsecure/splitWAL/WALs%2Fhscale-dev1-dn4%2C16020%2C1480686241892-splitting%2Fhscale-dev1-dn4%252C16020%252C1480686241892.null1.1480686244397
2016-12-02 19:21:29,064 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-0] wal.WALSplitter: Splitting wal: hdfs://hscale-dev1-nn:8020/apps/hbase/data/WALs/hscale-dev1-dn4,16020,1480686241892-splitting/hscale-dev1-dn4%2C16020%2C1480686241892.null1.1480686244397, length=91
2016-12-02 19:21:29,064 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-0] wal.WALSplitter: DistributedLogReplay = false
2016-12-02 19:21:29,079 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-0] util.FSHDFSUtils: Recovering lease on dfs file hdfs://hscale-dev1-nn:8020/apps/hbase/data/WALs/hscale-dev1-dn4,16020,1480686241892-splitting/hscale-dev1-dn4%2C16020%2C1480686241892.null1.1480686244397
2016-12-02 19:21:29,080 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-0] util.FSHDFSUtils: recoverLease=true, attempt=0 on file=hdfs://hscale-dev1-nn:8020/apps/hbase/data/WALs/hscale-dev1-dn4,16020,1480686241892-splitting/hscale-dev1-dn4%2C16020%2C1480686241892.null1.1480686244397 after 1ms
2016-12-02 19:21:29,103 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-0] wal.WALSplitter: Split writers finished
2016-12-02 19:21:29,103 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-0] wal.WALSplitter: Processed 0 edits across 0 regions; edits skipped=0; log file=hdfs://hscale-dev1-nn:8020/apps/hbase/data/WALs/hscale-dev1-dn4,16020,1480686241892-splitting/hscale-dev1-dn4%2C16020%2C1480686241892.null1.1480686244397, length=91, corrupted=false, progress failed=false
2016-12-02 19:21:29,150 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-0] coordination.ZkSplitLogWorkerCoordination: successfully transitioned task /hbase-unsecure/splitWAL/WALs%2Fhscale-dev1-dn4%2C16020%2C1480686241892-splitting%2Fhscale-dev1-dn4%252C16020%252C1480686241892.null1.1480686244397 to final state DONE hscale-dev1-dn1,16020,1480686617338
2016-12-02 19:21:29,150 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-0] handler.WALSplitterHandler: worker hscale-dev1-dn1,16020,1480686617338 done with task org.apache.hadoop.hbase.coordination.ZkSplitLogWorkerCoordination$ZkSplitTaskDetails@31ba2a2e in 117ms
2016-12-02 19:21:29,930 INFO  [SplitLogWorker-hscale-dev1-dn1:16020] coordination.ZkSplitLogWorkerCoordination: worker hscale-dev1-dn1,16020,1480686617338 acquired task /hbase-unsecure/splitWAL/WALs%2Fhscale-dev1-dn1%2C16020%2C1480686239118-splitting%2Fhscale-dev1-dn1%252C16020%252C1480686239118.null13.1480686242647
2016-12-02 19:21:29,957 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-1] wal.WALSplitter: Splitting wal: hdfs://hscale-dev1-nn:8020/apps/hbase/data/WALs/hscale-dev1-dn1,16020,1480686239118-splitting/hscale-dev1-dn1%2C16020%2C1480686239118.null13.1480686242647, length=91
2016-12-02 19:21:29,957 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-1] wal.WALSplitter: DistributedLogReplay = false
2016-12-02 19:21:29,974 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-1] util.FSHDFSUtils: Recovering lease on dfs file hdfs://hscale-dev1-nn:8020/apps/hbase/data/WALs/hscale-dev1-dn1,16020,1480686239118-splitting/hscale-dev1-dn1%2C16020%2C1480686239118.null13.1480686242647
2016-12-02 19:21:29,976 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-1] util.FSHDFSUtils: recoverLease=true, attempt=0 on file=hdfs://hscale-dev1-nn:8020/apps/hbase/data/WALs/hscale-dev1-dn1,16020,1480686239118-splitting/hscale-dev1-dn1%2C16020%2C1480686239118.null13.1480686242647 after 2ms
2016-12-02 19:21:30,002 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-1] wal.WALSplitter: Split writers finished
2016-12-02 19:21:30,002 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-1] wal.WALSplitter: Processed 0 edits across 0 regions; edits skipped=0; log file=hdfs://hscale-dev1-nn:8020/apps/hbase/data/WALs/hscale-dev1-dn1,16020,1480686239118-splitting/hscale-dev1-dn1%2C16020%2C1480686239118.null13.1480686242647, length=91, corrupted=false, progress failed=false
2016-12-02 19:21:30,017 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-1] coordination.ZkSplitLogWorkerCoordination: successfully transitioned task /hbase-unsecure/splitWAL/WALs%2Fhscale-dev1-dn1%2C16020%2C1480686239118-splitting%2Fhscale-dev1-dn1%252C16020%252C1480686239118.null13.1480686242647 to final state DONE hscale-dev1-dn1,16020,1480686617338
2016-12-02 19:21:30,017 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-1] handler.WALSplitterHandler: worker hscale-dev1-dn1,16020,1480686617338 done with task org.apache.hadoop.hbase.coordination.ZkSplitLogWorkerCoordination$ZkSplitTaskDetails@13bd7b35 in 87ms
2016-12-02 19:21:30,809 INFO  [SplitLogWorker-hscale-dev1-dn1:16020] coordination.ZkSplitLogWorkerCoordination: worker hscale-dev1-dn1,16020,1480686617338 acquired task /hbase-unsecure/splitWAL/WALs%2Fhscale-dev1-dn1%2C16020%2C1480686239118-splitting%2Fhscale-dev1-dn1%252C16020%252C1480686239118.null10.1480686242397
2016-12-02 19:21:30,828 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-0] wal.WALSplitter: Splitting wal: hdfs://hscale-dev1-nn:8020/apps/hbase/data/WALs/hscale-dev1-dn1,16020,1480686239118-splitting/hscale-dev1-dn1%2C16020%2C1480686239118.null10.1480686242397, length=91
2016-12-02 19:21:30,828 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-0] wal.WALSplitter: DistributedLogReplay = false
2016-12-02 19:21:30,846 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-0] util.FSHDFSUtils: Recovering lease on dfs file hdfs://hscale-dev1-nn:8020/apps/hbase/data/WALs/hscale-dev1-dn1,16020,1480686239118-splitting/hscale-dev1-dn1%2C16020%2C1480686239118.null10.1480686242397
2016-12-02 19:21:30,847 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-0] util.FSHDFSUtils: recoverLease=true, attempt=0 on file=hdfs://hscale-dev1-nn:8020/apps/hbase/data/WALs/hscale-dev1-dn1,16020,1480686239118-splitting/hscale-dev1-dn1%2C16020%2C1480686239118.null10.1480686242397 after 1ms
2016-12-02 19:21:30,872 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-0] wal.WALSplitter: Split writers finished
2016-12-02 19:21:30,872 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-0] wal.WALSplitter: Processed 0 edits across 0 regions; edits skipped=0; log file=hdfs://hscale-dev1-nn:8020/apps/hbase/data/WALs/hscale-dev1-dn1,16020,1480686239118-splitting/hscale-dev1-dn1%2C16020%2C1480686239118.null10.1480686242397, length=91, corrupted=false, progress failed=false
2016-12-02 19:21:30,888 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-0] coordination.ZkSplitLogWorkerCoordination: successfully transitioned task /hbase-unsecure/splitWAL/WALs%2Fhscale-dev1-dn1%2C16020%2C1480686239118-splitting%2Fhscale-dev1-dn1%252C16020%252C1480686239118.null10.1480686242397 to final state DONE hscale-dev1-dn1,16020,1480686617338
2016-12-02 19:21:30,888 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-0] handler.WALSplitterHandler: worker hscale-dev1-dn1,16020,1480686617338 done with task org.apache.hadoop.hbase.coordination.ZkSplitLogWorkerCoordination$ZkSplitTaskDetails@1cedd9f in 79ms
2016-12-02 19:21:31,060 ERROR [hbase-region-load-updater-0] hbase.HBaseRegionLoads: Unable to fetch region load info
org.apache.hadoop.hbase.client.RetriesExhaustedException: Failed after attempts=5, exceptions:
Fri Dec 02 19:21:29 IST 2016, RpcRetryingCaller{globalStartTime=1480686689019, pause=100, retries=5}, org.apache.hadoop.hbase.ipc.RemoteWithExtrasException(org.apache.hadoop.hbase.PleaseHoldException): org.apache.hadoop.hbase.PleaseHoldException: Master is initializing
    at org.apache.hadoop.hbase.master.HMaster.checkInitialized(HMaster.java:2324)
    at org.apache.hadoop.hbase.master.MasterRpcServices.getClusterStatus(MasterRpcServices.java:770)
    at org.apache.hadoop.hbase.protobuf.generated.MasterProtos$MasterService$2.callBlockingMethod(MasterProtos.java:55371)
    at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2114)
    at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:101)
    at org.apache.hadoop.hbase.ipc.RpcExecutor.consumerLoop(RpcExecutor.java:130)
    at org.apache.hadoop.hbase.ipc.RpcExecutor$1.run(RpcExecutor.java:107)
    at java.lang.Thread.run(Thread.java:745)

Fri Dec 02 19:21:29 IST 2016, RpcRetryingCaller{globalStartTime=1480686689019, pause=100, retries=5}, org.apache.hadoop.hbase.ipc.RemoteWithExtrasException(org.apache.hadoop.hbase.PleaseHoldException): org.apache.hadoop.hbase.PleaseHoldException: Master is initializing
    at org.apache.hadoop.hbase.master.HMaster.checkInitialized(HMaster.java:2324)
    at org.apache.hadoop.hbase.master.MasterRpcServices.getClusterStatus(MasterRpcServices.java:770)
    at org.apache.hadoop.hbase.protobuf.generated.MasterProtos$MasterService$2.callBlockingMethod(MasterProtos.java:55371)
    at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2114)
    at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:101)
    at org.apache.hadoop.hbase.ipc.RpcExecutor.consumerLoop(RpcExecutor.java:130)
    at org.apache.hadoop.hbase.ipc.RpcExecutor$1.run(RpcExecutor.java:107)
    at java.lang.Thread.run(Thread.java:745)

Fri Dec 02 19:21:29 IST 2016, RpcRetryingCaller{globalStartTime=1480686689019, pause=100, retries=5}, org.apache.hadoop.hbase.ipc.RemoteWithExtrasException(org.apache.hadoop.hbase.PleaseHoldException): org.apache.hadoop.hbase.PleaseHoldException: Master is initializing
    at org.apache.hadoop.hbase.master.HMaster.checkInitialized(HMaster.java:2324)
    at org.apache.hadoop.hbase.master.MasterRpcServices.getClusterStatus(MasterRpcServices.java:770)
    at org.apache.hadoop.hbase.protobuf.generated.MasterProtos$MasterService$2.callBlockingMethod(MasterProtos.java:55371)
    at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2114)
    at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:101)
    at org.apache.hadoop.hbase.ipc.RpcExecutor.consumerLoop(RpcExecutor.java:130)
    at org.apache.hadoop.hbase.ipc.RpcExecutor$1.run(RpcExecutor.java:107)
    at java.lang.Thread.run(Thread.java:745)

Fri Dec 02 19:21:30 IST 2016, RpcRetryingCaller{globalStartTime=1480686689019, pause=100, retries=5}, org.apache.hadoop.hbase.ipc.RemoteWithExtrasException(org.apache.hadoop.hbase.PleaseHoldException): org.apache.hadoop.hbase.PleaseHoldException: Master is initializing
    at org.apache.hadoop.hbase.master.HMaster.checkInitialized(HMaster.java:2324)
    at org.apache.hadoop.hbase.master.MasterRpcServices.getClusterStatus(MasterRpcServices.java:770)
    at org.apache.hadoop.hbase.protobuf.generated.MasterProtos$MasterService$2.callBlockingMethod(MasterProtos.java:55371)
    at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2114)
    at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:101)
    at org.apache.hadoop.hbase.ipc.RpcExecutor.consumerLoop(RpcExecutor.java:130)
    at org.apache.hadoop.hbase.ipc.RpcExecutor$1.run(RpcExecutor.java:107)
    at java.lang.Thread.run(Thread.java:745)

Fri Dec 02 19:21:31 IST 2016, RpcRetryingCaller{globalStartTime=1480686689019, pause=100, retries=5}, org.apache.hadoop.hbase.ipc.RemoteWithExtrasException(org.apache.hadoop.hbase.PleaseHoldException): org.apache.hadoop.hbase.PleaseHoldException: Master is initializing
    at org.apache.hadoop.hbase.master.HMaster.checkInitialized(HMaster.java:2324)
    at org.apache.hadoop.hbase.master.MasterRpcServices.getClusterStatus(MasterRpcServices.java:770)
    at org.apache.hadoop.hbase.protobuf.generated.MasterProtos$MasterService$2.callBlockingMethod(MasterProtos.java:55371)
    at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2114)
    at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:101)
    at org.apache.hadoop.hbase.ipc.RpcExecutor.consumerLoop(RpcExecutor.java:130)
    at org.apache.hadoop.hbase.ipc.RpcExecutor$1.run(RpcExecutor.java:107)
    at java.lang.Thread.run(Thread.java:745)


    at org.apache.hadoop.hbase.client.RpcRetryingCaller.callWithRetries(RpcRetryingCaller.java:147)
    at org.apache.hadoop.hbase.client.HBaseAdmin.executeCallable(HBaseAdmin.java:4083)
    at org.apache.hadoop.hbase.client.HBaseAdmin.getClusterStatus(HBaseAdmin.java:2765)
    at com.splicemachine.access.hbase.H10PartitionAdmin.allServers(H10PartitionAdmin.java:120)
    at com.splicemachine.hbase.HBaseRegionLoads.fetchRegionLoads(HBaseRegionLoads.java:164)
    at com.splicemachine.hbase.HBaseRegionLoads.access$000(HBaseRegionLoads.java:60)
    at com.splicemachine.hbase.HBaseRegionLoads$1.run(HBaseRegionLoads.java:81)
    at com.splicemachine.concurrent.LoggingScheduledThreadPoolExecutor$LoggingRunnable.run(LoggingScheduledThreadPoolExecutor.java:75)
    at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
    at java.util.concurrent.FutureTask.runAndReset(FutureTask.java:308)
    at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$301(ScheduledThreadPoolExecutor.java:180)
    at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:294)
    at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
    at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
    at java.lang.Thread.run(Thread.java:745)
Caused by: org.apache.hadoop.hbase.ipc.RemoteWithExtrasException(org.apache.hadoop.hbase.PleaseHoldException): org.apache.hadoop.hbase.PleaseHoldException: Master is initializing
    at org.apache.hadoop.hbase.master.HMaster.checkInitialized(HMaster.java:2324)
    at org.apache.hadoop.hbase.master.MasterRpcServices.getClusterStatus(MasterRpcServices.java:770)
    at org.apache.hadoop.hbase.protobuf.generated.MasterProtos$MasterService$2.callBlockingMethod(MasterProtos.java:55371)
    at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2114)
    at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:101)
    at org.apache.hadoop.hbase.ipc.RpcExecutor.consumerLoop(RpcExecutor.java:130)
    at org.apache.hadoop.hbase.ipc.RpcExecutor$1.run(RpcExecutor.java:107)
    at java.lang.Thread.run(Thread.java:745)

    at org.apache.hadoop.hbase.ipc.RpcClientImpl.call(RpcClientImpl.java:1226)
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient.callBlockingMethod(AbstractRpcClient.java:213)
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient$BlockingRpcChannelImplementation.callBlockingMethod(AbstractRpcClient.java:287)
    at org.apache.hadoop.hbase.protobuf.generated.MasterProtos$MasterService$BlockingStub.getClusterStatus(MasterProtos.java:58140)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation$4.getClusterStatus(ConnectionManager.java:2036)
    at org.apache.hadoop.hbase.client.HBaseAdmin$33.call(HBaseAdmin.java:2769)
    at org.apache.hadoop.hbase.client.HBaseAdmin$33.call(HBaseAdmin.java:2765)
    at org.apache.hadoop.hbase.client.RpcRetryingCaller.callWithRetries(RpcRetryingCaller.java:126)
    ... 14 more
2016-12-02 19:21:31,581 INFO  [SplitLogWorker-hscale-dev1-dn1:16020] coordination.ZkSplitLogWorkerCoordination: worker hscale-dev1-dn1,16020,1480686617338 acquired task /hbase-unsecure/splitWAL/WALs%2Fhscale-dev1-dn2%2C16020%2C1480686240889-splitting%2Fhscale-dev1-dn2%252C16020%252C1480686240889.null4.1480686243625
2016-12-02 19:21:31,607 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-1] wal.WALSplitter: Splitting wal: hdfs://hscale-dev1-nn:8020/apps/hbase/data/WALs/hscale-dev1-dn2,16020,1480686240889-splitting/hscale-dev1-dn2%2C16020%2C1480686240889.null4.1480686243625, length=380
2016-12-02 19:21:31,607 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-1] wal.WALSplitter: DistributedLogReplay = false
2016-12-02 19:21:31,625 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-1] util.FSHDFSUtils: Recovering lease on dfs file hdfs://hscale-dev1-nn:8020/apps/hbase/data/WALs/hscale-dev1-dn2,16020,1480686240889-splitting/hscale-dev1-dn2%2C16020%2C1480686240889.null4.1480686243625
2016-12-02 19:21:31,626 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-1] util.FSHDFSUtils: recoverLease=true, attempt=0 on file=hdfs://hscale-dev1-nn:8020/apps/hbase/data/WALs/hscale-dev1-dn2,16020,1480686240889-splitting/hscale-dev1-dn2%2C16020%2C1480686240889.null4.1480686243625 after 1ms
2016-12-02 19:21:31,685 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-1-Writer-2] wal.WALSplitter: Creating writer path=hdfs://hscale-dev1-nn:8020/apps/hbase/data/data/splice/SPLICE_TXN/f5b5be3447bafc9173850bc0400ff334/recovered.edits/0000000000000000003.temp region=f5b5be3447bafc9173850bc0400ff334
2016-12-02 19:21:31,685 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-1] wal.WALSplitter: Split writers finished
2016-12-02 19:21:31,750 INFO  [split-log-closeStream-1] wal.WALSplitter: Rename hdfs://hscale-dev1-nn:8020/apps/hbase/data/data/splice/SPLICE_TXN/f5b5be3447bafc9173850bc0400ff334/recovered.edits/0000000000000000003.temp to hdfs://hscale-dev1-nn:8020/apps/hbase/data/data/splice/SPLICE_TXN/f5b5be3447bafc9173850bc0400ff334/recovered.edits/0000000000000000003
2016-12-02 19:21:31,750 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-1] wal.WALSplitter: Processed 1 edits across 1 regions; edits skipped=0; log file=hdfs://hscale-dev1-nn:8020/apps/hbase/data/WALs/hscale-dev1-dn2,16020,1480686240889-splitting/hscale-dev1-dn2%2C16020%2C1480686240889.null4.1480686243625, length=380, corrupted=false, progress failed=false
2016-12-02 19:21:31,765 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-1] coordination.ZkSplitLogWorkerCoordination: successfully transitioned task /hbase-unsecure/splitWAL/WALs%2Fhscale-dev1-dn2%2C16020%2C1480686240889-splitting%2Fhscale-dev1-dn2%252C16020%252C1480686240889.null4.1480686243625 to final state DONE hscale-dev1-dn1,16020,1480686617338
2016-12-02 19:21:31,765 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-1] handler.WALSplitterHandler: worker hscale-dev1-dn1,16020,1480686617338 done with task org.apache.hadoop.hbase.coordination.ZkSplitLogWorkerCoordination$ZkSplitTaskDetails@42bc84c6 in 184ms
2016-12-02 19:21:32,211 INFO  [SplitLogWorker-hscale-dev1-dn1:16020] coordination.ZkSplitLogWorkerCoordination: worker hscale-dev1-dn1,16020,1480686617338 acquired task /hbase-unsecure/splitWAL/WALs%2Fhscale-dev1-dn2%2C16020%2C1480686240889-splitting%2Fhscale-dev1-dn2%252C16020%252C1480686240889.null12.1480686244284
2016-12-02 19:21:32,235 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-0] wal.WALSplitter: Splitting wal: hdfs://hscale-dev1-nn:8020/apps/hbase/data/WALs/hscale-dev1-dn2,16020,1480686240889-splitting/hscale-dev1-dn2%2C16020%2C1480686240889.null12.1480686244284, length=91
2016-12-02 19:21:32,235 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-0] wal.WALSplitter: DistributedLogReplay = false
2016-12-02 19:21:32,250 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-0] util.FSHDFSUtils: Recovering lease on dfs file hdfs://hscale-dev1-nn:8020/apps/hbase/data/WALs/hscale-dev1-dn2,16020,1480686240889-splitting/hscale-dev1-dn2%2C16020%2C1480686240889.null12.1480686244284
2016-12-02 19:21:32,252 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-0] util.FSHDFSUtils: recoverLease=true, attempt=0 on file=hdfs://hscale-dev1-nn:8020/apps/hbase/data/WALs/hscale-dev1-dn2,16020,1480686240889-splitting/hscale-dev1-dn2%2C16020%2C1480686240889.null12.1480686244284 after 2ms
2016-12-02 19:21:32,279 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-0] wal.WALSplitter: Split writers finished
2016-12-02 19:21:32,279 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-0] wal.WALSplitter: Processed 0 edits across 0 regions; edits skipped=0; log file=hdfs://hscale-dev1-nn:8020/apps/hbase/data/WALs/hscale-dev1-dn2,16020,1480686240889-splitting/hscale-dev1-dn2%2C16020%2C1480686240889.null12.1480686244284, length=91, corrupted=false, progress failed=false
2016-12-02 19:21:32,295 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-0] coordination.ZkSplitLogWorkerCoordination: successfully transitioned task /hbase-unsecure/splitWAL/WALs%2Fhscale-dev1-dn2%2C16020%2C1480686240889-splitting%2Fhscale-dev1-dn2%252C16020%252C1480686240889.null12.1480686244284 to final state DONE hscale-dev1-dn1,16020,1480686617338
2016-12-02 19:21:32,295 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-0] handler.WALSplitterHandler: worker hscale-dev1-dn1,16020,1480686617338 done with task org.apache.hadoop.hbase.coordination.ZkSplitLogWorkerCoordination$ZkSplitTaskDetails@47ed54de in 84ms
2016-12-02 19:21:33,212 INFO  [SplitLogWorker-hscale-dev1-dn1:16020] coordination.ZkSplitLogWorkerCoordination: worker hscale-dev1-dn1,16020,1480686617338 acquired task /hbase-unsecure/splitWAL/WALs%2Fhscale-dev1-dn1%2C16020%2C1480686239118-splitting%2Fhscale-dev1-dn1%252C16020%252C1480686239118.null14.1480686242725
2016-12-02 19:21:33,239 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-1] wal.WALSplitter: Splitting wal: hdfs://hscale-dev1-nn:8020/apps/hbase/data/WALs/hscale-dev1-dn1,16020,1480686239118-splitting/hscale-dev1-dn1%2C16020%2C1480686239118.null14.1480686242725, length=91
2016-12-02 19:21:33,239 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-1] wal.WALSplitter: DistributedLogReplay = false
2016-12-02 19:21:33,256 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-1] util.FSHDFSUtils: Recovering lease on dfs file hdfs://hscale-dev1-nn:8020/apps/hbase/data/WALs/hscale-dev1-dn1,16020,1480686239118-splitting/hscale-dev1-dn1%2C16020%2C1480686239118.null14.1480686242725
2016-12-02 19:21:33,257 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-1] util.FSHDFSUtils: recoverLease=true, attempt=0 on file=hdfs://hscale-dev1-nn:8020/apps/hbase/data/WALs/hscale-dev1-dn1,16020,1480686239118-splitting/hscale-dev1-dn1%2C16020%2C1480686239118.null14.1480686242725 after 1ms
2016-12-02 19:21:33,285 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-1] wal.WALSplitter: Split writers finished
2016-12-02 19:21:33,285 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-1] wal.WALSplitter: Processed 0 edits across 0 regions; edits skipped=0; log file=hdfs://hscale-dev1-nn:8020/apps/hbase/data/WALs/hscale-dev1-dn1,16020,1480686239118-splitting/hscale-dev1-dn1%2C16020%2C1480686239118.null14.1480686242725, length=91, corrupted=false, progress failed=false
2016-12-02 19:21:33,303 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-1] coordination.ZkSplitLogWorkerCoordination: successfully transitioned task /hbase-unsecure/splitWAL/WALs%2Fhscale-dev1-dn1%2C16020%2C1480686239118-splitting%2Fhscale-dev1-dn1%252C16020%252C1480686239118.null14.1480686242725 to final state DONE hscale-dev1-dn1,16020,1480686617338
2016-12-02 19:21:33,303 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-1] handler.WALSplitterHandler: worker hscale-dev1-dn1,16020,1480686617338 done with task org.apache.hadoop.hbase.coordination.ZkSplitLogWorkerCoordination$ZkSplitTaskDetails@3ca28517 in 91ms
2016-12-02 19:21:34,070 INFO  [SplitLogWorker-hscale-dev1-dn1:16020] coordination.ZkSplitLogWorkerCoordination: worker hscale-dev1-dn1,16020,1480686617338 acquired task /hbase-unsecure/splitWAL/WALs%2Fhscale-dev1-dn1%2C16020%2C1480686239118-splitting%2Fhscale-dev1-dn1%252C16020%252C1480686239118.null7.1480686242154
2016-12-02 19:21:34,094 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-0] wal.WALSplitter: Splitting wal: hdfs://hscale-dev1-nn:8020/apps/hbase/data/WALs/hscale-dev1-dn1,16020,1480686239118-splitting/hscale-dev1-dn1%2C16020%2C1480686239118.null7.1480686242154, length=389
2016-12-02 19:21:34,094 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-0] wal.WALSplitter: DistributedLogReplay = false
2016-12-02 19:21:34,109 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-0] util.FSHDFSUtils: Recovering lease on dfs file hdfs://hscale-dev1-nn:8020/apps/hbase/data/WALs/hscale-dev1-dn1,16020,1480686239118-splitting/hscale-dev1-dn1%2C16020%2C1480686239118.null7.1480686242154
2016-12-02 19:21:34,110 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-0] util.FSHDFSUtils: recoverLease=true, attempt=0 on file=hdfs://hscale-dev1-nn:8020/apps/hbase/data/WALs/hscale-dev1-dn1,16020,1480686239118-splitting/hscale-dev1-dn1%2C16020%2C1480686239118.null7.1480686242154 after 1ms
2016-12-02 19:21:34,165 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-0-Writer-2] wal.WALSplitter: Creating writer path=hdfs://hscale-dev1-nn:8020/apps/hbase/data/data/splice/SPLICE_SEQUENCES/969f9fff45a516829701ab973a11c9ce/recovered.edits/0000000000000000003.temp region=969f9fff45a516829701ab973a11c9ce
2016-12-02 19:21:34,165 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-0] wal.WALSplitter: Split writers finished
2016-12-02 19:21:34,227 INFO  [split-log-closeStream-1] wal.WALSplitter: Rename hdfs://hscale-dev1-nn:8020/apps/hbase/data/data/splice/SPLICE_SEQUENCES/969f9fff45a516829701ab973a11c9ce/recovered.edits/0000000000000000003.temp to hdfs://hscale-dev1-nn:8020/apps/hbase/data/data/splice/SPLICE_SEQUENCES/969f9fff45a516829701ab973a11c9ce/recovered.edits/0000000000000000003
2016-12-02 19:21:34,227 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-0] wal.WALSplitter: Processed 1 edits across 1 regions; edits skipped=0; log file=hdfs://hscale-dev1-nn:8020/apps/hbase/data/WALs/hscale-dev1-dn1,16020,1480686239118-splitting/hscale-dev1-dn1%2C16020%2C1480686239118.null7.1480686242154, length=389, corrupted=false, progress failed=false
2016-12-02 19:21:34,249 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-0] coordination.ZkSplitLogWorkerCoordination: successfully transitioned task /hbase-unsecure/splitWAL/WALs%2Fhscale-dev1-dn1%2C16020%2C1480686239118-splitting%2Fhscale-dev1-dn1%252C16020%252C1480686239118.null7.1480686242154 to final state DONE hscale-dev1-dn1,16020,1480686617338
2016-12-02 19:21:34,249 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-0] handler.WALSplitterHandler: worker hscale-dev1-dn1,16020,1480686617338 done with task org.apache.hadoop.hbase.coordination.ZkSplitLogWorkerCoordination$ZkSplitTaskDetails@7b50c47e in 178ms
2016-12-02 19:21:34,971 INFO  [SplitLogWorker-hscale-dev1-dn1:16020] coordination.ZkSplitLogWorkerCoordination: worker hscale-dev1-dn1,16020,1480686617338 acquired task /hbase-unsecure/splitWAL/WALs%2Fhscale-dev1-dn3%2C16020%2C1480686232696-splitting%2Fhscale-dev1-dn3%252C16020%252C1480686232696.null3.1480686235333
2016-12-02 19:21:35,071 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-1] wal.WALSplitter: Splitting wal: hdfs://hscale-dev1-nn:8020/apps/hbase/data/WALs/hscale-dev1-dn3,16020,1480686232696-splitting/hscale-dev1-dn3%2C16020%2C1480686232696.null3.1480686235333, length=380
2016-12-02 19:21:35,071 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-1] wal.WALSplitter: DistributedLogReplay = false
2016-12-02 19:21:35,088 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-1] util.FSHDFSUtils: Recovering lease on dfs file hdfs://hscale-dev1-nn:8020/apps/hbase/data/WALs/hscale-dev1-dn3,16020,1480686232696-splitting/hscale-dev1-dn3%2C16020%2C1480686232696.null3.1480686235333
2016-12-02 19:21:35,089 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-1] util.FSHDFSUtils: recoverLease=true, attempt=0 on file=hdfs://hscale-dev1-nn:8020/apps/hbase/data/WALs/hscale-dev1-dn3,16020,1480686232696-splitting/hscale-dev1-dn3%2C16020%2C1480686232696.null3.1480686235333 after 1ms
2016-12-02 19:21:35,153 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-1-Writer-2] wal.WALSplitter: Creating writer path=hdfs://hscale-dev1-nn:8020/apps/hbase/data/data/splice/SPLICE_TXN/e480b1b3cb833cb61cea26e4c0aa966e/recovered.edits/0000000000000000003.temp region=e480b1b3cb833cb61cea26e4c0aa966e
2016-12-02 19:21:35,153 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-1] wal.WALSplitter: Split writers finished
2016-12-02 19:21:35,234 INFO  [split-log-closeStream-1] wal.WALSplitter: Rename hdfs://hscale-dev1-nn:8020/apps/hbase/data/data/splice/SPLICE_TXN/e480b1b3cb833cb61cea26e4c0aa966e/recovered.edits/0000000000000000003.temp to hdfs://hscale-dev1-nn:8020/apps/hbase/data/data/splice/SPLICE_TXN/e480b1b3cb833cb61cea26e4c0aa966e/recovered.edits/0000000000000000003
2016-12-02 19:21:35,234 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-1] wal.WALSplitter: Processed 1 edits across 1 regions; edits skipped=0; log file=hdfs://hscale-dev1-nn:8020/apps/hbase/data/WALs/hscale-dev1-dn3,16020,1480686232696-splitting/hscale-dev1-dn3%2C16020%2C1480686232696.null3.1480686235333, length=380, corrupted=false, progress failed=false
2016-12-02 19:21:35,255 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-1] coordination.ZkSplitLogWorkerCoordination: successfully transitioned task /hbase-unsecure/splitWAL/WALs%2Fhscale-dev1-dn3%2C16020%2C1480686232696-splitting%2Fhscale-dev1-dn3%252C16020%252C1480686232696.null3.1480686235333 to final state DONE hscale-dev1-dn1,16020,1480686617338
2016-12-02 19:21:35,255 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-1] handler.WALSplitterHandler: worker hscale-dev1-dn1,16020,1480686617338 done with task org.apache.hadoop.hbase.coordination.ZkSplitLogWorkerCoordination$ZkSplitTaskDetails@253c9c33 in 283ms
2016-12-02 19:21:35,871 INFO  [SplitLogWorker-hscale-dev1-dn1:16020] coordination.ZkSplitLogWorkerCoordination: worker hscale-dev1-dn1,16020,1480686617338 acquired task /hbase-unsecure/splitWAL/WALs%2Fhscale-dev1-dn4%2C16020%2C1480686241892-splitting%2Fhscale-dev1-dn4%252C16020%252C1480686241892.null12.1480686245260
2016-12-02 19:21:35,895 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-0] wal.WALSplitter: Splitting wal: hdfs://hscale-dev1-nn:8020/apps/hbase/data/WALs/hscale-dev1-dn4,16020,1480686241892-splitting/hscale-dev1-dn4%2C16020%2C1480686241892.null12.1480686245260, length=91
2016-12-02 19:21:35,895 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-0] wal.WALSplitter: DistributedLogReplay = false
2016-12-02 19:21:35,909 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-0] util.FSHDFSUtils: Recovering lease on dfs file hdfs://hscale-dev1-nn:8020/apps/hbase/data/WALs/hscale-dev1-dn4,16020,1480686241892-splitting/hscale-dev1-dn4%2C16020%2C1480686241892.null12.1480686245260
2016-12-02 19:21:35,911 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-0] util.FSHDFSUtils: recoverLease=true, attempt=0 on file=hdfs://hscale-dev1-nn:8020/apps/hbase/data/WALs/hscale-dev1-dn4,16020,1480686241892-splitting/hscale-dev1-dn4%2C16020%2C1480686241892.null12.1480686245260 after 2ms
2016-12-02 19:21:35,939 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-0] wal.WALSplitter: Split writers finished
2016-12-02 19:21:35,939 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-0] wal.WALSplitter: Processed 0 edits across 0 regions; edits skipped=0; log file=hdfs://hscale-dev1-nn:8020/apps/hbase/data/WALs/hscale-dev1-dn4,16020,1480686241892-splitting/hscale-dev1-dn4%2C16020%2C1480686241892.null12.1480686245260, length=91, corrupted=false, progress failed=false
2016-12-02 19:21:35,954 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-0] coordination.ZkSplitLogWorkerCoordination: successfully transitioned task /hbase-unsecure/splitWAL/WALs%2Fhscale-dev1-dn4%2C16020%2C1480686241892-splitting%2Fhscale-dev1-dn4%252C16020%252C1480686241892.null12.1480686245260 to final state DONE hscale-dev1-dn1,16020,1480686617338
2016-12-02 19:21:35,954 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-0] handler.WALSplitterHandler: worker hscale-dev1-dn1,16020,1480686617338 done with task org.apache.hadoop.hbase.coordination.ZkSplitLogWorkerCoordination$ZkSplitTaskDetails@6a2e50da in 83ms
2016-12-02 19:21:36,077 ERROR [hbase-region-load-updater-0] hbase.HBaseRegionLoads: Unable to fetch region load info
org.apache.hadoop.hbase.client.RetriesExhaustedException: Failed after attempts=5, exceptions:
Fri Dec 02 19:21:34 IST 2016, RpcRetryingCaller{globalStartTime=1480686694019, pause=100, retries=5}, org.apache.hadoop.hbase.ipc.RemoteWithExtrasException(org.apache.hadoop.hbase.PleaseHoldException): org.apache.hadoop.hbase.PleaseHoldException: Master is initializing
    at org.apache.hadoop.hbase.master.HMaster.checkInitialized(HMaster.java:2324)
    at org.apache.hadoop.hbase.master.MasterRpcServices.getClusterStatus(MasterRpcServices.java:770)
    at org.apache.hadoop.hbase.protobuf.generated.MasterProtos$MasterService$2.callBlockingMethod(MasterProtos.java:55371)
    at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2114)
    at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:101)
    at org.apache.hadoop.hbase.ipc.RpcExecutor.consumerLoop(RpcExecutor.java:130)
    at org.apache.hadoop.hbase.ipc.RpcExecutor$1.run(RpcExecutor.java:107)
    at java.lang.Thread.run(Thread.java:745)

Fri Dec 02 19:21:34 IST 2016, RpcRetryingCaller{globalStartTime=1480686694019, pause=100, retries=5}, org.apache.hadoop.hbase.ipc.RemoteWithExtrasException(org.apache.hadoop.hbase.PleaseHoldException): org.apache.hadoop.hbase.PleaseHoldException: Master is initializing
    at org.apache.hadoop.hbase.master.HMaster.checkInitialized(HMaster.java:2324)
    at org.apache.hadoop.hbase.master.MasterRpcServices.getClusterStatus(MasterRpcServices.java:770)
    at org.apache.hadoop.hbase.protobuf.generated.MasterProtos$MasterService$2.callBlockingMethod(MasterProtos.java:55371)
    at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2114)
    at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:101)
    at org.apache.hadoop.hbase.ipc.RpcExecutor.consumerLoop(RpcExecutor.java:130)
    at org.apache.hadoop.hbase.ipc.RpcExecutor$1.run(RpcExecutor.java:107)
    at java.lang.Thread.run(Thread.java:745)

Fri Dec 02 19:21:34 IST 2016, RpcRetryingCaller{globalStartTime=1480686694019, pause=100, retries=5}, org.apache.hadoop.hbase.ipc.RemoteWithExtrasException(org.apache.hadoop.hbase.PleaseHoldException): org.apache.hadoop.hbase.PleaseHoldException: Master is initializing
    at org.apache.hadoop.hbase.master.HMaster.checkInitialized(HMaster.java:2324)
    at org.apache.hadoop.hbase.master.MasterRpcServices.getClusterStatus(MasterRpcServices.java:770)
    at org.apache.hadoop.hbase.protobuf.generated.MasterProtos$MasterService$2.callBlockingMethod(MasterProtos.java:55371)
    at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2114)
    at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:101)
    at org.apache.hadoop.hbase.ipc.RpcExecutor.consumerLoop(RpcExecutor.java:130)
    at org.apache.hadoop.hbase.ipc.RpcExecutor$1.run(RpcExecutor.java:107)
    at java.lang.Thread.run(Thread.java:745)

Fri Dec 02 19:21:35 IST 2016, RpcRetryingCaller{globalStartTime=1480686694019, pause=100, retries=5}, org.apache.hadoop.hbase.ipc.RemoteWithExtrasException(org.apache.hadoop.hbase.PleaseHoldException): org.apache.hadoop.hbase.PleaseHoldException: Master is initializing
    at org.apache.hadoop.hbase.master.HMaster.checkInitialized(HMaster.java:2324)
    at org.apache.hadoop.hbase.master.MasterRpcServices.getClusterStatus(MasterRpcServices.java:770)
    at org.apache.hadoop.hbase.protobuf.generated.MasterProtos$MasterService$2.callBlockingMethod(MasterProtos.java:55371)
    at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2114)
    at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:101)
    at org.apache.hadoop.hbase.ipc.RpcExecutor.consumerLoop(RpcExecutor.java:130)
    at org.apache.hadoop.hbase.ipc.RpcExecutor$1.run(RpcExecutor.java:107)
    at java.lang.Thread.run(Thread.java:745)

Fri Dec 02 19:21:36 IST 2016, RpcRetryingCaller{globalStartTime=1480686694019, pause=100, retries=5}, org.apache.hadoop.hbase.ipc.RemoteWithExtrasException(org.apache.hadoop.hbase.PleaseHoldException): org.apache.hadoop.hbase.PleaseHoldException: Master is initializing
    at org.apache.hadoop.hbase.master.HMaster.checkInitialized(HMaster.java:2324)
    at org.apache.hadoop.hbase.master.MasterRpcServices.getClusterStatus(MasterRpcServices.java:770)
    at org.apache.hadoop.hbase.protobuf.generated.MasterProtos$MasterService$2.callBlockingMethod(MasterProtos.java:55371)
    at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2114)
    at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:101)
    at org.apache.hadoop.hbase.ipc.RpcExecutor.consumerLoop(RpcExecutor.java:130)
    at org.apache.hadoop.hbase.ipc.RpcExecutor$1.run(RpcExecutor.java:107)
    at java.lang.Thread.run(Thread.java:745)


    at org.apache.hadoop.hbase.client.RpcRetryingCaller.callWithRetries(RpcRetryingCaller.java:147)
    at org.apache.hadoop.hbase.client.HBaseAdmin.executeCallable(HBaseAdmin.java:4083)
    at org.apache.hadoop.hbase.client.HBaseAdmin.getClusterStatus(HBaseAdmin.java:2765)
    at com.splicemachine.access.hbase.H10PartitionAdmin.allServers(H10PartitionAdmin.java:120)
    at com.splicemachine.hbase.HBaseRegionLoads.fetchRegionLoads(HBaseRegionLoads.java:164)
    at com.splicemachine.hbase.HBaseRegionLoads.access$000(HBaseRegionLoads.java:60)
    at com.splicemachine.hbase.HBaseRegionLoads$1.run(HBaseRegionLoads.java:81)
    at com.splicemachine.concurrent.LoggingScheduledThreadPoolExecutor$LoggingRunnable.run(LoggingScheduledThreadPoolExecutor.java:75)
    at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
    at java.util.concurrent.FutureTask.runAndReset(FutureTask.java:308)
    at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$301(ScheduledThreadPoolExecutor.java:180)
    at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:294)
    at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
    at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
    at java.lang.Thread.run(Thread.java:745)
Caused by: org.apache.hadoop.hbase.ipc.RemoteWithExtrasException(org.apache.hadoop.hbase.PleaseHoldException): org.apache.hadoop.hbase.PleaseHoldException: Master is initializing
    at org.apache.hadoop.hbase.master.HMaster.checkInitialized(HMaster.java:2324)
    at org.apache.hadoop.hbase.master.MasterRpcServices.getClusterStatus(MasterRpcServices.java:770)
    at org.apache.hadoop.hbase.protobuf.generated.MasterProtos$MasterService$2.callBlockingMethod(MasterProtos.java:55371)
    at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2114)
    at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:101)
    at org.apache.hadoop.hbase.ipc.RpcExecutor.consumerLoop(RpcExecutor.java:130)
    at org.apache.hadoop.hbase.ipc.RpcExecutor$1.run(RpcExecutor.java:107)
    at java.lang.Thread.run(Thread.java:745)

    at org.apache.hadoop.hbase.ipc.RpcClientImpl.call(RpcClientImpl.java:1226)
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient.callBlockingMethod(AbstractRpcClient.java:213)
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient$BlockingRpcChannelImplementation.callBlockingMethod(AbstractRpcClient.java:287)
    at org.apache.hadoop.hbase.protobuf.generated.MasterProtos$MasterService$BlockingStub.getClusterStatus(MasterProtos.java:58140)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation$4.getClusterStatus(ConnectionManager.java:2036)
    at org.apache.hadoop.hbase.client.HBaseAdmin$33.call(HBaseAdmin.java:2769)
    at org.apache.hadoop.hbase.client.HBaseAdmin$33.call(HBaseAdmin.java:2765)
    at org.apache.hadoop.hbase.client.RpcRetryingCaller.callWithRetries(RpcRetryingCaller.java:126)
    ... 14 more
2016-12-02 19:21:36,562 INFO  [SplitLogWorker-hscale-dev1-dn1:16020] coordination.ZkSplitLogWorkerCoordination: worker hscale-dev1-dn1,16020,1480686617338 acquired task /hbase-unsecure/splitWAL/WALs%2Fhscale-dev1-dn4%2C16020%2C1480686241892-splitting%2Fhscale-dev1-dn4%252C16020%252C1480686241892.null6.1480686244723
2016-12-02 19:21:36,595 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-1] wal.WALSplitter: Splitting wal: hdfs://hscale-dev1-nn:8020/apps/hbase/data/WALs/hscale-dev1-dn4,16020,1480686241892-splitting/hscale-dev1-dn4%2C16020%2C1480686241892.null6.1480686244723, length=91
2016-12-02 19:21:36,595 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-1] wal.WALSplitter: DistributedLogReplay = false
2016-12-02 19:21:36,648 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-1] util.FSHDFSUtils: Recovering lease on dfs file hdfs://hscale-dev1-nn:8020/apps/hbase/data/WALs/hscale-dev1-dn4,16020,1480686241892-splitting/hscale-dev1-dn4%2C16020%2C1480686241892.null6.1480686244723
2016-12-02 19:21:36,650 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-1] util.FSHDFSUtils: recoverLease=true, attempt=0 on file=hdfs://hscale-dev1-nn:8020/apps/hbase/data/WALs/hscale-dev1-dn4,16020,1480686241892-splitting/hscale-dev1-dn4%2C16020%2C1480686241892.null6.1480686244723 after 2ms
2016-12-02 19:21:36,669 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-1] wal.WALSplitter: Split writers finished
2016-12-02 19:21:36,670 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-1] wal.WALSplitter: Processed 0 edits across 0 regions; edits skipped=0; log file=hdfs://hscale-dev1-nn:8020/apps/hbase/data/WALs/hscale-dev1-dn4,16020,1480686241892-splitting/hscale-dev1-dn4%2C16020%2C1480686241892.null6.1480686244723, length=91, corrupted=false, progress failed=false
2016-12-02 19:21:36,686 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-1] coordination.ZkSplitLogWorkerCoordination: successfully transitioned task /hbase-unsecure/splitWAL/WALs%2Fhscale-dev1-dn4%2C16020%2C1480686241892-splitting%2Fhscale-dev1-dn4%252C16020%252C1480686241892.null6.1480686244723 to final state DONE hscale-dev1-dn1,16020,1480686617338
2016-12-02 19:21:36,686 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-1] handler.WALSplitterHandler: worker hscale-dev1-dn1,16020,1480686617338 done with task org.apache.hadoop.hbase.coordination.ZkSplitLogWorkerCoordination$ZkSplitTaskDetails@78fd51d4 in 124ms
2016-12-02 19:21:37,504 INFO  [SplitLogWorker-hscale-dev1-dn1:16020] coordination.ZkSplitLogWorkerCoordination: worker hscale-dev1-dn1,16020,1480686617338 acquired task /hbase-unsecure/splitWAL/WALs%2Fhscale-dev1-dn2%2C16020%2C1480686240889-splitting%2Fhscale-dev1-dn2%252C16020%252C1480686240889.null6.1480686243768
2016-12-02 19:21:37,533 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-0] wal.WALSplitter: Splitting wal: hdfs://hscale-dev1-nn:8020/apps/hbase/data/WALs/hscale-dev1-dn2,16020,1480686240889-splitting/hscale-dev1-dn2%2C16020%2C1480686240889.null6.1480686243768, length=91
2016-12-02 19:21:37,533 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-0] wal.WALSplitter: DistributedLogReplay = false
2016-12-02 19:21:37,563 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-0] util.FSHDFSUtils: Recovering lease on dfs file hdfs://hscale-dev1-nn:8020/apps/hbase/data/WALs/hscale-dev1-dn2,16020,1480686240889-splitting/hscale-dev1-dn2%2C16020%2C1480686240889.null6.1480686243768
2016-12-02 19:21:37,564 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-0] util.FSHDFSUtils: recoverLease=true, attempt=0 on file=hdfs://hscale-dev1-nn:8020/apps/hbase/data/WALs/hscale-dev1-dn2,16020,1480686240889-splitting/hscale-dev1-dn2%2C16020%2C1480686240889.null6.1480686243768 after 1ms
2016-12-02 19:21:37,590 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-0] wal.WALSplitter: Split writers finished
2016-12-02 19:21:37,590 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-0] wal.WALSplitter: Processed 0 edits across 0 regions; edits skipped=0; log file=hdfs://hscale-dev1-nn:8020/apps/hbase/data/WALs/hscale-dev1-dn2,16020,1480686240889-splitting/hscale-dev1-dn2%2C16020%2C1480686240889.null6.1480686243768, length=91, corrupted=false, progress failed=false
2016-12-02 19:21:37,606 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-0] coordination.ZkSplitLogWorkerCoordination: successfully transitioned task /hbase-unsecure/splitWAL/WALs%2Fhscale-dev1-dn2%2C16020%2C1480686240889-splitting%2Fhscale-dev1-dn2%252C16020%252C1480686240889.null6.1480686243768 to final state DONE hscale-dev1-dn1,16020,1480686617338
2016-12-02 19:21:37,606 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-0] handler.WALSplitterHandler: worker hscale-dev1-dn1,16020,1480686617338 done with task org.apache.hadoop.hbase.coordination.ZkSplitLogWorkerCoordination$ZkSplitTaskDetails@2e56fda0 in 102ms
2016-12-02 19:21:38,059 INFO  [SplitLogWorker-hscale-dev1-dn1:16020] coordination.ZkSplitLogWorkerCoordination: worker hscale-dev1-dn1,16020,1480686617338 acquired task /hbase-unsecure/splitWAL/WALs%2Fhscale-dev1-dn4%2C16020%2C1480686241892-splitting%2Fhscale-dev1-dn4%252C16020%252C1480686241892.null7.1480686244777
2016-12-02 19:21:38,080 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-1] wal.WALSplitter: Splitting wal: hdfs://hscale-dev1-nn:8020/apps/hbase/data/WALs/hscale-dev1-dn4,16020,1480686241892-splitting/hscale-dev1-dn4%2C16020%2C1480686241892.null7.1480686244777, length=91
2016-12-02 19:21:38,080 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-1] wal.WALSplitter: DistributedLogReplay = false
2016-12-02 19:21:38,098 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-1] util.FSHDFSUtils: Recovering lease on dfs file hdfs://hscale-dev1-nn:8020/apps/hbase/data/WALs/hscale-dev1-dn4,16020,1480686241892-splitting/hscale-dev1-dn4%2C16020%2C1480686241892.null7.1480686244777
2016-12-02 19:21:38,099 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-1] util.FSHDFSUtils: recoverLease=true, attempt=0 on file=hdfs://hscale-dev1-nn:8020/apps/hbase/data/WALs/hscale-dev1-dn4,16020,1480686241892-splitting/hscale-dev1-dn4%2C16020%2C1480686241892.null7.1480686244777 after 1ms
2016-12-02 19:21:38,126 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-1] wal.WALSplitter: Split writers finished
2016-12-02 19:21:38,126 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-1] wal.WALSplitter: Processed 0 edits across 0 regions; edits skipped=0; log file=hdfs://hscale-dev1-nn:8020/apps/hbase/data/WALs/hscale-dev1-dn4,16020,1480686241892-splitting/hscale-dev1-dn4%2C16020%2C1480686241892.null7.1480686244777, length=91, corrupted=false, progress failed=false
2016-12-02 19:21:38,143 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-1] coordination.ZkSplitLogWorkerCoordination: successfully transitioned task /hbase-unsecure/splitWAL/WALs%2Fhscale-dev1-dn4%2C16020%2C1480686241892-splitting%2Fhscale-dev1-dn4%252C16020%252C1480686241892.null7.1480686244777 to final state DONE hscale-dev1-dn1,16020,1480686617338
2016-12-02 19:21:38,143 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-1] handler.WALSplitterHandler: worker hscale-dev1-dn1,16020,1480686617338 done with task org.apache.hadoop.hbase.coordination.ZkSplitLogWorkerCoordination$ZkSplitTaskDetails@6514f7f2 in 84ms
2016-12-02 19:21:38,210 WARN  [pool-10-thread-1] client.HBaseAdmin: failed to get the procedure result procId=6
org.apache.hadoop.hbase.client.RetriesExhaustedException: Failed after attempts=5, exceptions:
Fri Dec 02 19:21:36 IST 2016, RpcRetryingCaller{globalStartTime=1480686696166, pause=100, retries=5}, org.apache.hadoop.hbase.ipc.RemoteWithExtrasException(org.apache.hadoop.hbase.PleaseHoldException): org.apache.hadoop.hbase.PleaseHoldException: Master is initializing
    at org.apache.hadoop.hbase.master.HMaster.checkInitialized(HMaster.java:2324)
    at org.apache.hadoop.hbase.master.MasterRpcServices.getProcedureResult(MasterRpcServices.java:1023)
    at org.apache.hadoop.hbase.protobuf.generated.MasterProtos$MasterService$2.callBlockingMethod(MasterProtos.java:55469)
    at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2114)
    at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:101)
    at org.apache.hadoop.hbase.ipc.RpcExecutor.consumerLoop(RpcExecutor.java:130)
    at org.apache.hadoop.hbase.ipc.RpcExecutor$1.run(RpcExecutor.java:107)
    at java.lang.Thread.run(Thread.java:745)

Fri Dec 02 19:21:36 IST 2016, RpcRetryingCaller{globalStartTime=1480686696166, pause=100, retries=5}, org.apache.hadoop.hbase.ipc.RemoteWithExtrasException(org.apache.hadoop.hbase.PleaseHoldException): org.apache.hadoop.hbase.PleaseHoldException: Master is initializing
    at org.apache.hadoop.hbase.master.HMaster.checkInitialized(HMaster.java:2324)
    at org.apache.hadoop.hbase.master.MasterRpcServices.getProcedureResult(MasterRpcServices.java:1023)
    at org.apache.hadoop.hbase.protobuf.generated.MasterProtos$MasterService$2.callBlockingMethod(MasterProtos.java:55469)
    at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2114)
    at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:101)
    at org.apache.hadoop.hbase.ipc.RpcExecutor.consumerLoop(RpcExecutor.java:130)
    at org.apache.hadoop.hbase.ipc.RpcExecutor$1.run(RpcExecutor.java:107)
    at java.lang.Thread.run(Thread.java:745)

Fri Dec 02 19:21:36 IST 2016, RpcRetryingCaller{globalStartTime=1480686696166, pause=100, retries=5}, org.apache.hadoop.hbase.ipc.RemoteWithExtrasException(org.apache.hadoop.hbase.PleaseHoldException): org.apache.hadoop.hbase.PleaseHoldException: Master is initializing
    at org.apache.hadoop.hbase.master.HMaster.checkInitialized(HMaster.java:2324)
    at org.apache.hadoop.hbase.master.MasterRpcServices.getProcedureResult(MasterRpcServices.java:1023)
    at org.apache.hadoop.hbase.protobuf.generated.MasterProtos$MasterService$2.callBlockingMethod(MasterProtos.java:55469)
    at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2114)
    at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:101)
    at org.apache.hadoop.hbase.ipc.RpcExecutor.consumerLoop(RpcExecutor.java:130)
    at org.apache.hadoop.hbase.ipc.RpcExecutor$1.run(RpcExecutor.java:107)
    at java.lang.Thread.run(Thread.java:745)

Fri Dec 02 19:21:37 IST 2016, RpcRetryingCaller{globalStartTime=1480686696166, pause=100, retries=5}, org.apache.hadoop.hbase.ipc.RemoteWithExtrasException(org.apache.hadoop.hbase.PleaseHoldException): org.apache.hadoop.hbase.PleaseHoldException: Master is initializing
    at org.apache.hadoop.hbase.master.HMaster.checkInitialized(HMaster.java:2324)
    at org.apache.hadoop.hbase.master.MasterRpcServices.getProcedureResult(MasterRpcServices.java:1023)
    at org.apache.hadoop.hbase.protobuf.generated.MasterProtos$MasterService$2.callBlockingMethod(MasterProtos.java:55469)
    at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2114)
    at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:101)
    at org.apache.hadoop.hbase.ipc.RpcExecutor.consumerLoop(RpcExecutor.java:130)
    at org.apache.hadoop.hbase.ipc.RpcExecutor$1.run(RpcExecutor.java:107)
    at java.lang.Thread.run(Thread.java:745)

Fri Dec 02 19:21:38 IST 2016, RpcRetryingCaller{globalStartTime=1480686696166, pause=100, retries=5}, org.apache.hadoop.hbase.ipc.RemoteWithExtrasException(org.apache.hadoop.hbase.PleaseHoldException): org.apache.hadoop.hbase.PleaseHoldException: Master is initializing
    at org.apache.hadoop.hbase.master.HMaster.checkInitialized(HMaster.java:2324)
    at org.apache.hadoop.hbase.master.MasterRpcServices.getProcedureResult(MasterRpcServices.java:1023)
    at org.apache.hadoop.hbase.protobuf.generated.MasterProtos$MasterService$2.callBlockingMethod(MasterProtos.java:55469)
    at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2114)
    at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:101)
    at org.apache.hadoop.hbase.ipc.RpcExecutor.consumerLoop(RpcExecutor.java:130)
    at org.apache.hadoop.hbase.ipc.RpcExecutor$1.run(RpcExecutor.java:107)
    at java.lang.Thread.run(Thread.java:745)


    at org.apache.hadoop.hbase.client.RpcRetryingCaller.callWithRetries(RpcRetryingCaller.java:147)
    at org.apache.hadoop.hbase.client.HBaseAdmin.executeCallable(HBaseAdmin.java:4083)
    at org.apache.hadoop.hbase.client.HBaseAdmin.access$700(HBaseAdmin.java:194)
    at org.apache.hadoop.hbase.client.HBaseAdmin$ProcedureFuture.getProcedureResult(HBaseAdmin.java:4383)
    at org.apache.hadoop.hbase.client.HBaseAdmin$ProcedureFuture.waitProcedureResult(HBaseAdmin.java:4335)
    at org.apache.hadoop.hbase.client.HBaseAdmin$ProcedureFuture.get(HBaseAdmin.java:4291)
    at org.apache.hadoop.hbase.client.HBaseAdmin.createTable(HBaseAdmin.java:647)
    at org.apache.hadoop.hbase.client.HBaseAdmin.createTable(HBaseAdmin.java:577)
    at com.splicemachine.lifecycle.RegionServerLifecycle.distributedStart(RegionServerLifecycle.java:66)
    at com.splicemachine.derby.lifecycle.EngineLifecycleService.start(EngineLifecycleService.java:81)
    at com.splicemachine.lifecycle.DatabaseLifecycleManager$Startup.bootServices(DatabaseLifecycleManager.java:229)
    at com.splicemachine.lifecycle.DatabaseLifecycleManager$Startup.run(DatabaseLifecycleManager.java:220)
    at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
    at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
    at java.lang.Thread.run(Thread.java:745)
Caused by: org.apache.hadoop.hbase.ipc.RemoteWithExtrasException(org.apache.hadoop.hbase.PleaseHoldException): org.apache.hadoop.hbase.PleaseHoldException: Master is initializing
    at org.apache.hadoop.hbase.master.HMaster.checkInitialized(HMaster.java:2324)
    at org.apache.hadoop.hbase.master.MasterRpcServices.getProcedureResult(MasterRpcServices.java:1023)
    at org.apache.hadoop.hbase.protobuf.generated.MasterProtos$MasterService$2.callBlockingMethod(MasterProtos.java:55469)
    at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2114)
    at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:101)
    at org.apache.hadoop.hbase.ipc.RpcExecutor.consumerLoop(RpcExecutor.java:130)
    at org.apache.hadoop.hbase.ipc.RpcExecutor$1.run(RpcExecutor.java:107)
    at java.lang.Thread.run(Thread.java:745)

    at org.apache.hadoop.hbase.ipc.RpcClientImpl.call(RpcClientImpl.java:1226)
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient.callBlockingMethod(AbstractRpcClient.java:213)
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient$BlockingRpcChannelImplementation.callBlockingMethod(AbstractRpcClient.java:287)
    at org.apache.hadoop.hbase.protobuf.generated.MasterProtos$MasterService$BlockingStub.getProcedureResult(MasterProtos.java:58728)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation$4.getProcedureResult(ConnectionManager.java:1951)
    at org.apache.hadoop.hbase.client.HBaseAdmin$ProcedureFuture$2.call(HBaseAdmin.java:4387)
    at org.apache.hadoop.hbase.client.HBaseAdmin$ProcedureFuture$2.call(HBaseAdmin.java:4384)
    at org.apache.hadoop.hbase.client.RpcRetryingCaller.callWithRetries(RpcRetryingCaller.java:126)
    ... 14 more
2016-12-02 19:21:38,596 INFO  [SplitLogWorker-hscale-dev1-dn1:16020] coordination.ZkSplitLogWorkerCoordination: worker hscale-dev1-dn1,16020,1480686617338 acquired task /hbase-unsecure/splitWAL/WALs%2Fhscale-dev1-dn2%2C16020%2C1480686240889-splitting%2Fhscale-dev1-dn2%252C16020%252C1480686240889.null9.1480686244030
2016-12-02 19:21:38,624 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-0] wal.WALSplitter: Splitting wal: hdfs://hscale-dev1-nn:8020/apps/hbase/data/WALs/hscale-dev1-dn2,16020,1480686240889-splitting/hscale-dev1-dn2%2C16020%2C1480686240889.null9.1480686244030, length=91
2016-12-02 19:21:38,624 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-0] wal.WALSplitter: DistributedLogReplay = false
2016-12-02 19:21:38,643 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-0] util.FSHDFSUtils: Recovering lease on dfs file hdfs://hscale-dev1-nn:8020/apps/hbase/data/WALs/hscale-dev1-dn2,16020,1480686240889-splitting/hscale-dev1-dn2%2C16020%2C1480686240889.null9.1480686244030
2016-12-02 19:21:38,644 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-0] util.FSHDFSUtils: recoverLease=true, attempt=0 on file=hdfs://hscale-dev1-nn:8020/apps/hbase/data/WALs/hscale-dev1-dn2,16020,1480686240889-splitting/hscale-dev1-dn2%2C16020%2C1480686240889.null9.1480686244030 after 1ms
2016-12-02 19:21:38,665 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-0] wal.WALSplitter: Split writers finished
2016-12-02 19:21:38,665 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-0] wal.WALSplitter: Processed 0 edits across 0 regions; edits skipped=0; log file=hdfs://hscale-dev1-nn:8020/apps/hbase/data/WALs/hscale-dev1-dn2,16020,1480686240889-splitting/hscale-dev1-dn2%2C16020%2C1480686240889.null9.1480686244030, length=91, corrupted=false, progress failed=false
2016-12-02 19:21:38,680 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-0] coordination.ZkSplitLogWorkerCoordination: successfully transitioned task /hbase-unsecure/splitWAL/WALs%2Fhscale-dev1-dn2%2C16020%2C1480686240889-splitting%2Fhscale-dev1-dn2%252C16020%252C1480686240889.null9.1480686244030 to final state DONE hscale-dev1-dn1,16020,1480686617338
2016-12-02 19:21:38,680 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-0] handler.WALSplitterHandler: worker hscale-dev1-dn1,16020,1480686617338 done with task org.apache.hadoop.hbase.coordination.ZkSplitLogWorkerCoordination$ZkSplitTaskDetails@d21d97 in 83ms
2016-12-02 19:21:39,613 INFO  [SplitLogWorker-hscale-dev1-dn1:16020] coordination.ZkSplitLogWorkerCoordination: worker hscale-dev1-dn1,16020,1480686617338 acquired task /hbase-unsecure/splitWAL/WALs%2Fhscale-dev1-dn2%2C16020%2C1480686240889-splitting%2Fhscale-dev1-dn2%252C16020%252C1480686240889.null3.1480686243573
2016-12-02 19:21:39,643 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-1] wal.WALSplitter: Splitting wal: hdfs://hscale-dev1-nn:8020/apps/hbase/data/WALs/hscale-dev1-dn2,16020,1480686240889-splitting/hscale-dev1-dn2%2C16020%2C1480686240889.null3.1480686243573, length=380
2016-12-02 19:21:39,643 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-1] wal.WALSplitter: DistributedLogReplay = false
2016-12-02 19:21:39,659 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-1] util.FSHDFSUtils: Recovering lease on dfs file hdfs://hscale-dev1-nn:8020/apps/hbase/data/WALs/hscale-dev1-dn2,16020,1480686240889-splitting/hscale-dev1-dn2%2C16020%2C1480686240889.null3.1480686243573
2016-12-02 19:21:39,661 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-1] util.FSHDFSUtils: recoverLease=true, attempt=0 on file=hdfs://hscale-dev1-nn:8020/apps/hbase/data/WALs/hscale-dev1-dn2,16020,1480686240889-splitting/hscale-dev1-dn2%2C16020%2C1480686240889.null3.1480686243573 after 2ms
2016-12-02 19:21:39,719 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-1-Writer-1] wal.WALSplitter: Creating writer path=hdfs://hscale-dev1-nn:8020/apps/hbase/data/data/splice/SPLICE_TXN/939509c5192425f14707b8d185bf68de/recovered.edits/0000000000000000003.temp region=939509c5192425f14707b8d185bf68de
2016-12-02 19:21:39,719 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-1] wal.WALSplitter: Split writers finished
2016-12-02 19:21:39,797 INFO  [split-log-closeStream-1] wal.WALSplitter: Rename hdfs://hscale-dev1-nn:8020/apps/hbase/data/data/splice/SPLICE_TXN/939509c5192425f14707b8d185bf68de/recovered.edits/0000000000000000003.temp to hdfs://hscale-dev1-nn:8020/apps/hbase/data/data/splice/SPLICE_TXN/939509c5192425f14707b8d185bf68de/recovered.edits/0000000000000000003
2016-12-02 19:21:39,797 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-1] wal.WALSplitter: Processed 1 edits across 1 regions; edits skipped=0; log file=hdfs://hscale-dev1-nn:8020/apps/hbase/data/WALs/hscale-dev1-dn2,16020,1480686240889-splitting/hscale-dev1-dn2%2C16020%2C1480686240889.null3.1480686243573, length=380, corrupted=false, progress failed=false
2016-12-02 19:21:39,813 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-1] coordination.ZkSplitLogWorkerCoordination: successfully transitioned task /hbase-unsecure/splitWAL/WALs%2Fhscale-dev1-dn2%2C16020%2C1480686240889-splitting%2Fhscale-dev1-dn2%252C16020%252C1480686240889.null3.1480686243573 to final state DONE hscale-dev1-dn1,16020,1480686617338
2016-12-02 19:21:39,813 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-1] handler.WALSplitterHandler: worker hscale-dev1-dn1,16020,1480686617338 done with task org.apache.hadoop.hbase.coordination.ZkSplitLogWorkerCoordination$ZkSplitTaskDetails@4011c4d9 in 200ms
2016-12-02 19:21:40,452 INFO  [SplitLogWorker-hscale-dev1-dn1:16020] coordination.ZkSplitLogWorkerCoordination: worker hscale-dev1-dn1,16020,1480686617338 acquired task /hbase-unsecure/splitWAL/WALs%2Fhscale-dev1-dn1%2C16020%2C1480686239118-splitting%2Fhscale-dev1-dn1%252C16020%252C1480686239118.null2.1480686241826
2016-12-02 19:21:40,484 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-0] wal.WALSplitter: Splitting wal: hdfs://hscale-dev1-nn:8020/apps/hbase/data/WALs/hscale-dev1-dn1,16020,1480686239118-splitting/hscale-dev1-dn1%2C16020%2C1480686239118.null2.1480686241826, length=380
2016-12-02 19:21:40,484 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-0] wal.WALSplitter: DistributedLogReplay = false
2016-12-02 19:21:40,501 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-0] util.FSHDFSUtils: Recovering lease on dfs file hdfs://hscale-dev1-nn:8020/apps/hbase/data/WALs/hscale-dev1-dn1,16020,1480686239118-splitting/hscale-dev1-dn1%2C16020%2C1480686239118.null2.1480686241826
2016-12-02 19:21:40,502 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-0] util.FSHDFSUtils: recoverLease=true, attempt=0 on file=hdfs://hscale-dev1-nn:8020/apps/hbase/data/WALs/hscale-dev1-dn1,16020,1480686239118-splitting/hscale-dev1-dn1%2C16020%2C1480686239118.null2.1480686241826 after 1ms
2016-12-02 19:21:40,559 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-0-Writer-2] wal.WALSplitter: Creating writer path=hdfs://hscale-dev1-nn:8020/apps/hbase/data/data/splice/SPLICE_TXN/ace1654b29a8315feb0793d463b78246/recovered.edits/0000000000000000003.temp region=ace1654b29a8315feb0793d463b78246
2016-12-02 19:21:40,559 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-0] wal.WALSplitter: Split writers finished
2016-12-02 19:21:40,622 INFO  [split-log-closeStream-1] wal.WALSplitter: Rename hdfs://hscale-dev1-nn:8020/apps/hbase/data/data/splice/SPLICE_TXN/ace1654b29a8315feb0793d463b78246/recovered.edits/0000000000000000003.temp to hdfs://hscale-dev1-nn:8020/apps/hbase/data/data/splice/SPLICE_TXN/ace1654b29a8315feb0793d463b78246/recovered.edits/0000000000000000003
2016-12-02 19:21:40,622 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-0] wal.WALSplitter: Processed 1 edits across 1 regions; edits skipped=0; log file=hdfs://hscale-dev1-nn:8020/apps/hbase/data/WALs/hscale-dev1-dn1,16020,1480686239118-splitting/hscale-dev1-dn1%2C16020%2C1480686239118.null2.1480686241826, length=380, corrupted=false, progress failed=false
2016-12-02 19:21:40,639 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-0] coordination.ZkSplitLogWorkerCoordination: successfully transitioned task /hbase-unsecure/splitWAL/WALs%2Fhscale-dev1-dn1%2C16020%2C1480686239118-splitting%2Fhscale-dev1-dn1%252C16020%252C1480686239118.null2.1480686241826 to final state DONE hscale-dev1-dn1,16020,1480686617338
2016-12-02 19:21:40,639 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-0] handler.WALSplitterHandler: worker hscale-dev1-dn1,16020,1480686617338 done with task org.apache.hadoop.hbase.coordination.ZkSplitLogWorkerCoordination$ZkSplitTaskDetails@1582fc8b in 187ms
2016-12-02 19:21:40,772 INFO  [PriorityRpcServer.handler=9,queue=1,port=16020] regionserver.RSRpcServices: Open splice:TENTATIVE_DDL,,1480686246654.7ba2eae2b5a4d7d5a163e2c50aa33a6b.
2016-12-02 19:21:40,779 INFO  [PriorityRpcServer.handler=9,queue=1,port=16020] regionserver.RSRpcServices: Open splice:SPLICE_TXN,,1480686242157.befbb055ce59b866459081d27215d354.
2016-12-02 19:21:40,797 INFO  [PriorityRpcServer.handler=9,queue=1,port=16020] regionserver.RSRpcServices: Open splice:SPLICE_TXN,\x09,1480686242157.ace1654b29a8315feb0793d463b78246.
2016-12-02 19:21:40,800 INFO  [PriorityRpcServer.handler=9,queue=1,port=16020] regionserver.RSRpcServices: Open splice:SPLICE_TXN,\x0F,1480686242157.c161d85dba1212e9f283596f80de56c8.
2016-12-02 19:21:40,816 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-0] coprocessor.CoprocessorHost: System coprocessor org.apache.hadoop.hbase.security.access.SecureBulkLoadEndpoint was loaded successfully with priority (536870911).
2016-12-02 19:21:40,817 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-0] coprocessor.CoprocessorHost: System coprocessor com.splicemachine.hbase.MemstoreAwareObserver was loaded successfully with priority (536870912).
2016-12-02 19:21:40,817 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-0] coprocessor.CoprocessorHost: System coprocessor com.splicemachine.derby.hbase.SpliceIndexObserver was loaded successfully with priority (536870913).
2016-12-02 19:21:40,817 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-0] coprocessor.CoprocessorHost: System coprocessor com.splicemachine.derby.hbase.SpliceIndexEndpoint was loaded successfully with priority (536870914).
2016-12-02 19:21:40,817 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-0] coprocessor.CoprocessorHost: System coprocessor com.splicemachine.hbase.RegionSizeEndpoint was loaded successfully with priority (536870915).
2016-12-02 19:21:40,818 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-0] coprocessor.CoprocessorHost: System coprocessor com.splicemachine.si.data.hbase.coprocessor.TxnLifecycleEndpoint was loaded successfully with priority (536870916).
2016-12-02 19:21:40,831 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-0] coprocessor.CoprocessorHost: System coprocessor com.splicemachine.si.data.hbase.coprocessor.SIObserver was loaded successfully with priority (536870917).
2016-12-02 19:21:40,832 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-0] coprocessor.CoprocessorHost: System coprocessor com.splicemachine.hbase.BackupEndpointObserver was loaded successfully with priority (536870918).
2016-12-02 19:21:40,837 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-0] compress.CodecPool: Got brand-new compressor [.snappy]
2016-12-02 19:21:40,841 INFO  [StoreOpener-7ba2eae2b5a4d7d5a163e2c50aa33a6b-1] hfile.CacheConfig: blockCache=LruBlockCache{blockCount=7, currentSize=852992, freeSize=804453376, maxSize=805306368, heapSize=852992, minSize=765041024, minFactor=0.95, multiSize=382520512, multiFactor=0.5, singleSize=191260256, singleFactor=0.25}, cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=true, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false
2016-12-02 19:21:40,841 INFO  [StoreOpener-7ba2eae2b5a4d7d5a163e2c50aa33a6b-1] compactions.CompactionConfiguration: size [16777216, 260046848); files [5, 10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000
2016-12-02 19:21:40,844 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-1] coprocessor.CoprocessorHost: System coprocessor org.apache.hadoop.hbase.security.access.SecureBulkLoadEndpoint was loaded successfully with priority (536870911).
2016-12-02 19:21:40,845 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-1] coprocessor.CoprocessorHost: System coprocessor com.splicemachine.hbase.MemstoreAwareObserver was loaded successfully with priority (536870912).
2016-12-02 19:21:40,845 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-1] coprocessor.CoprocessorHost: System coprocessor com.splicemachine.derby.hbase.SpliceIndexObserver was loaded successfully with priority (536870913).
2016-12-02 19:21:40,845 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-1] coprocessor.CoprocessorHost: System coprocessor com.splicemachine.derby.hbase.SpliceIndexEndpoint was loaded successfully with priority (536870914).
2016-12-02 19:21:40,846 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-1] coprocessor.CoprocessorHost: System coprocessor com.splicemachine.hbase.RegionSizeEndpoint was loaded successfully with priority (536870915).
2016-12-02 19:21:40,852 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-0] regionserver.HRegion: Replaying edits from hdfs://hscale-dev1-nn:8020/apps/hbase/data/data/splice/TENTATIVE_DDL/7ba2eae2b5a4d7d5a163e2c50aa33a6b/recovered.edits/0000000000000000003
2016-12-02 19:21:40,856 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-2] coprocessor.CoprocessorHost: System coprocessor org.apache.hadoop.hbase.security.access.SecureBulkLoadEndpoint was loaded successfully with priority (536870911).
2016-12-02 19:21:40,856 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-2] coprocessor.CoprocessorHost: System coprocessor com.splicemachine.hbase.MemstoreAwareObserver was loaded successfully with priority (536870912).
2016-12-02 19:21:40,856 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-2] coprocessor.CoprocessorHost: System coprocessor com.splicemachine.derby.hbase.SpliceIndexObserver was loaded successfully with priority (536870913).
2016-12-02 19:21:40,856 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-2] coprocessor.CoprocessorHost: System coprocessor com.splicemachine.derby.hbase.SpliceIndexEndpoint was loaded successfully with priority (536870914).
2016-12-02 19:21:40,856 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-2] coprocessor.CoprocessorHost: System coprocessor com.splicemachine.hbase.RegionSizeEndpoint was loaded successfully with priority (536870915).
2016-12-02 19:21:40,864 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-2] coprocessor.CoprocessorHost: System coprocessor com.splicemachine.si.data.hbase.coprocessor.TxnLifecycleEndpoint was loaded successfully with priority (536870916).
2016-12-02 19:21:40,864 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-1] coprocessor.CoprocessorHost: System coprocessor com.splicemachine.si.data.hbase.coprocessor.TxnLifecycleEndpoint was loaded successfully with priority (536870916).
2016-12-02 19:21:40,864 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-2] coprocessor.CoprocessorHost: System coprocessor com.splicemachine.si.data.hbase.coprocessor.SIObserver was loaded successfully with priority (536870917).
2016-12-02 19:21:40,864 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-1] coprocessor.CoprocessorHost: System coprocessor com.splicemachine.si.data.hbase.coprocessor.SIObserver was loaded successfully with priority (536870917).
2016-12-02 19:21:40,864 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-2] coprocessor.CoprocessorHost: System coprocessor com.splicemachine.hbase.BackupEndpointObserver was loaded successfully with priority (536870918).
2016-12-02 19:21:40,864 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-1] coprocessor.CoprocessorHost: System coprocessor com.splicemachine.hbase.BackupEndpointObserver was loaded successfully with priority (536870918).
2016-12-02 19:21:40,875 INFO  [StoreOpener-ace1654b29a8315feb0793d463b78246-1] hfile.CacheConfig: blockCache=LruBlockCache{blockCount=7, currentSize=852992, freeSize=804453376, maxSize=805306368, heapSize=852992, minSize=765041024, minFactor=0.95, multiSize=382520512, multiFactor=0.5, singleSize=191260256, singleFactor=0.25}, cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=true, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false
2016-12-02 19:21:40,875 INFO  [StoreOpener-ace1654b29a8315feb0793d463b78246-1] compactions.CompactionConfiguration: size [16777216, 260046848); files [5, 10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000
2016-12-02 19:21:40,876 INFO  [StoreOpener-befbb055ce59b866459081d27215d354-1] hfile.CacheConfig: blockCache=LruBlockCache{blockCount=7, currentSize=852992, freeSize=804453376, maxSize=805306368, heapSize=852992, minSize=765041024, minFactor=0.95, multiSize=382520512, multiFactor=0.5, singleSize=191260256, singleFactor=0.25}, cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=true, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false
2016-12-02 19:21:40,877 INFO  [StoreOpener-befbb055ce59b866459081d27215d354-1] compactions.CompactionConfiguration: size [16777216, 260046848); files [5, 10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000
2016-12-02 19:21:40,877 INFO  [StoreOpener-ace1654b29a8315feb0793d463b78246-1] hfile.CacheConfig: blockCache=LruBlockCache{blockCount=7, currentSize=852992, freeSize=804453376, maxSize=805306368, heapSize=852992, minSize=765041024, minFactor=0.95, multiSize=382520512, multiFactor=0.5, singleSize=191260256, singleFactor=0.25}, cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=true, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false
2016-12-02 19:21:40,878 INFO  [StoreOpener-ace1654b29a8315feb0793d463b78246-1] compactions.CompactionConfiguration: size [16777216, 260046848); files [5, 10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000
2016-12-02 19:21:40,880 INFO  [StoreOpener-befbb055ce59b866459081d27215d354-1] hfile.CacheConfig: blockCache=LruBlockCache{blockCount=7, currentSize=852992, freeSize=804453376, maxSize=805306368, heapSize=852992, minSize=765041024, minFactor=0.95, multiSize=382520512, multiFactor=0.5, singleSize=191260256, singleFactor=0.25}, cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=true, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false
2016-12-02 19:21:40,881 INFO  [StoreOpener-befbb055ce59b866459081d27215d354-1] compactions.CompactionConfiguration: size [16777216, 260046848); files [5, 10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000
2016-12-02 19:21:40,885 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-2] regionserver.HRegion: Replaying edits from hdfs://hscale-dev1-nn:8020/apps/hbase/data/data/splice/SPLICE_TXN/ace1654b29a8315feb0793d463b78246/recovered.edits/0000000000000000003
2016-12-02 19:21:40,889 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-1] regionserver.HRegion: Replaying edits from hdfs://hscale-dev1-nn:8020/apps/hbase/data/data/splice/SPLICE_TXN/befbb055ce59b866459081d27215d354/recovered.edits/0000000000000000003
2016-12-02 19:21:40,946 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-0] regionserver.HRegion: Onlined 7ba2eae2b5a4d7d5a163e2c50aa33a6b; next sequenceid=4
2016-12-02 19:21:40,953 INFO  [PostOpenDeployTasks:7ba2eae2b5a4d7d5a163e2c50aa33a6b] regionserver.HRegionServer: Post open deploy tasks for splice:TENTATIVE_DDL,,1480686246654.7ba2eae2b5a4d7d5a163e2c50aa33a6b.
2016-12-02 19:21:40,983 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-2] regionserver.HRegion: Onlined ace1654b29a8315feb0793d463b78246; next sequenceid=4
2016-12-02 19:21:40,990 INFO  [PriorityRpcServer.handler=5,queue=1,port=16020] regionserver.RSRpcServices: Open splice:SPLICE_TXN,\x04,1480686242157.5147876c22787fb8040fbe21bfff47ad.
2016-12-02 19:21:40,999 INFO  [PostOpenDeployTasks:ace1654b29a8315feb0793d463b78246] regionserver.HRegionServer: Post open deploy tasks for splice:SPLICE_TXN,\x09,1480686242157.ace1654b29a8315feb0793d463b78246.
2016-12-02 19:21:40,999 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-1] regionserver.HRegion: Onlined befbb055ce59b866459081d27215d354; next sequenceid=4
2016-12-02 19:21:41,001 INFO  [PriorityRpcServer.handler=5,queue=1,port=16020] regionserver.RSRpcServices: Open splice:SPLICE_TXN,\x08,1480686242157.10320ae24a73ddfe6576272c451d3845.
2016-12-02 19:21:41,002 INFO  [PriorityRpcServer.handler=5,queue=1,port=16020] regionserver.RSRpcServices: Open splice:SPLICE_TXN,\x0D,1480686242157.5ea395aa4dcee819ecb70dff508ef3f6.
2016-12-02 19:21:41,004 INFO  [PostOpenDeployTasks:befbb055ce59b866459081d27215d354] regionserver.HRegionServer: Post open deploy tasks for splice:SPLICE_TXN,,1480686242157.befbb055ce59b866459081d27215d354.
2016-12-02 19:21:41,039 INFO  [PostOpenDeployTasks:7ba2eae2b5a4d7d5a163e2c50aa33a6b] hbase.MetaTableAccessor: Updated row splice:TENTATIVE_DDL,,1480686246654.7ba2eae2b5a4d7d5a163e2c50aa33a6b. with server=hscale-dev1-dn1,16020,1480686617338
2016-12-02 19:21:41,040 INFO  [PostOpenDeployTasks:ace1654b29a8315feb0793d463b78246] hbase.MetaTableAccessor: Updated row splice:SPLICE_TXN,\x09,1480686242157.ace1654b29a8315feb0793d463b78246. with server=hscale-dev1-dn1,16020,1480686617338
2016-12-02 19:21:41,040 INFO  [PostOpenDeployTasks:befbb055ce59b866459081d27215d354] hbase.MetaTableAccessor: Updated row splice:SPLICE_TXN,,1480686242157.befbb055ce59b866459081d27215d354. with server=hscale-dev1-dn1,16020,1480686617338
2016-12-02 19:21:41,045 ERROR [hbase-region-load-updater-0] hbase.HBaseRegionLoads: Unable to fetch region load info
org.apache.hadoop.hbase.client.RetriesExhaustedException: Failed after attempts=5, exceptions:
Fri Dec 02 19:21:39 IST 2016, RpcRetryingCaller{globalStartTime=1480686699019, pause=100, retries=5}, org.apache.hadoop.hbase.ipc.RemoteWithExtrasException(org.apache.hadoop.hbase.PleaseHoldException): org.apache.hadoop.hbase.PleaseHoldException: Master is initializing
    at org.apache.hadoop.hbase.master.HMaster.checkInitialized(HMaster.java:2324)
    at org.apache.hadoop.hbase.master.MasterRpcServices.getClusterStatus(MasterRpcServices.java:770)
    at org.apache.hadoop.hbase.protobuf.generated.MasterProtos$MasterService$2.callBlockingMethod(MasterProtos.java:55371)
    at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2114)
    at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:101)
    at org.apache.hadoop.hbase.ipc.RpcExecutor.consumerLoop(RpcExecutor.java:130)
    at org.apache.hadoop.hbase.ipc.RpcExecutor$1.run(RpcExecutor.java:107)
    at java.lang.Thread.run(Thread.java:745)

Fri Dec 02 19:21:39 IST 2016, RpcRetryingCaller{globalStartTime=1480686699019, pause=100, retries=5}, org.apache.hadoop.hbase.ipc.RemoteWithExtrasException(org.apache.hadoop.hbase.PleaseHoldException): org.apache.hadoop.hbase.PleaseHoldException: Master is initializing
    at org.apache.hadoop.hbase.master.HMaster.checkInitialized(HMaster.java:2324)
    at org.apache.hadoop.hbase.master.MasterRpcServices.getClusterStatus(MasterRpcServices.java:770)
    at org.apache.hadoop.hbase.protobuf.generated.MasterProtos$MasterService$2.callBlockingMethod(MasterProtos.java:55371)
    at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2114)
    at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:101)
    at org.apache.hadoop.hbase.ipc.RpcExecutor.consumerLoop(RpcExecutor.java:130)
    at org.apache.hadoop.hbase.ipc.RpcExecutor$1.run(RpcExecutor.java:107)
    at java.lang.Thread.run(Thread.java:745)

Fri Dec 02 19:21:39 IST 2016, RpcRetryingCaller{globalStartTime=1480686699019, pause=100, retries=5}, org.apache.hadoop.hbase.ipc.RemoteWithExtrasException(org.apache.hadoop.hbase.PleaseHoldException): org.apache.hadoop.hbase.PleaseHoldException: Master is initializing
    at org.apache.hadoop.hbase.master.HMaster.checkInitialized(HMaster.java:2324)
    at org.apache.hadoop.hbase.master.MasterRpcServices.getClusterStatus(MasterRpcServices.java:770)
    at org.apache.hadoop.hbase.protobuf.generated.MasterProtos$MasterService$2.callBlockingMethod(MasterProtos.java:55371)
    at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2114)
    at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:101)
    at org.apache.hadoop.hbase.ipc.RpcExecutor.consumerLoop(RpcExecutor.java:130)
    at org.apache.hadoop.hbase.ipc.RpcExecutor$1.run(RpcExecutor.java:107)
    at java.lang.Thread.run(Thread.java:745)

Fri Dec 02 19:21:40 IST 2016, RpcRetryingCaller{globalStartTime=1480686699019, pause=100, retries=5}, org.apache.hadoop.hbase.ipc.RemoteWithExtrasException(org.apache.hadoop.hbase.PleaseHoldException): org.apache.hadoop.hbase.PleaseHoldException: Master is initializing
    at org.apache.hadoop.hbase.master.HMaster.checkInitialized(HMaster.java:2324)
    at org.apache.hadoop.hbase.master.MasterRpcServices.getClusterStatus(MasterRpcServices.java:770)
    at org.apache.hadoop.hbase.protobuf.generated.MasterProtos$MasterService$2.callBlockingMethod(MasterProtos.java:55371)
    at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2114)
    at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:101)
    at org.apache.hadoop.hbase.ipc.RpcExecutor.consumerLoop(RpcExecutor.java:130)
    at org.apache.hadoop.hbase.ipc.RpcExecutor$1.run(RpcExecutor.java:107)
    at java.lang.Thread.run(Thread.java:745)

Fri Dec 02 19:21:41 IST 2016, RpcRetryingCaller{globalStartTime=1480686699019, pause=100, retries=5}, org.apache.hadoop.hbase.ipc.RemoteWithExtrasException(org.apache.hadoop.hbase.PleaseHoldException): org.apache.hadoop.hbase.PleaseHoldException: Master is initializing
    at org.apache.hadoop.hbase.master.HMaster.checkInitialized(HMaster.java:2324)
    at org.apache.hadoop.hbase.master.MasterRpcServices.getClusterStatus(MasterRpcServices.java:770)
    at org.apache.hadoop.hbase.protobuf.generated.MasterProtos$MasterService$2.callBlockingMethod(MasterProtos.java:55371)
    at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2114)
    at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:101)
    at org.apache.hadoop.hbase.ipc.RpcExecutor.consumerLoop(RpcExecutor.java:130)
    at org.apache.hadoop.hbase.ipc.RpcExecutor$1.run(RpcExecutor.java:107)
    at java.lang.Thread.run(Thread.java:745)


    at org.apache.hadoop.hbase.client.RpcRetryingCaller.callWithRetries(RpcRetryingCaller.java:147)
    at org.apache.hadoop.hbase.client.HBaseAdmin.executeCallable(HBaseAdmin.java:4083)
    at org.apache.hadoop.hbase.client.HBaseAdmin.getClusterStatus(HBaseAdmin.java:2765)
    at com.splicemachine.access.hbase.H10PartitionAdmin.allServers(H10PartitionAdmin.java:120)
    at com.splicemachine.hbase.HBaseRegionLoads.fetchRegionLoads(HBaseRegionLoads.java:164)
    at com.splicemachine.hbase.HBaseRegionLoads.access$000(HBaseRegionLoads.java:60)
    at com.splicemachine.hbase.HBaseRegionLoads$1.run(HBaseRegionLoads.java:81)
    at com.splicemachine.concurrent.LoggingScheduledThreadPoolExecutor$LoggingRunnable.run(LoggingScheduledThreadPoolExecutor.java:75)
    at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
    at java.util.concurrent.FutureTask.runAndReset(FutureTask.java:308)
    at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$301(ScheduledThreadPoolExecutor.java:180)
    at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:294)
    at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
    at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
    at java.lang.Thread.run(Thread.java:745)
Caused by: org.apache.hadoop.hbase.ipc.RemoteWithExtrasException(org.apache.hadoop.hbase.PleaseHoldException): org.apache.hadoop.hbase.PleaseHoldException: Master is initializing
    at org.apache.hadoop.hbase.master.HMaster.checkInitialized(HMaster.java:2324)
    at org.apache.hadoop.hbase.master.MasterRpcServices.getClusterStatus(MasterRpcServices.java:770)
    at org.apache.hadoop.hbase.protobuf.generated.MasterProtos$MasterService$2.callBlockingMethod(MasterProtos.java:55371)
    at org.apache.hadoop.hbase.ipc.RpcServer.call(RpcServer.java:2114)
    at org.apache.hadoop.hbase.ipc.CallRunner.run(CallRunner.java:101)
    at org.apache.hadoop.hbase.ipc.RpcExecutor.consumerLoop(RpcExecutor.java:130)
    at org.apache.hadoop.hbase.ipc.RpcExecutor$1.run(RpcExecutor.java:107)
    at java.lang.Thread.run(Thread.java:745)

    at org.apache.hadoop.hbase.ipc.RpcClientImpl.call(RpcClientImpl.java:1226)
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient.callBlockingMethod(AbstractRpcClient.java:213)
    at org.apache.hadoop.hbase.ipc.AbstractRpcClient$BlockingRpcChannelImplementation.callBlockingMethod(AbstractRpcClient.java:287)
    at org.apache.hadoop.hbase.protobuf.generated.MasterProtos$MasterService$BlockingStub.getClusterStatus(MasterProtos.java:58140)
    at org.apache.hadoop.hbase.client.ConnectionManager$HConnectionImplementation$4.getClusterStatus(ConnectionManager.java:2036)
    at org.apache.hadoop.hbase.client.HBaseAdmin$33.call(HBaseAdmin.java:2769)
    at org.apache.hadoop.hbase.client.HBaseAdmin$33.call(HBaseAdmin.java:2765)
    at org.apache.hadoop.hbase.client.RpcRetryingCaller.callWithRetries(RpcRetryingCaller.java:126)
    ... 14 more
2016-12-02 19:21:41,111 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-1] coprocessor.CoprocessorHost: System coprocessor org.apache.hadoop.hbase.security.access.SecureBulkLoadEndpoint was loaded successfully with priority (536870911).
2016-12-02 19:21:41,112 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-1] coprocessor.CoprocessorHost: System coprocessor com.splicemachine.hbase.MemstoreAwareObserver was loaded successfully with priority (536870912).
2016-12-02 19:21:41,112 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-1] coprocessor.CoprocessorHost: System coprocessor com.splicemachine.derby.hbase.SpliceIndexObserver was loaded successfully with priority (536870913).
2016-12-02 19:21:41,112 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-1] coprocessor.CoprocessorHost: System coprocessor com.splicemachine.derby.hbase.SpliceIndexEndpoint was loaded successfully with priority (536870914).
2016-12-02 19:21:41,113 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-1] coprocessor.CoprocessorHost: System coprocessor com.splicemachine.hbase.RegionSizeEndpoint was loaded successfully with priority (536870915).
2016-12-02 19:21:41,113 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-1] coprocessor.CoprocessorHost: System coprocessor com.splicemachine.si.data.hbase.coprocessor.TxnLifecycleEndpoint was loaded successfully with priority (536870916).
2016-12-02 19:21:41,114 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-1] coprocessor.CoprocessorHost: System coprocessor com.splicemachine.si.data.hbase.coprocessor.SIObserver was loaded successfully with priority (536870917).
2016-12-02 19:21:41,114 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-1] coprocessor.CoprocessorHost: System coprocessor com.splicemachine.hbase.BackupEndpointObserver was loaded successfully with priority (536870918).
2016-12-02 19:21:41,119 INFO  [StoreOpener-c161d85dba1212e9f283596f80de56c8-1] hfile.CacheConfig: blockCache=LruBlockCache{blockCount=7, currentSize=852992, freeSize=804453376, maxSize=805306368, heapSize=852992, minSize=765041024, minFactor=0.95, multiSize=382520512, multiFactor=0.5, singleSize=191260256, singleFactor=0.25}, cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=true, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false
2016-12-02 19:21:41,120 INFO  [StoreOpener-c161d85dba1212e9f283596f80de56c8-1] compactions.CompactionConfiguration: size [16777216, 260046848); files [5, 10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000
2016-12-02 19:21:41,123 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-0] coprocessor.CoprocessorHost: System coprocessor org.apache.hadoop.hbase.security.access.SecureBulkLoadEndpoint was loaded successfully with priority (536870911).
2016-12-02 19:21:41,123 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-0] coprocessor.CoprocessorHost: System coprocessor com.splicemachine.hbase.MemstoreAwareObserver was loaded successfully with priority (536870912).
2016-12-02 19:21:41,123 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-0] coprocessor.CoprocessorHost: System coprocessor com.splicemachine.derby.hbase.SpliceIndexObserver was loaded successfully with priority (536870913).
2016-12-02 19:21:41,123 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-0] coprocessor.CoprocessorHost: System coprocessor com.splicemachine.derby.hbase.SpliceIndexEndpoint was loaded successfully with priority (536870914).
2016-12-02 19:21:41,123 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-0] coprocessor.CoprocessorHost: System coprocessor com.splicemachine.hbase.RegionSizeEndpoint was loaded successfully with priority (536870915).
2016-12-02 19:21:41,124 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-0] coprocessor.CoprocessorHost: System coprocessor com.splicemachine.si.data.hbase.coprocessor.TxnLifecycleEndpoint was loaded successfully with priority (536870916).
2016-12-02 19:21:41,124 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-0] coprocessor.CoprocessorHost: System coprocessor com.splicemachine.si.data.hbase.coprocessor.SIObserver was loaded successfully with priority (536870917).
2016-12-02 19:21:41,124 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-0] coprocessor.CoprocessorHost: System coprocessor com.splicemachine.hbase.BackupEndpointObserver was loaded successfully with priority (536870918).
2016-12-02 19:21:41,125 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-2] coprocessor.CoprocessorHost: System coprocessor org.apache.hadoop.hbase.security.access.SecureBulkLoadEndpoint was loaded successfully with priority (536870911).
2016-12-02 19:21:41,125 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-2] coprocessor.CoprocessorHost: System coprocessor com.splicemachine.hbase.MemstoreAwareObserver was loaded successfully with priority (536870912).
2016-12-02 19:21:41,125 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-2] coprocessor.CoprocessorHost: System coprocessor com.splicemachine.derby.hbase.SpliceIndexObserver was loaded successfully with priority (536870913).
2016-12-02 19:21:41,125 INFO  [StoreOpener-c161d85dba1212e9f283596f80de56c8-1] hfile.CacheConfig: blockCache=LruBlockCache{blockCount=7, currentSize=852992, freeSize=804453376, maxSize=805306368, heapSize=852992, minSize=765041024, minFactor=0.95, multiSize=382520512, multiFactor=0.5, singleSize=191260256, singleFactor=0.25}, cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=true, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false
2016-12-02 19:21:41,125 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-2] coprocessor.CoprocessorHost: System coprocessor com.splicemachine.derby.hbase.SpliceIndexEndpoint was loaded successfully with priority (536870914).
2016-12-02 19:21:41,125 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-2] coprocessor.CoprocessorHost: System coprocessor com.splicemachine.hbase.RegionSizeEndpoint was loaded successfully with priority (536870915).
2016-12-02 19:21:41,125 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-2] coprocessor.CoprocessorHost: System coprocessor com.splicemachine.si.data.hbase.coprocessor.TxnLifecycleEndpoint was loaded successfully with priority (536870916).
2016-12-02 19:21:41,125 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-2] coprocessor.CoprocessorHost: System coprocessor com.splicemachine.si.data.hbase.coprocessor.SIObserver was loaded successfully with priority (536870917).
2016-12-02 19:21:41,125 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-2] coprocessor.CoprocessorHost: System coprocessor com.splicemachine.hbase.BackupEndpointObserver was loaded successfully with priority (536870918).
2016-12-02 19:21:41,126 INFO  [StoreOpener-c161d85dba1212e9f283596f80de56c8-1] compactions.CompactionConfiguration: size [16777216, 260046848); files [5, 10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000
2016-12-02 19:21:41,130 INFO  [StoreOpener-10320ae24a73ddfe6576272c451d3845-1] hfile.CacheConfig: blockCache=LruBlockCache{blockCount=7, currentSize=852992, freeSize=804453376, maxSize=805306368, heapSize=852992, minSize=765041024, minFactor=0.95, multiSize=382520512, multiFactor=0.5, singleSize=191260256, singleFactor=0.25}, cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=true, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false
2016-12-02 19:21:41,131 INFO  [StoreOpener-5147876c22787fb8040fbe21bfff47ad-1] hfile.CacheConfig: blockCache=LruBlockCache{blockCount=7, currentSize=852992, freeSize=804453376, maxSize=805306368, heapSize=852992, minSize=765041024, minFactor=0.95, multiSize=382520512, multiFactor=0.5, singleSize=191260256, singleFactor=0.25}, cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=true, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false
2016-12-02 19:21:41,131 INFO  [StoreOpener-10320ae24a73ddfe6576272c451d3845-1] compactions.CompactionConfiguration: size [16777216, 260046848); files [5, 10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000
2016-12-02 19:21:41,131 INFO  [StoreOpener-5147876c22787fb8040fbe21bfff47ad-1] compactions.CompactionConfiguration: size [16777216, 260046848); files [5, 10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000
2016-12-02 19:21:41,134 INFO  [StoreOpener-10320ae24a73ddfe6576272c451d3845-1] hfile.CacheConfig: blockCache=LruBlockCache{blockCount=7, currentSize=852992, freeSize=804453376, maxSize=805306368, heapSize=852992, minSize=765041024, minFactor=0.95, multiSize=382520512, multiFactor=0.5, singleSize=191260256, singleFactor=0.25}, cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=true, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false
2016-12-02 19:21:41,134 INFO  [StoreOpener-10320ae24a73ddfe6576272c451d3845-1] compactions.CompactionConfiguration: size [16777216, 260046848); files [5, 10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000
2016-12-02 19:21:41,134 INFO  [StoreOpener-5147876c22787fb8040fbe21bfff47ad-1] hfile.CacheConfig: blockCache=LruBlockCache{blockCount=7, currentSize=852992, freeSize=804453376, maxSize=805306368, heapSize=852992, minSize=765041024, minFactor=0.95, multiSize=382520512, multiFactor=0.5, singleSize=191260256, singleFactor=0.25}, cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=true, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false
2016-12-02 19:21:41,135 INFO  [StoreOpener-5147876c22787fb8040fbe21bfff47ad-1] compactions.CompactionConfiguration: size [16777216, 260046848); files [5, 10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000
2016-12-02 19:21:41,138 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-1] regionserver.HRegion: Replaying edits from hdfs://hscale-dev1-nn:8020/apps/hbase/data/data/splice/SPLICE_TXN/c161d85dba1212e9f283596f80de56c8/recovered.edits/0000000000000000003
2016-12-02 19:21:41,141 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-2] regionserver.HRegion: Replaying edits from hdfs://hscale-dev1-nn:8020/apps/hbase/data/data/splice/SPLICE_TXN/10320ae24a73ddfe6576272c451d3845/recovered.edits/0000000000000000003
2016-12-02 19:21:41,141 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-0] regionserver.HRegion: Replaying edits from hdfs://hscale-dev1-nn:8020/apps/hbase/data/data/splice/SPLICE_TXN/5147876c22787fb8040fbe21bfff47ad/recovered.edits/0000000000000000003
2016-12-02 19:21:41,241 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-1] regionserver.HRegion: Onlined c161d85dba1212e9f283596f80de56c8; next sequenceid=4
2016-12-02 19:21:41,249 INFO  [PostOpenDeployTasks:c161d85dba1212e9f283596f80de56c8] regionserver.HRegionServer: Post open deploy tasks for splice:SPLICE_TXN,\x0F,1480686242157.c161d85dba1212e9f283596f80de56c8.
2016-12-02 19:21:41,254 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-0] regionserver.HRegion: Onlined 5147876c22787fb8040fbe21bfff47ad; next sequenceid=4
2016-12-02 19:21:41,254 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-2] regionserver.HRegion: Onlined 10320ae24a73ddfe6576272c451d3845; next sequenceid=4
2016-12-02 19:21:41,259 INFO  [PostOpenDeployTasks:c161d85dba1212e9f283596f80de56c8] hbase.MetaTableAccessor: Updated row splice:SPLICE_TXN,\x0F,1480686242157.c161d85dba1212e9f283596f80de56c8. with server=hscale-dev1-dn1,16020,1480686617338
2016-12-02 19:21:41,262 INFO  [PostOpenDeployTasks:5147876c22787fb8040fbe21bfff47ad] regionserver.HRegionServer: Post open deploy tasks for splice:SPLICE_TXN,\x04,1480686242157.5147876c22787fb8040fbe21bfff47ad.
2016-12-02 19:21:41,262 INFO  [PostOpenDeployTasks:10320ae24a73ddfe6576272c451d3845] regionserver.HRegionServer: Post open deploy tasks for splice:SPLICE_TXN,\x08,1480686242157.10320ae24a73ddfe6576272c451d3845.
2016-12-02 19:21:41,272 INFO  [PostOpenDeployTasks:5147876c22787fb8040fbe21bfff47ad] hbase.MetaTableAccessor: Updated row splice:SPLICE_TXN,\x04,1480686242157.5147876c22787fb8040fbe21bfff47ad. with server=hscale-dev1-dn1,16020,1480686617338
2016-12-02 19:21:41,275 INFO  [PostOpenDeployTasks:10320ae24a73ddfe6576272c451d3845] hbase.MetaTableAccessor: Updated row splice:SPLICE_TXN,\x08,1480686242157.10320ae24a73ddfe6576272c451d3845. with server=hscale-dev1-dn1,16020,1480686617338
2016-12-02 19:21:41,335 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-1] coprocessor.CoprocessorHost: System coprocessor org.apache.hadoop.hbase.security.access.SecureBulkLoadEndpoint was loaded successfully with priority (536870911).
2016-12-02 19:21:41,335 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-1] coprocessor.CoprocessorHost: System coprocessor com.splicemachine.hbase.MemstoreAwareObserver was loaded successfully with priority (536870912).
2016-12-02 19:21:41,335 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-1] coprocessor.CoprocessorHost: System coprocessor com.splicemachine.derby.hbase.SpliceIndexObserver was loaded successfully with priority (536870913).
2016-12-02 19:21:41,335 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-1] coprocessor.CoprocessorHost: System coprocessor com.splicemachine.derby.hbase.SpliceIndexEndpoint was loaded successfully with priority (536870914).
2016-12-02 19:21:41,335 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-1] coprocessor.CoprocessorHost: System coprocessor com.splicemachine.hbase.RegionSizeEndpoint was loaded successfully with priority (536870915).
2016-12-02 19:21:41,335 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-1] coprocessor.CoprocessorHost: System coprocessor com.splicemachine.si.data.hbase.coprocessor.TxnLifecycleEndpoint was loaded successfully with priority (536870916).
2016-12-02 19:21:41,336 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-1] coprocessor.CoprocessorHost: System coprocessor com.splicemachine.si.data.hbase.coprocessor.SIObserver was loaded successfully with priority (536870917).
2016-12-02 19:21:41,336 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-1] coprocessor.CoprocessorHost: System coprocessor com.splicemachine.hbase.BackupEndpointObserver was loaded successfully with priority (536870918).
2016-12-02 19:21:41,340 INFO  [StoreOpener-5ea395aa4dcee819ecb70dff508ef3f6-1] hfile.CacheConfig: blockCache=LruBlockCache{blockCount=7, currentSize=852992, freeSize=804453376, maxSize=805306368, heapSize=852992, minSize=765041024, minFactor=0.95, multiSize=382520512, multiFactor=0.5, singleSize=191260256, singleFactor=0.25}, cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=true, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false
2016-12-02 19:21:41,341 INFO  [StoreOpener-5ea395aa4dcee819ecb70dff508ef3f6-1] compactions.CompactionConfiguration: size [16777216, 260046848); files [5, 10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000
2016-12-02 19:21:41,344 INFO  [StoreOpener-5ea395aa4dcee819ecb70dff508ef3f6-1] hfile.CacheConfig: blockCache=LruBlockCache{blockCount=7, currentSize=852992, freeSize=804453376, maxSize=805306368, heapSize=852992, minSize=765041024, minFactor=0.95, multiSize=382520512, multiFactor=0.5, singleSize=191260256, singleFactor=0.25}, cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=true, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false
2016-12-02 19:21:41,344 INFO  [StoreOpener-5ea395aa4dcee819ecb70dff508ef3f6-1] compactions.CompactionConfiguration: size [16777216, 260046848); files [5, 10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000
2016-12-02 19:21:41,352 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-1] regionserver.HRegion: Replaying edits from hdfs://hscale-dev1-nn:8020/apps/hbase/data/data/splice/SPLICE_TXN/5ea395aa4dcee819ecb70dff508ef3f6/recovered.edits/0000000000000000003
2016-12-02 19:21:41,371 INFO  [SplitLogWorker-hscale-dev1-dn1:16020] coordination.ZkSplitLogWorkerCoordination: worker hscale-dev1-dn1,16020,1480686617338 acquired task /hbase-unsecure/splitWAL/WALs%2Fhscale-dev1-dn4%2C16020%2C1480686241892-splitting%2Fhscale-dev1-dn4%252C16020%252C1480686241892.null10.1480686245082
2016-12-02 19:21:41,392 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-1] wal.WALSplitter: Splitting wal: hdfs://hscale-dev1-nn:8020/apps/hbase/data/WALs/hscale-dev1-dn4,16020,1480686241892-splitting/hscale-dev1-dn4%2C16020%2C1480686241892.null10.1480686245082, length=91
2016-12-02 19:21:41,392 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-1] wal.WALSplitter: DistributedLogReplay = false
2016-12-02 19:21:41,411 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-1] util.FSHDFSUtils: Recovering lease on dfs file hdfs://hscale-dev1-nn:8020/apps/hbase/data/WALs/hscale-dev1-dn4,16020,1480686241892-splitting/hscale-dev1-dn4%2C16020%2C1480686241892.null10.1480686245082
2016-12-02 19:21:41,412 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-1] util.FSHDFSUtils: recoverLease=true, attempt=0 on file=hdfs://hscale-dev1-nn:8020/apps/hbase/data/WALs/hscale-dev1-dn4,16020,1480686241892-splitting/hscale-dev1-dn4%2C16020%2C1480686241892.null10.1480686245082 after 1ms
2016-12-02 19:21:41,433 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-1] regionserver.HRegion: Onlined 5ea395aa4dcee819ecb70dff508ef3f6; next sequenceid=4
2016-12-02 19:21:41,434 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-1] wal.WALSplitter: Split writers finished
2016-12-02 19:21:41,435 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-1] wal.WALSplitter: Processed 0 edits across 0 regions; edits skipped=0; log file=hdfs://hscale-dev1-nn:8020/apps/hbase/data/WALs/hscale-dev1-dn4,16020,1480686241892-splitting/hscale-dev1-dn4%2C16020%2C1480686241892.null10.1480686245082, length=91, corrupted=false, progress failed=false
2016-12-02 19:21:41,470 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-1] coordination.ZkSplitLogWorkerCoordination: successfully transitioned task /hbase-unsecure/splitWAL/WALs%2Fhscale-dev1-dn4%2C16020%2C1480686241892-splitting%2Fhscale-dev1-dn4%252C16020%252C1480686241892.null10.1480686245082 to final state DONE hscale-dev1-dn1,16020,1480686617338
2016-12-02 19:21:41,470 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-1] handler.WALSplitterHandler: worker hscale-dev1-dn1,16020,1480686617338 done with task org.apache.hadoop.hbase.coordination.ZkSplitLogWorkerCoordination$ZkSplitTaskDetails@131aed3c in 99ms
2016-12-02 19:21:41,470 INFO  [PostOpenDeployTasks:5ea395aa4dcee819ecb70dff508ef3f6] regionserver.HRegionServer: Post open deploy tasks for splice:SPLICE_TXN,\x0D,1480686242157.5ea395aa4dcee819ecb70dff508ef3f6.
2016-12-02 19:21:41,479 INFO  [PostOpenDeployTasks:5ea395aa4dcee819ecb70dff508ef3f6] hbase.MetaTableAccessor: Updated row splice:SPLICE_TXN,\x0D,1480686242157.5ea395aa4dcee819ecb70dff508ef3f6. with server=hscale-dev1-dn1,16020,1480686617338
2016-12-02 19:21:42,074 INFO  [SplitLogWorker-hscale-dev1-dn1:16020] coordination.ZkSplitLogWorkerCoordination: worker hscale-dev1-dn1,16020,1480686617338 acquired task /hbase-unsecure/splitWAL/WALs%2Fhscale-dev1-dn4%2C16020%2C1480686241892-splitting%2Fhscale-dev1-dn4%252C16020%252C1480686241892.null5.1480686244668
2016-12-02 19:21:42,105 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-0] wal.WALSplitter: Splitting wal: hdfs://hscale-dev1-nn:8020/apps/hbase/data/WALs/hscale-dev1-dn4,16020,1480686241892-splitting/hscale-dev1-dn4%2C16020%2C1480686241892.null5.1480686244668, length=91
2016-12-02 19:21:42,105 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-0] wal.WALSplitter: DistributedLogReplay = false
2016-12-02 19:21:42,119 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-0] util.FSHDFSUtils: Recovering lease on dfs file hdfs://hscale-dev1-nn:8020/apps/hbase/data/WALs/hscale-dev1-dn4,16020,1480686241892-splitting/hscale-dev1-dn4%2C16020%2C1480686241892.null5.1480686244668
2016-12-02 19:21:42,121 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-0] util.FSHDFSUtils: recoverLease=true, attempt=0 on file=hdfs://hscale-dev1-nn:8020/apps/hbase/data/WALs/hscale-dev1-dn4,16020,1480686241892-splitting/hscale-dev1-dn4%2C16020%2C1480686241892.null5.1480686244668 after 2ms
2016-12-02 19:21:42,140 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-0] wal.WALSplitter: Split writers finished
2016-12-02 19:21:42,140 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-0] wal.WALSplitter: Processed 0 edits across 0 regions; edits skipped=0; log file=hdfs://hscale-dev1-nn:8020/apps/hbase/data/WALs/hscale-dev1-dn4,16020,1480686241892-splitting/hscale-dev1-dn4%2C16020%2C1480686241892.null5.1480686244668, length=91, corrupted=false, progress failed=false
2016-12-02 19:21:42,157 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-0] coordination.ZkSplitLogWorkerCoordination: successfully transitioned task /hbase-unsecure/splitWAL/WALs%2Fhscale-dev1-dn4%2C16020%2C1480686241892-splitting%2Fhscale-dev1-dn4%252C16020%252C1480686241892.null5.1480686244668 to final state DONE hscale-dev1-dn1,16020,1480686617338
2016-12-02 19:21:42,157 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-0] handler.WALSplitterHandler: worker hscale-dev1-dn1,16020,1480686617338 done with task org.apache.hadoop.hbase.coordination.ZkSplitLogWorkerCoordination$ZkSplitTaskDetails@638d9192 in 82ms
2016-12-02 19:21:42,785 INFO  [SplitLogWorker-hscale-dev1-dn1:16020] coordination.ZkSplitLogWorkerCoordination: worker hscale-dev1-dn1,16020,1480686617338 acquired task /hbase-unsecure/splitWAL/WALs%2Fhscale-dev1-dn4%2C16020%2C1480686241892-splitting%2Fhscale-dev1-dn4%252C16020%252C1480686241892.null3.1480686244552
2016-12-02 19:21:42,814 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-1] wal.WALSplitter: Splitting wal: hdfs://hscale-dev1-nn:8020/apps/hbase/data/WALs/hscale-dev1-dn4,16020,1480686241892-splitting/hscale-dev1-dn4%2C16020%2C1480686241892.null3.1480686244552, length=91
2016-12-02 19:21:42,814 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-1] wal.WALSplitter: DistributedLogReplay = false
2016-12-02 19:21:42,831 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-1] util.FSHDFSUtils: Recovering lease on dfs file hdfs://hscale-dev1-nn:8020/apps/hbase/data/WALs/hscale-dev1-dn4,16020,1480686241892-splitting/hscale-dev1-dn4%2C16020%2C1480686241892.null3.1480686244552
2016-12-02 19:21:42,832 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-1] util.FSHDFSUtils: recoverLease=true, attempt=0 on file=hdfs://hscale-dev1-nn:8020/apps/hbase/data/WALs/hscale-dev1-dn4,16020,1480686241892-splitting/hscale-dev1-dn4%2C16020%2C1480686241892.null3.1480686244552 after 1ms
2016-12-02 19:21:42,855 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-1] wal.WALSplitter: Split writers finished
2016-12-02 19:21:42,855 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-1] wal.WALSplitter: Processed 0 edits across 0 regions; edits skipped=0; log file=hdfs://hscale-dev1-nn:8020/apps/hbase/data/WALs/hscale-dev1-dn4,16020,1480686241892-splitting/hscale-dev1-dn4%2C16020%2C1480686241892.null3.1480686244552, length=91, corrupted=false, progress failed=false
2016-12-02 19:21:42,869 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-1] coordination.ZkSplitLogWorkerCoordination: successfully transitioned task /hbase-unsecure/splitWAL/WALs%2Fhscale-dev1-dn4%2C16020%2C1480686241892-splitting%2Fhscale-dev1-dn4%252C16020%252C1480686241892.null3.1480686244552 to final state DONE hscale-dev1-dn1,16020,1480686617338
2016-12-02 19:21:42,870 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-1] handler.WALSplitterHandler: worker hscale-dev1-dn1,16020,1480686617338 done with task org.apache.hadoop.hbase.coordination.ZkSplitLogWorkerCoordination$ZkSplitTaskDetails@46e5d85c in 85ms
2016-12-02 19:21:43,531 INFO  [PriorityRpcServer.handler=14,queue=0,port=16020] regionserver.RSRpcServices: Open splice:SPLICE_CONGLOMERATE,,1480686248935.ae9b17e1268b869b219108292bc8808b.
2016-12-02 19:21:43,543 INFO  [PriorityRpcServer.handler=14,queue=0,port=16020] regionserver.RSRpcServices: Open splice:SPLICE_TXN,\x03,1480686242157.e480b1b3cb833cb61cea26e4c0aa966e.
2016-12-02 19:21:43,556 INFO  [PriorityRpcServer.handler=14,queue=0,port=16020] regionserver.RSRpcServices: Open splice:SPLICE_TXN,\x0C,1480686242157.7347e789234a06dd5d95df6cbbbbe149.
2016-12-02 19:21:43,591 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-0] coprocessor.CoprocessorHost: System coprocessor org.apache.hadoop.hbase.security.access.SecureBulkLoadEndpoint was loaded successfully with priority (536870911).
2016-12-02 19:21:43,591 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-0] coprocessor.CoprocessorHost: System coprocessor com.splicemachine.hbase.MemstoreAwareObserver was loaded successfully with priority (536870912).
2016-12-02 19:21:43,591 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-0] coprocessor.CoprocessorHost: System coprocessor com.splicemachine.derby.hbase.SpliceIndexObserver was loaded successfully with priority (536870913).
2016-12-02 19:21:43,591 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-0] coprocessor.CoprocessorHost: System coprocessor com.splicemachine.derby.hbase.SpliceIndexEndpoint was loaded successfully with priority (536870914).
2016-12-02 19:21:43,591 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-0] coprocessor.CoprocessorHost: System coprocessor com.splicemachine.hbase.RegionSizeEndpoint was loaded successfully with priority (536870915).
2016-12-02 19:21:43,591 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-0] coprocessor.CoprocessorHost: System coprocessor com.splicemachine.si.data.hbase.coprocessor.TxnLifecycleEndpoint was loaded successfully with priority (536870916).
2016-12-02 19:21:43,591 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-0] coprocessor.CoprocessorHost: System coprocessor com.splicemachine.si.data.hbase.coprocessor.SIObserver was loaded successfully with priority (536870917).
2016-12-02 19:21:43,591 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-0] coprocessor.CoprocessorHost: System coprocessor com.splicemachine.hbase.BackupEndpointObserver was loaded successfully with priority (536870918).
2016-12-02 19:21:43,601 INFO  [StoreOpener-ae9b17e1268b869b219108292bc8808b-1] hfile.CacheConfig: blockCache=LruBlockCache{blockCount=7, currentSize=852992, freeSize=804453376, maxSize=805306368, heapSize=852992, minSize=765041024, minFactor=0.95, multiSize=382520512, multiFactor=0.5, singleSize=191260256, singleFactor=0.25}, cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=true, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false
2016-12-02 19:21:43,602 INFO  [StoreOpener-ae9b17e1268b869b219108292bc8808b-1] compactions.CompactionConfiguration: size [16777216, 260046848); files [5, 10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000
2016-12-02 19:21:43,612 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-0] regionserver.HRegion: Replaying edits from hdfs://hscale-dev1-nn:8020/apps/hbase/data/data/splice/SPLICE_CONGLOMERATE/ae9b17e1268b869b219108292bc8808b/recovered.edits/0000000000000000003
2016-12-02 19:21:43,624 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-2] coprocessor.CoprocessorHost: System coprocessor org.apache.hadoop.hbase.security.access.SecureBulkLoadEndpoint was loaded successfully with priority (536870911).
2016-12-02 19:21:43,624 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-2] coprocessor.CoprocessorHost: System coprocessor com.splicemachine.hbase.MemstoreAwareObserver was loaded successfully with priority (536870912).
2016-12-02 19:21:43,624 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-2] coprocessor.CoprocessorHost: System coprocessor com.splicemachine.derby.hbase.SpliceIndexObserver was loaded successfully with priority (536870913).
2016-12-02 19:21:43,624 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-2] coprocessor.CoprocessorHost: System coprocessor com.splicemachine.derby.hbase.SpliceIndexEndpoint was loaded successfully with priority (536870914).
2016-12-02 19:21:43,624 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-2] coprocessor.CoprocessorHost: System coprocessor com.splicemachine.hbase.RegionSizeEndpoint was loaded successfully with priority (536870915).
2016-12-02 19:21:43,624 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-2] coprocessor.CoprocessorHost: System coprocessor com.splicemachine.si.data.hbase.coprocessor.TxnLifecycleEndpoint was loaded successfully with priority (536870916).
2016-12-02 19:21:43,625 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-2] coprocessor.CoprocessorHost: System coprocessor com.splicemachine.si.data.hbase.coprocessor.SIObserver was loaded successfully with priority (536870917).
2016-12-02 19:21:43,625 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-2] coprocessor.CoprocessorHost: System coprocessor com.splicemachine.hbase.BackupEndpointObserver was loaded successfully with priority (536870918).
2016-12-02 19:21:43,627 INFO  [StoreOpener-e480b1b3cb833cb61cea26e4c0aa966e-1] hfile.CacheConfig: blockCache=LruBlockCache{blockCount=7, currentSize=852992, freeSize=804453376, maxSize=805306368, heapSize=852992, minSize=765041024, minFactor=0.95, multiSize=382520512, multiFactor=0.5, singleSize=191260256, singleFactor=0.25}, cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=true, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false
2016-12-02 19:21:43,628 INFO  [StoreOpener-e480b1b3cb833cb61cea26e4c0aa966e-1] compactions.CompactionConfiguration: size [16777216, 260046848); files [5, 10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000
2016-12-02 19:21:43,630 INFO  [StoreOpener-e480b1b3cb833cb61cea26e4c0aa966e-1] hfile.CacheConfig: blockCache=LruBlockCache{blockCount=7, currentSize=852992, freeSize=804453376, maxSize=805306368, heapSize=852992, minSize=765041024, minFactor=0.95, multiSize=382520512, multiFactor=0.5, singleSize=191260256, singleFactor=0.25}, cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=true, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false
2016-12-02 19:21:43,631 INFO  [StoreOpener-e480b1b3cb833cb61cea26e4c0aa966e-1] compactions.CompactionConfiguration: size [16777216, 260046848); files [5, 10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000
2016-12-02 19:21:43,637 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-2] regionserver.HRegion: Replaying edits from hdfs://hscale-dev1-nn:8020/apps/hbase/data/data/splice/SPLICE_TXN/e480b1b3cb833cb61cea26e4c0aa966e/recovered.edits/0000000000000000003
2016-12-02 19:21:43,673 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-1] coprocessor.CoprocessorHost: System coprocessor org.apache.hadoop.hbase.security.access.SecureBulkLoadEndpoint was loaded successfully with priority (536870911).
2016-12-02 19:21:43,673 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-1] coprocessor.CoprocessorHost: System coprocessor com.splicemachine.hbase.MemstoreAwareObserver was loaded successfully with priority (536870912).
2016-12-02 19:21:43,673 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-1] coprocessor.CoprocessorHost: System coprocessor com.splicemachine.derby.hbase.SpliceIndexObserver was loaded successfully with priority (536870913).
2016-12-02 19:21:43,673 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-1] coprocessor.CoprocessorHost: System coprocessor com.splicemachine.derby.hbase.SpliceIndexEndpoint was loaded successfully with priority (536870914).
2016-12-02 19:21:43,674 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-1] coprocessor.CoprocessorHost: System coprocessor com.splicemachine.hbase.RegionSizeEndpoint was loaded successfully with priority (536870915).
2016-12-02 19:21:43,674 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-1] coprocessor.CoprocessorHost: System coprocessor com.splicemachine.si.data.hbase.coprocessor.TxnLifecycleEndpoint was loaded successfully with priority (536870916).
2016-12-02 19:21:43,674 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-1] coprocessor.CoprocessorHost: System coprocessor com.splicemachine.si.data.hbase.coprocessor.SIObserver was loaded successfully with priority (536870917).
2016-12-02 19:21:43,674 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-1] coprocessor.CoprocessorHost: System coprocessor com.splicemachine.hbase.BackupEndpointObserver was loaded successfully with priority (536870918).
2016-12-02 19:21:43,679 INFO  [StoreOpener-7347e789234a06dd5d95df6cbbbbe149-1] hfile.CacheConfig: blockCache=LruBlockCache{blockCount=7, currentSize=852992, freeSize=804453376, maxSize=805306368, heapSize=852992, minSize=765041024, minFactor=0.95, multiSize=382520512, multiFactor=0.5, singleSize=191260256, singleFactor=0.25}, cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=true, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false
2016-12-02 19:21:43,680 INFO  [StoreOpener-7347e789234a06dd5d95df6cbbbbe149-1] compactions.CompactionConfiguration: size [16777216, 260046848); files [5, 10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000
2016-12-02 19:21:43,685 INFO  [StoreOpener-7347e789234a06dd5d95df6cbbbbe149-1] hfile.CacheConfig: blockCache=LruBlockCache{blockCount=7, currentSize=852992, freeSize=804453376, maxSize=805306368, heapSize=852992, minSize=765041024, minFactor=0.95, multiSize=382520512, multiFactor=0.5, singleSize=191260256, singleFactor=0.25}, cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=true, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false
2016-12-02 19:21:43,686 INFO  [StoreOpener-7347e789234a06dd5d95df6cbbbbe149-1] compactions.CompactionConfiguration: size [16777216, 260046848); files [5, 10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000
2016-12-02 19:21:43,689 INFO  [SplitLogWorker-hscale-dev1-dn1:16020] coordination.ZkSplitLogWorkerCoordination: worker hscale-dev1-dn1,16020,1480686617338 acquired task /hbase-unsecure/splitWAL/WALs%2Fhscale-dev1-dn4%2C16020%2C1480686241892-splitting%2Fhscale-dev1-dn4%252C16020%252C1480686241892.null2.1480686244470
2016-12-02 19:21:43,698 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-1] regionserver.HRegion: Replaying edits from hdfs://hscale-dev1-nn:8020/apps/hbase/data/data/splice/SPLICE_TXN/7347e789234a06dd5d95df6cbbbbe149/recovered.edits/0000000000000000003
2016-12-02 19:21:43,718 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-0] wal.WALSplitter: Splitting wal: hdfs://hscale-dev1-nn:8020/apps/hbase/data/WALs/hscale-dev1-dn4,16020,1480686241892-splitting/hscale-dev1-dn4%2C16020%2C1480686241892.null2.1480686244470, length=91
2016-12-02 19:21:43,718 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-0] wal.WALSplitter: DistributedLogReplay = false
2016-12-02 19:21:43,736 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-0] util.FSHDFSUtils: Recovering lease on dfs file hdfs://hscale-dev1-nn:8020/apps/hbase/data/WALs/hscale-dev1-dn4,16020,1480686241892-splitting/hscale-dev1-dn4%2C16020%2C1480686241892.null2.1480686244470
2016-12-02 19:21:43,738 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-0] util.FSHDFSUtils: recoverLease=true, attempt=0 on file=hdfs://hscale-dev1-nn:8020/apps/hbase/data/WALs/hscale-dev1-dn4,16020,1480686241892-splitting/hscale-dev1-dn4%2C16020%2C1480686241892.null2.1480686244470 after 2ms
2016-12-02 19:21:43,738 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-0] regionserver.HRegion: Onlined ae9b17e1268b869b219108292bc8808b; next sequenceid=4
2016-12-02 19:21:43,743 INFO  [PostOpenDeployTasks:ae9b17e1268b869b219108292bc8808b] regionserver.HRegionServer: Post open deploy tasks for splice:SPLICE_CONGLOMERATE,,1480686248935.ae9b17e1268b869b219108292bc8808b.
2016-12-02 19:21:43,747 INFO  [PostOpenDeployTasks:ae9b17e1268b869b219108292bc8808b] hbase.MetaTableAccessor: Updated row splice:SPLICE_CONGLOMERATE,,1480686248935.ae9b17e1268b869b219108292bc8808b. with server=hscale-dev1-dn1,16020,1480686617338
2016-12-02 19:21:43,750 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-2] regionserver.HRegion: Onlined e480b1b3cb833cb61cea26e4c0aa966e; next sequenceid=4
2016-12-02 19:21:43,767 INFO  [PostOpenDeployTasks:e480b1b3cb833cb61cea26e4c0aa966e] regionserver.HRegionServer: Post open deploy tasks for splice:SPLICE_TXN,\x03,1480686242157.e480b1b3cb833cb61cea26e4c0aa966e.
2016-12-02 19:21:43,771 INFO  [PostOpenDeployTasks:e480b1b3cb833cb61cea26e4c0aa966e] hbase.MetaTableAccessor: Updated row splice:SPLICE_TXN,\x03,1480686242157.e480b1b3cb833cb61cea26e4c0aa966e. with server=hscale-dev1-dn1,16020,1480686617338
2016-12-02 19:21:43,788 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-0] wal.WALSplitter: Split writers finished
2016-12-02 19:21:43,788 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-0] wal.WALSplitter: Processed 0 edits across 0 regions; edits skipped=0; log file=hdfs://hscale-dev1-nn:8020/apps/hbase/data/WALs/hscale-dev1-dn4,16020,1480686241892-splitting/hscale-dev1-dn4%2C16020%2C1480686241892.null2.1480686244470, length=91, corrupted=false, progress failed=false
2016-12-02 19:21:43,800 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-1] regionserver.HRegion: Onlined 7347e789234a06dd5d95df6cbbbbe149; next sequenceid=4
2016-12-02 19:21:43,817 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-0] coordination.ZkSplitLogWorkerCoordination: successfully transitioned task /hbase-unsecure/splitWAL/WALs%2Fhscale-dev1-dn4%2C16020%2C1480686241892-splitting%2Fhscale-dev1-dn4%252C16020%252C1480686241892.null2.1480686244470 to final state DONE hscale-dev1-dn1,16020,1480686617338
2016-12-02 19:21:43,817 INFO  [RS_LOG_REPLAY_OPS-hscale-dev1-dn1:16020-0] handler.WALSplitterHandler: worker hscale-dev1-dn1,16020,1480686617338 done with task org.apache.hadoop.hbase.coordination.ZkSplitLogWorkerCoordination$ZkSplitTaskDetails@46522228 in 127ms
2016-12-02 19:21:43,817 INFO  [PostOpenDeployTasks:7347e789234a06dd5d95df6cbbbbe149] regionserver.HRegionServer: Post open deploy tasks for splice:SPLICE_TXN,\x0C,1480686242157.7347e789234a06dd5d95df6cbbbbe149.
2016-12-02 19:21:43,830 INFO  [PostOpenDeployTasks:7347e789234a06dd5d95df6cbbbbe149] hbase.MetaTableAccessor: Updated row splice:SPLICE_TXN,\x0C,1480686242157.7347e789234a06dd5d95df6cbbbbe149. with server=hscale-dev1-dn1,16020,1480686617338
2016-12-02 19:21:44,176 INFO  [PriorityRpcServer.handler=2,queue=0,port=16020] regionserver.RSRpcServices: Open splice:16,,1480686254027.75c25e3e9e9f5281c8face7c909f0b04.
2016-12-02 19:21:44,222 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-0] coprocessor.CoprocessorHost: System coprocessor org.apache.hadoop.hbase.security.access.SecureBulkLoadEndpoint was loaded successfully with priority (536870911).
2016-12-02 19:21:44,222 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-0] coprocessor.CoprocessorHost: System coprocessor com.splicemachine.hbase.MemstoreAwareObserver was loaded successfully with priority (536870912).
2016-12-02 19:21:44,223 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-0] coprocessor.CoprocessorHost: System coprocessor com.splicemachine.derby.hbase.SpliceIndexObserver was loaded successfully with priority (536870913).
2016-12-02 19:21:44,223 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-0] coprocessor.CoprocessorHost: System coprocessor com.splicemachine.derby.hbase.SpliceIndexEndpoint was loaded successfully with priority (536870914).
2016-12-02 19:21:44,223 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-0] coprocessor.CoprocessorHost: System coprocessor com.splicemachine.hbase.RegionSizeEndpoint was loaded successfully with priority (536870915).
2016-12-02 19:21:44,223 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-0] coprocessor.CoprocessorHost: System coprocessor com.splicemachine.si.data.hbase.coprocessor.TxnLifecycleEndpoint was loaded successfully with priority (536870916).
2016-12-02 19:21:44,223 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-0] coprocessor.CoprocessorHost: System coprocessor com.splicemachine.si.data.hbase.coprocessor.SIObserver was loaded successfully with priority (536870917).
2016-12-02 19:21:44,223 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-0] coprocessor.CoprocessorHost: System coprocessor com.splicemachine.hbase.BackupEndpointObserver was loaded successfully with priority (536870918).
2016-12-02 19:21:44,227 INFO  [StoreOpener-75c25e3e9e9f5281c8face7c909f0b04-1] hfile.CacheConfig: blockCache=LruBlockCache{blockCount=7, currentSize=852992, freeSize=804453376, maxSize=805306368, heapSize=852992, minSize=765041024, minFactor=0.95, multiSize=382520512, multiFactor=0.5, singleSize=191260256, singleFactor=0.25}, cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=true, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false
2016-12-02 19:21:44,227 INFO  [StoreOpener-75c25e3e9e9f5281c8face7c909f0b04-1] compactions.CompactionConfiguration: size [16777216, 260046848); files [5, 10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000
2016-12-02 19:21:44,236 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-0] regionserver.HRegion: Replaying edits from hdfs://hscale-dev1-nn:8020/apps/hbase/data/data/splice/16/75c25e3e9e9f5281c8face7c909f0b04/recovered.edits/0000000000000000004
2016-12-02 19:21:44,245 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-0] regionserver.HRegion: Replaying edits from hdfs://hscale-dev1-nn:8020/apps/hbase/data/data/splice/16/75c25e3e9e9f5281c8face7c909f0b04/recovered.edits/0000000000000000007
2016-12-02 19:21:44,253 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-0] regionserver.HRegion: Replaying edits from hdfs://hscale-dev1-nn:8020/apps/hbase/data/data/splice/16/75c25e3e9e9f5281c8face7c909f0b04/recovered.edits/0000000000000000011
2016-12-02 19:21:44,354 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-0] regionserver.HRegion: Onlined 75c25e3e9e9f5281c8face7c909f0b04; next sequenceid=12
2016-12-02 19:21:44,370 INFO  [PostOpenDeployTasks:75c25e3e9e9f5281c8face7c909f0b04] regionserver.HRegionServer: Post open deploy tasks for splice:16,,1480686254027.75c25e3e9e9f5281c8face7c909f0b04.
2016-12-02 19:21:44,386 INFO  [PostOpenDeployTasks:75c25e3e9e9f5281c8face7c909f0b04] hbase.MetaTableAccessor: Updated row splice:16,,1480686254027.75c25e3e9e9f5281c8face7c909f0b04. with server=hscale-dev1-dn1,16020,1480686617338
2016-12-02 19:21:45,462 INFO  [PriorityRpcServer.handler=19,queue=1,port=16020] regionserver.RSRpcServices: Open SPLICE_INIT,,1480686619560.54d34b28fb70efee998fdacb4efb262a.
2016-12-02 19:21:45,514 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-2] coprocessor.CoprocessorHost: System coprocessor org.apache.hadoop.hbase.security.access.SecureBulkLoadEndpoint was loaded successfully with priority (536870911).
2016-12-02 19:21:45,515 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-2] coprocessor.CoprocessorHost: System coprocessor com.splicemachine.hbase.MemstoreAwareObserver was loaded successfully with priority (536870912).
2016-12-02 19:21:45,515 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-2] coprocessor.CoprocessorHost: System coprocessor com.splicemachine.derby.hbase.SpliceIndexObserver was loaded successfully with priority (536870913).
2016-12-02 19:21:45,515 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-2] coprocessor.CoprocessorHost: System coprocessor com.splicemachine.derby.hbase.SpliceIndexEndpoint was loaded successfully with priority (536870914).
2016-12-02 19:21:45,515 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-2] coprocessor.CoprocessorHost: System coprocessor com.splicemachine.hbase.RegionSizeEndpoint was loaded successfully with priority (536870915).
2016-12-02 19:21:45,516 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-2] coprocessor.CoprocessorHost: System coprocessor com.splicemachine.si.data.hbase.coprocessor.TxnLifecycleEndpoint was loaded successfully with priority (536870916).
2016-12-02 19:21:45,516 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-2] coprocessor.CoprocessorHost: System coprocessor com.splicemachine.si.data.hbase.coprocessor.SIObserver was loaded successfully with priority (536870917).
2016-12-02 19:21:45,516 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-2] coprocessor.CoprocessorHost: System coprocessor com.splicemachine.hbase.BackupEndpointObserver was loaded successfully with priority (536870918).
2016-12-02 19:21:45,540 INFO  [StoreOpener-54d34b28fb70efee998fdacb4efb262a-1] hfile.CacheConfig: blockCache=LruBlockCache{blockCount=7, currentSize=852992, freeSize=804453376, maxSize=805306368, heapSize=852992, minSize=765041024, minFactor=0.95, multiSize=382520512, multiFactor=0.5, singleSize=191260256, singleFactor=0.25}, cacheDataOnRead=true, cacheDataOnWrite=false, cacheIndexesOnWrite=false, cacheBloomsOnWrite=true, cacheEvictOnClose=false, cacheDataCompressed=false, prefetchOnOpen=false
2016-12-02 19:21:45,541 INFO  [StoreOpener-54d34b28fb70efee998fdacb4efb262a-1] compactions.CompactionConfiguration: size [16777216, 260046848); files [5, 10); ratio 1.200000; off-peak ratio 5.000000; throttle point 2684354560; major period 604800000, major jitter 0.500000, min locality to compact 0.000000
2016-12-02 19:21:45,583 INFO  [RS_OPEN_REGION-hscale-dev1-dn1:16020-2] regionserver.HRegion: Onlined 54d34b28fb70efee998fdacb4efb262a; next sequenceid=2
2016-12-02 19:21:45,592 INFO  [PostOpenDeployTasks:54d34b28fb70efee998fdacb4efb262a] regionserver.HRegionServer: Post open deploy tasks for SPLICE_INIT,,1480686619560.54d34b28fb70efee998fdacb4efb262a.
2016-12-02 19:21:45,600 INFO  [PostOpenDeployTasks:54d34b28fb70efee998fdacb4efb262a] hbase.MetaTableAccessor: Updated row SPLICE_INIT,,1480686619560.54d34b28fb70efee998fdacb4efb262a. with server=hscale-dev1-dn1,16020,1480686617338
2016-12-02 19:21:58,216 INFO  [pool-10-thread-1] client.HBaseAdmin: Created SPLICE_INIT
2016-12-02 19:21:58,766 INFO  [pool-10-thread-1] db.SpliceDatabase: Booting the Splice Machine database
2016-12-02 19:21:58,817 ERROR [pool-10-thread-1] lifecycle.DatabaseLifecycleManager: Error during during startup of service com.splicemachine.derby.lifecycle.MonitoredLifecycleService@5f15ea70:
java.sql.SQLException: Failed to start database 'splicedb' with class loader sun.misc.Launcher$AppClassLoader@18b4aac2, see the next exception for details.
    at com.splicemachine.db.impl.jdbc.SQLExceptionFactory40.getSQLException(SQLExceptionFactory40.java:102)
    at com.splicemachine.db.impl.jdbc.Util.newEmbedSQLException(Util.java:170)
    at com.splicemachine.db.impl.jdbc.Util.seeNextException(Util.java:306)
    at com.splicemachine.db.impl.jdbc.EmbedConnection.bootDatabase(EmbedConnection.java:2326)
    at com.splicemachine.db.impl.jdbc.EmbedConnection.<init>(EmbedConnection.java:302)
    at com.splicemachine.db.impl.jdbc.EmbedConnection30.<init>(EmbedConnection30.java:72)
    at com.splicemachine.db.impl.jdbc.EmbedConnection40.<init>(EmbedConnection40.java:57)
    at com.splicemachine.db.jdbc.Driver40.getNewEmbedConnection(Driver40.java:69)
    at com.splicemachine.db.jdbc.InternalDriver.connect(InternalDriver.java:256)
    at com.splicemachine.db.jdbc.EmbeddedDriver.connect(EmbeddedDriver.java:125)
    at com.splicemachine.tools.EmbedConnectionMaker.createNew(EmbedConnectionMaker.java:42)
    at com.splicemachine.derby.lifecycle.EngineLifecycleService.start(EngineLifecycleService.java:98)
    at com.splicemachine.lifecycle.DatabaseLifecycleManager$Startup.bootServices(DatabaseLifecycleManager.java:229)
    at com.splicemachine.lifecycle.DatabaseLifecycleManager$Startup.run(DatabaseLifecycleManager.java:220)
    at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
    at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
    at java.lang.Thread.run(Thread.java:745)
Caused by: java.sql.SQLException: Failed to start database 'splicedb' with class loader sun.misc.Launcher$AppClassLoader@18b4aac2, see the next exception for details.
    at com.splicemachine.db.impl.jdbc.SQLExceptionFactory.getSQLException(SQLExceptionFactory.java:46)
    at com.splicemachine.db.impl.jdbc.SQLExceptionFactory40.wrapArgsForTransportAcrossDRDA(SQLExceptionFactory40.java:126)
    at com.splicemachine.db.impl.jdbc.SQLExceptionFactory40.getSQLException(SQLExceptionFactory40.java:75)
    ... 16 more
Caused by: java.sql.SQLException: Java exception: ': java.lang.NullPointerException'.
    at com.splicemachine.db.impl.jdbc.SQLExceptionFactory.getSQLException(SQLExceptionFactory.java:46)
    at com.splicemachine.db.impl.jdbc.SQLExceptionFactory40.wrapArgsForTransportAcrossDRDA(SQLExceptionFactory40.java:126)
    at com.splicemachine.db.impl.jdbc.SQLExceptionFactory40.getSQLException(SQLExceptionFactory40.java:75)
    at com.splicemachine.db.impl.jdbc.Util.newEmbedSQLException(Util.java:170)
    at com.splicemachine.db.impl.jdbc.Util.javaException(Util.java:327)
    at com.splicemachine.db.impl.jdbc.EmbedConnection.bootDatabase(EmbedConnection.java:2322)
    ... 13 more
Caused by: java.lang.NullPointerException
    at com.splicemachine.pipeline.Exceptions.parseException(Exceptions.java:37)
    at com.splicemachine.derby.impl.sql.ZkPropertyManager.getProperty(ZkPropertyManager.java:65)
    at com.splicemachine.derby.impl.store.access.PropertyConglomerate.<init>(PropertyConglomerate.java:64)
    at com.splicemachine.derby.impl.store.access.SpliceAccessManager.boot(SpliceAccessManager.java:687)
    at com.splicemachine.db.impl.services.monitor.BaseMonitor.boot(BaseMonitor.java:1996)
    at com.splicemachine.db.impl.services.monitor.TopService.bootModule(TopService.java:337)
    at com.splicemachine.db.impl.services.monitor.BaseMonitor.startModule(BaseMonitor.java:545)
    at com.splicemachine.db.impl.services.monitor.FileMonitor.startModule(FileMonitor.java:51)
    at com.splicemachine.db.iapi.services.monitor.Monitor.bootServiceModule(Monitor.java:430)
    at com.splicemachine.derby.impl.db.SpliceDatabase.bootStore(SpliceDatabase.java:447)
    at com.splicemachine.db.impl.db.BasicDatabase.boot(BasicDatabase.java:164)
    at com.splicemachine.derby.impl.db.SpliceDatabase.boot(SpliceDatabase.java:115)
    at com.splicemachine.db.impl.services.monitor.BaseMonitor.boot(BaseMonitor.java:1996)
    at com.splicemachine.db.impl.services.monitor.TopService.bootModule(TopService.java:337)
    at com.splicemachine.db.impl.services.monitor.BaseMonitor.bootService(BaseMonitor.java:1830)
    at com.splicemachine.db.impl.services.monitor.BaseMonitor.startProviderService(BaseMonitor.java:1696)
    at com.splicemachine.db.impl.services.monitor.BaseMonitor.findProviderAndStartService(BaseMonitor.java:1574)
    at com.splicemachine.db.impl.services.monitor.BaseMonitor.startPersistentService(BaseMonitor.java:993)
    at com.splicemachine.db.iapi.services.monitor.Monitor.startPersistentService(Monitor.java:553)
    at com.splicemachine.db.impl.jdbc.EmbedConnection.bootDatabase(EmbedConnection.java:2287)
    ... 13 more
2016-12-02 19:21:58,819 ERROR [pool-10-thread-1] lifecycle.DatabaseLifecycleManager: Error during shutdown of service com.splicemachine.derby.lifecycle.NetworkLifecycleService@6075dcea:
java.lang.NullPointerException
    at com.splicemachine.derby.lifecycle.NetworkLifecycleService.shutdown(NetworkLifecycleService.java:65)
    at com.splicemachine.lifecycle.DatabaseLifecycleManager$Shutdown.run(DatabaseLifecycleManager.java:268)
    at com.splicemachine.lifecycle.DatabaseLifecycleManager$Startup.bootServices(DatabaseLifecycleManager.java:233)
    at com.splicemachine.lifecycle.DatabaseLifecycleManager$Startup.run(DatabaseLifecycleManager.java:220)
    at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
    at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
    at java.lang.Thread.run(Thread.java:745)
2016-12-02 19:21:58,819 INFO  [pool-10-thread-1] impl.TimestampClient: shutting down TimestampClient state=SHUTDOWN
2016-12-02 19:22:17,395 WARN  [timeline] timeline.HadoopTimelineMetricsSink: Unable to send metrics to collector by address:http://hscale-dev1-nn:6188/ws/v1/timeline/metrics
2016-12-02 19:22:17,396 WARN  [timeline] timeline.HadoopTimelineMetricsSink: Unable to send metrics to collector by address:http://hscale-dev1-nn:6188/ws/v1/timeline/metrics
2016-12-02 19:22:17,400 WARN  [timeline] timeline.HadoopTimelineMetricsSink: Unable to send metrics to collector by address:http://hscale-dev1-nn:6188/ws/v1/timeline/metrics
2016-12-02 19:22:17,401 WARN  [timeline] timeline.HadoopTimelineMetricsSink: Unable to send metrics to collector by address:http://hscale-dev1-nn:6188/ws/v1/timeline/metrics
2016-12-02 19:22:17,402 WARN  [timeline] timeline.HadoopTimelineMetricsSink: Unable to send metrics to collector by address:http://hscale-dev1-nn:6188/ws/v1/timeline/metrics
2016-12-02 19:23:17,338 WARN  [timeline] timeline.HadoopTimelineMetricsSink: Unable to send metrics to collector by address:http://hscale-dev1-nn:6188/ws/v1/timeline/metrics
2016-12-02 19:23:17,342 WARN  [timeline] timeline.HadoopTimelineMetricsSink: Unable to send metrics to collector by address:http://hscale-dev1-nn:6188/ws/v1/timeline/metrics