per_node_diag = [{'ns_1@10.1.5.227', [{version, [{kernel,"2.13.4"}, {mnesia,"4.4.12"}, {stdlib,"1.16.4"}, {os_mon,"2.2.4"}, {inets,"5.2"}, {ns_server,"1.6.5.4r"}, {menelaus,"1.6.5.4r"}, {sasl,"2.1.8"}]}, {manifest, ["bucket_engine_1.6.5.3_1_g5ba5371-Linux.x86_64.tar.gz", "curl-7.21.1-w64_patched.tar.gz", "ep-engine_1.6.5.3.1_13_g81a6cf1-Linux.x86_64.tar.gz", "google-perftools-1.6","google-perftools-1.6.tar.gz", "libconflate_1.6.5-Linux.x86_64.tar.gz", "libevent-2.0.7-rc.tar.gz", "libmemcached-0.41_trond-norbye_mingw32-revno895.tar.gz", "libvbucket_1.6.4-Linux.x86_64.tar.gz", "membase-cli_1.6.5-Linux.x86_64.tar.gz", "memcached_1.4.4_382_g9df3289-Linux.x86_64.tar.gz", "moxi_1.6.5.2-Linux.x86_64.tar.gz", "ns_server_1.6.5.4r.tar.gz", "vbucketmigrator_1.6.5_2_gd2c8834-Linux.x86_64.tar.gz", "wallace_1.6.5.4r-8-g00eec2b-Linux.x86_64"]}, {config, [{{node,'ns_1@127.0.0.1',ns_log}, [{filename, "/var/opt/membase/1.6.5.4r/data/ns_1/ns_log"}]}, {otp, [{'_vclock',[{'ns_1@10.1.5.229',{1,63472280023}}]}, {cookie,bujipmggocppuqmx}]}, {memory_quota,1570}, {buckets, [{'_vclock', [{'ns_1@10.1.5.227',{2,63472280025}}, {'ns_1@127.0.0.1',{3,63472280007}}]}, {configs, [{"default", [{num_replicas,1}, {ram_quota,1096810496}, {auth_type,sasl}, {sasl_password,[]}, {type,membase}, {num_vbuckets,1024}, {db_shards,4}, {ht_size,3079}, {tap_keepalive,0}, {tap_noop_interval,20}, {max_txn_size,1000}, {ht_locks,5}, {servers,['ns_1@10.1.5.227','ns_1@10.1.5.229']}, {map, [['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined]]}]}]}]}, {{node,'ns_1@127.0.0.1',membership},active}, {rest_creds, [{'_vclock',[{'ns_1@127.0.0.1',{1,63472279813}}]}, {creds, [{"Administrator",[{password,'filtered-out'}]}]}]}, {port_servers, [{moxi,"./bin/moxi/moxi", ["-Z", {"port_listen=~B,default_bucket_name=default,downstream_max=1024,downstream_conn_max=4,connect_max_errors=5,connect_retry_interval=30000,connect_timeout=400,auth_timeout=100,cycle=200,downstream_conn_queue_timeout=200,downstream_timeout=5000,wait_queue_timeout=200", [port]}, "-z", {"url=http://127.0.0.1:~B/pools/default/saslBucketsStreaming", [{rest,port}]}, "-p","0","-Y","y","-O","stderr", {"~s",[verbosity]}], [{env, [{"EVENT_NOSELECT","1"}, {"MOXI_SASL_PLAIN_USR", {"~s",[{ns_moxi_sup,rest_user,[]}]}}, {"MOXI_SASL_PLAIN_PWD", {"~s",[{ns_moxi_sup,rest_pass,[]}]}}]}, use_stdio,stderr_to_stdout,stream]}, {memcached,"./bin/memcached/memcached", ["-X","./bin/memcached/stdin_term_handler.so","-p", {"~B",[port]}, "-E","./bin/bucket_engine/bucket_engine.so","-B", "binary","-r","-c","10000","-e", {"admin=~s;default_bucket_name=default;auto_create=false", [admin_user]}, {"~s",[verbosity]}], [{env, [{"EVENT_NOSELECT","1"}, {"MEMCACHED_TOP_KEYS","100"}, {"ISASL_PWFILE",{"~s",[{isasl,path}]}}, {"ISASL_DB_CHECK_TIME","1"}]}, use_stdio,stderr_to_stdout,stream]}]}, {alerts, [{email,[]}, {email_alerts,false}, {email_server, [{user,undefined}, {pass,'filtered-out'}, {addr,undefined}, {port,undefined}, {encrypt,false}]}, {alerts, [server_down,server_unresponsive,server_up, server_joined,server_left,bucket_created, bucket_deleted,bucket_auth_failed]}]}, {rebalance_status,running}, {nodes_wanted, [{'_vclock',[{'ns_1@10.1.5.227',{2,63472280023}}]}, 'ns_1@10.1.5.227','ns_1@10.1.5.229']}, {rest, [{'_vclock',[{'ns_1@127.0.0.1',{1,63472279813}}]}, {port,8091}]}, {{node,'ns_1@10.1.5.227',memcached}, [{'_vclock',[{'ns_1@10.1.5.227',{1,63472280022}}]}, {port,11210}, {dbdir,"/var/opt/membase/1.6.5.4r/data/ns_1"}, {admin_user,"_admin"}, {admin_pass,"_admin"}, {bucket_engine,"./bin/bucket_engine/bucket_engine.so"}, {engines, [{membase, [{engine,"bin/ep_engine/ep.so"}, {initfile,"/etc/opt/membase/1.6.5.4r/init.sql"}]}, {memcached, [{engine,"bin/memcached/default_engine.so"}]}]}, {verbosity,[]}]}, {{node,'ns_1@10.1.5.227',ns_log}, [{'_vclock',[{'ns_1@10.1.5.227',{1,63472280022}}]}, {filename, "/var/opt/membase/1.6.5.4r/data/ns_1/ns_log"}]}, {{node,'ns_1@127.0.0.1',isasl}, [{path,"/var/opt/membase/1.6.5.4r/data/ns_1/isasl.pw"}]}, {{node,'ns_1@10.1.5.229',memcached}, [{'_vclock',[{'ns_1@10.1.5.229',{1,63472280023}}]}, {port,11210}, {dbdir,"/var/opt/membase/1.6.5.4r/data/ns_1"}, {admin_user,"_admin"}, {admin_pass,"_admin"}, {bucket_engine,"./bin/bucket_engine/bucket_engine.so"}, {engines, [{membase, [{engine,"bin/ep_engine/ep.so"}, {initfile,"/etc/opt/membase/1.6.5.4r/init.sql"}]}, {memcached, [{engine,"bin/memcached/default_engine.so"}]}]}, {verbosity,[]}]}, {{node,'ns_1@10.1.5.229',ns_log}, [{'_vclock',[{'ns_1@10.1.5.229',{1,63472280023}}]}, {filename, "/var/opt/membase/1.6.5.4r/data/ns_1/ns_log"}]}, {moxi,[{port,11211},{verbosity,[]}]}, {{node,'ns_1@10.1.5.227',membership},active}, {{node,'ns_1@10.1.5.229',membership},active}, {{node,'ns_1@10.1.5.227',isasl}, [{'_vclock',[{'ns_1@10.1.5.227',{1,63472280022}}]}, {path,"/var/opt/membase/1.6.5.4r/data/ns_1/isasl.pw"}]}, {replication,[{enabled,true}]}, {{node,'ns_1@127.0.0.1',memcached}, [{port,11210}, {dbdir,"/var/opt/membase/1.6.5.4r/data/ns_1"}, {admin_user,"_admin"}, {admin_pass,"_admin"}, {bucket_engine,"./bin/bucket_engine/bucket_engine.so"}, {engines, [{membase, [{engine,"bin/ep_engine/ep.so"}, {initfile,"/etc/opt/membase/1.6.5.4r/init.sql"}]}, {memcached, [{engine,"bin/memcached/default_engine.so"}]}]}, {verbosity,[]}]}, {{node,'ns_1@10.1.5.229',isasl}, [{'_vclock',[{'ns_1@10.1.5.229',{1,63472280023}}]}, {path, "/var/opt/membase/1.6.5.4r/data/ns_1/isasl.pw"}]}]}, {basic_info, [{version, [{kernel,"2.13.4"}, {mnesia,"4.4.12"}, {stdlib,"1.16.4"}, {os_mon,"2.2.4"}, {inets,"5.2"}, {ns_server,"1.6.5.4r"}, {menelaus,"1.6.5.4r"}, {sasl,"2.1.8"}]}, {system_arch,"x86_64-unknown-linux-gnu"}, {wall_clock,1480}, {memory_data,{2058731520,680501248,{<0.12.0>,6657168}}}, {disk_data, [{"/",11204616,49}, {"/boot",101086,14}, {"/dev/shm",1005240,0}]}]}, {processes, [{<0.0.0>, [{registered_name,init}, {status,waiting}, {initial_call,{otp_ring0,start,2}}, {backtrace, ["Program counter: 0x00002b7413276468 (init:loop/1 + 40)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002aaaaac4d6a8 Return addr 0x0000000000846c78 ()", "y(0) {state,[{'-root',[<<45 bytes>>]},{'-progname',[<<3 bytes>>]},{'-home',[<<12 bytes"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,190}]}, {heap_size,4181}, {total_heap_size,5778}, {links,[<0.6.0>,<0.7.0>,<0.3.0>]}, {memory,47048}, {message_queue_len,0}, {reductions,70095}, {trap_exit,true}]}, {<0.3.0>, [{registered_name,erl_prim_loader}, {status,waiting}, {initial_call,{erlang,apply,2}}, {backtrace, ["Program counter: 0x00002b74132d5ef8 (erl_prim_loader:loop/3 + 184)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002aaaabb9c218 Return addr 0x0000000000846c78 ()", "y(0) []", "y(1) [\"/opt/membase/1.6.5.4r/bin/ns_server/ebin\",\"/opt/membase/1.6.5.4r/bin/ns_server/", "y(2) <0.2.0>", "y(3) {state,efile,[],none,#Port<0.2>,infinity,undefined,true,{prim_state,false,undefin", "y(4) infinity"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,497}]}, {heap_size,4181}, {total_heap_size,21892}, {links,[#Port<0.2>,<0.0.0>]}, {memory,175920}, {message_queue_len,0}, {reductions,1481481}, {trap_exit,true}]}, {<0.6.0>, [{registered_name,error_logger}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x00002b74132c6c48 (gen_event:fetch_msg/5 + 88)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002aaaaac6b850 Return addr 0x00002b74132f1110 (proc_lib:init_p_do_apply/3 + 56)", "y(0) false","y(1) []", "y(2) [{handler,ns_log_mf_h,false,{state,\"/var/opt/membase/1.6.5.4r/logs\",10485760,10,8", "y(3) error_logger","y(4) <0.2.0>", "0x00002aaaaac6b880 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002b74132f1130 (proc_lib:init_p_do_apply/3 + 88)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,20}]}, {heap_size,2584}, {total_heap_size,5168}, {links,[<0.0.0>,<0.25.0>,#Port<0.1102>]}, {memory,42272}, {message_queue_len,0}, {reductions,390552}, {trap_exit,true}]}, {<0.7.0>, [{registered_name,application_controller}, {status,waiting}, {initial_call,{erlang,apply,2}}, {backtrace, ["Program counter: 0x00002aaaaaafedf0 (gen_server:loop/6 + 288)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002aaaaaca56f8 Return addr 0x0000000000846c78 ()", "y(0) []","y(1) infinity", "y(2) application_controller", "y(3) {state,[],[],[],[{inets,<0.513.0>},{mnesia,<0.410.0>},{ns_server,<0.51.0>},{os_mo", "y(4) application_controller", "y(5) <0.2.0>"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,5}]}, {heap_size,4181}, {total_heap_size,21892}, {links, [<0.41.0>,<0.410.0>,<0.513.0>,<0.51.0>,<0.9.0>, <0.32.0>,<0.0.0>]}, {memory,176224}, {message_queue_len,0}, {reductions,36554}, {trap_exit,true}]}, {<0.9.0>, [{registered_name,[]}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x00002aaaaaaf6d08 (application_master:main_loop/2 + 64)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002b7413322dd0 Return addr 0x00002b74132f1110 (proc_lib:init_p_do_apply/3 + 56)", "y(0) []", "y(1) {state,<0.10.0>,{appl_data,kernel,[application_controller,erl_reply,auth,boot_ser", "y(2) <0.7.0>", "0x00002b7413322df0 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002b74132f1130 (proc_lib:init_p_do_apply/3 + 88)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,1}]}, {heap_size,377}, {total_heap_size,754}, {links,[<0.7.0>,<0.10.0>]}, {memory,6920}, {message_queue_len,0}, {reductions,44}, {trap_exit,true}]}, {<0.10.0>, [{registered_name,[]}, {status,waiting}, {initial_call,{application_master,start_it,4}}, {backtrace, ["Program counter: 0x00002aaaaaaf9390 (application_master:loop_it/4 + 80)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002b7413325e10 Return addr 0x0000000000846c78 ()", "y(0) []","y(1) kernel","y(2) <0.11.0>", "y(3) <0.9.0>"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,0}]}, {heap_size,233}, {total_heap_size,233}, {links,[<0.9.0>,<0.11.0>]}, {memory,2648}, {message_queue_len,0}, {reductions,72}, {trap_exit,true}]}, {<0.11.0>, [{registered_name,kernel_sup}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x00002aaaaaafedf0 (gen_server:loop/6 + 288)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002aaaab05c2a0 Return addr 0x00002b74132f1110 (proc_lib:init_p_do_apply/3 + 56)", "y(0) []","y(1) infinity", "y(2) supervisor", "y(3) {state,{local,kernel_sup},one_for_all,[{child,<0.388.0>,net_sup_dynamic,{erl_dist", "y(4) kernel_sup","y(5) <0.10.0>", "0x00002aaaab05c2d8 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002b74132f1130 (proc_lib:init_p_do_apply/3 + 88)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,13}]}, {heap_size,1597}, {total_heap_size,1974}, {links, [<0.22.0>,<0.26.0>,<0.27.0>,<0.388.0>,<0.24.0>, <0.17.0>,<0.20.0>,<0.21.0>,<0.19.0>,<0.12.0>, <0.13.0>,<0.10.0>]}, {memory,17080}, {message_queue_len,0}, {reductions,1614}, {trap_exit,true}]}, {<0.12.0>, [{registered_name,rex}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x00002aaaaaafedf0 (gen_server:loop/6 + 288)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002aaaae0c0b78 Return addr 0x00002b74132f1110 (proc_lib:init_p_do_apply/3 + 56)", "y(0) []","y(1) infinity","y(2) rpc", "y(3) {1,{<0.6877.0>,{<0.6876.0>,{#Ref<0.0.0.150229>,'ns_1@10.1.5.227'}},nil,nil}}", "y(4) rex","y(5) <0.11.0>", "0x00002aaaae0c0bb0 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002b74132f1130 (proc_lib:init_p_do_apply/3 + 88)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,2}]}, {heap_size,317811}, {total_heap_size,832040}, {links,[<0.11.0>]}, {memory,6657312}, {message_queue_len,0}, {reductions,3603}, {trap_exit,true}]}, {<0.13.0>, [{registered_name,global_name_server}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x00002aaaaaafedf0 (gen_server:loop/6 + 288)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002b7413350130 Return addr 0x00002b74132f1110 (proc_lib:init_p_do_apply/3 + 56)", "y(0) []","y(1) infinity","y(2) global", "y(3) {state,true,['ns_1@10.1.5.229'],['ns_1@10.1.5.229'],[],[],'ns_1@10.1.5.227',<0.14", "y(4) global_name_server","y(5) <0.11.0>", "0x00002b7413350168 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002b74132f1130 (proc_lib:init_p_do_apply/3 + 88)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,532}]}, {heap_size,610}, {total_heap_size,987}, {links,[<0.14.0>,<0.16.0>,<0.15.0>,<0.11.0>]}, {memory,9008}, {message_queue_len,0}, {reductions,113404}, {trap_exit,true}]}, {<0.14.0>, [{registered_name,[]}, {status,waiting}, {initial_call,{erlang,apply,2}}, {backtrace, ["Program counter: 0x00002aaaaab85598 (global:loop_the_locker/1 + 1176)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002aaaaacabd80 Return addr 0x00002aaaaab850c0 (global:init_the_locker/1 + 384)", "y(0) {multi,[],[],['ns_1@10.1.5.229'],'ns_1@10.1.5.229',false,false}", "y(1) infinity", "0x00002aaaaacabd98 Return addr 0x0000000000846c78 ()"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,2}]}, {heap_size,377}, {total_heap_size,754}, {links,[<0.13.0>]}, {memory,6776}, {message_queue_len,0}, {reductions,384}, {trap_exit,true}]}, {<0.15.0>, [{registered_name,[]}, {status,waiting}, {initial_call,{erlang,apply,2}}, {backtrace, ["Program counter: 0x00002aaaaab8d6f0 (global:collect_deletions/2 + 152)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002aaaaac6f1b0 Return addr 0x00002aaaaab8d5f0 (global:loop_the_deleter/1 + 72)", "y(0) infinity","y(1) []", "y(2) <0.13.0>", "0x00002aaaaac6f1d0 Return addr 0x0000000000846c78 ()", "y(0) <0.13.0>"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,1}]}, {heap_size,377}, {total_heap_size,754}, {links,[<0.13.0>]}, {memory,6776}, {message_queue_len,0}, {reductions,113}, {trap_exit,false}]}, {<0.16.0>, [{registered_name,[]}, {status,waiting}, {initial_call,{erlang,apply,2}}, {backtrace, ["Program counter: 0x00002aaaaab8d930 (global:loop_the_registrar/0 + 24)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002aaaaacab1c0 Return addr 0x0000000000846c78 ()", "y(0) []"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,3}]}, {heap_size,610}, {total_heap_size,987}, {links,[<0.13.0>]}, {memory,8640}, {message_queue_len,0}, {reductions,364}, {trap_exit,false}]}, {<0.17.0>, [{registered_name,inet_db}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x00002aaaaaafedf0 (gen_server:loop/6 + 288)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002b74133333d8 Return addr 0x00002b74132f1110 (proc_lib:init_p_do_apply/3 + 56)", "y(0) []","y(1) infinity","y(2) inet_db", "y(3) {state,inet_db,inet_cache,inet_hosts_byname,inet_hosts_byaddr,inet_hosts_file_byn", "y(4) inet_db","y(5) <0.11.0>", "0x00002b7413333410 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002b74132f1130 (proc_lib:init_p_do_apply/3 + 88)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,1}]}, {heap_size,987}, {total_heap_size,2584}, {links,[<0.11.0>]}, {memory,21520}, {message_queue_len,0}, {reductions,1467}, {trap_exit,true}]}, {<0.19.0>, [{registered_name,global_group}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x00002aaaaaafedf0 (gen_server:loop/6 + 288)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002b741339bf48 Return addr 0x00002b74132f1110 (proc_lib:init_p_do_apply/3 + 56)", "y(0) []","y(1) infinity", "y(2) global_group", "y(3) {state,no_conf,true,[],[],[],[],[],'nonode@nohost',[],normal,normal}", "y(4) global_group","y(5) <0.11.0>", "0x00002b741339bf80 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002b74132f1130 (proc_lib:init_p_do_apply/3 + 88)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,0}]}, {heap_size,233}, {total_heap_size,233}, {links,[<0.11.0>]}, {memory,2712}, {message_queue_len,0}, {reductions,92}, {trap_exit,true}]}, {<0.20.0>, [{registered_name,file_server_2}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x00002aaaaaafedf0 (gen_server:loop/6 + 288)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002aaaab0d3180 Return addr 0x00002b74132f1110 (proc_lib:init_p_do_apply/3 + 56)", "y(0) []","y(1) infinity", "y(2) file_server","y(3) #Port<0.262>", "y(4) file_server_2","y(5) <0.11.0>", "0x00002aaaab0d31b8 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002b74132f1130 (proc_lib:init_p_do_apply/3 + 88)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,311}]}, {heap_size,6765}, {total_heap_size,17711}, {links,[#Port<0.262>,<0.11.0>]}, {memory,142576}, {message_queue_len,0}, {reductions,714315}, {trap_exit,true}]}, {<0.21.0>, [{registered_name,code_server}, {status,waiting}, {initial_call,{erlang,apply,2}}, {backtrace, ["Program counter: 0x00002aaaaad22fc8 (code_server:loop/1 + 128)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002aaaab0db460 Return addr 0x0000000000846c78 ()", "y(0) {state,<0.11.0>,\"/opt/membase/1.6.5.4r/erlang-13b03/lib/erlang\",[\"/opt/membase/1.", "y(1) <0.11.0>"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,353}]}, {heap_size,4181}, {total_heap_size,21892}, {links,[<0.11.0>]}, {memory,175880}, {message_queue_len,0}, {reductions,318010}, {trap_exit,true}]}, {<0.22.0>, [{registered_name,standard_error_sup}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x00002aaaaaafedf0 (gen_server:loop/6 + 288)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002b7413324af8 Return addr 0x00002b74132f1110 (proc_lib:init_p_do_apply/3 + 56)", "y(0) []","y(1) infinity", "y(2) supervisor_bridge", "y(3) {state,standard_error,<0.23.0>,<0.23.0>,{local,standard_error_sup}}", "y(4) standard_error_sup","y(5) <0.11.0>", "0x00002b7413324b30 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002b74132f1130 (proc_lib:init_p_do_apply/3 + 88)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,0}]}, {heap_size,233}, {total_heap_size,233}, {links,[<0.11.0>,<0.23.0>]}, {memory,2752}, {message_queue_len,0}, {reductions,40}, {trap_exit,true}]}, {<0.23.0>, [{registered_name,standard_error}, {status,waiting}, {initial_call,{standard_error,server,2}}, {backtrace, ["Program counter: 0x00002aaaaad60458 (standard_error:server_loop/1 + 40)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002b7413325280 Return addr 0x0000000000846c78 ()", "y(0) #Port<0.568>"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,0}]}, {heap_size,233}, {total_heap_size,233}, {links,[<0.22.0>,#Port<0.568>]}, {memory,2752}, {message_queue_len,0}, {reductions,7}, {trap_exit,true}]}, {<0.24.0>, [{registered_name,[]}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x00002aaaaaafedf0 (gen_server:loop/6 + 288)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002b74133397b8 Return addr 0x00002b74132f1110 (proc_lib:init_p_do_apply/3 + 56)", "y(0) []","y(1) infinity", "y(2) supervisor_bridge", "y(3) {state,user_sup,<0.25.0>,<0.25.0>,{<0.24.0>,user_sup}}", "y(4) <0.24.0>","y(5) <0.11.0>", "0x00002b74133397f0 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002b74132f1130 (proc_lib:init_p_do_apply/3 + 88)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,0}]}, {heap_size,1597}, {total_heap_size,1597}, {links,[<0.11.0>,<0.25.0>]}, {memory,13664}, {message_queue_len,0}, {reductions,71}, {trap_exit,true}]}, {<0.25.0>, [{registered_name,user}, {status,waiting}, {initial_call,{user,server,2}}, {backtrace, ["Program counter: 0x00002aaaaad77ea0 (user:server_loop/2 + 56)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002aaaab0c5e30 Return addr 0x0000000000846c78 ()", "y(0) []","y(1) []","y(2) {[],[]}", "y(3) #Port<0.587>"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,21}]}, {heap_size,6765}, {total_heap_size,13530}, {links,[<0.6.0>,<0.24.0>,#Port<0.587>]}, {memory,109168}, {message_queue_len,0}, {reductions,44952}, {trap_exit,true}]}, {<0.26.0>, [{registered_name,[]}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x00002aaaaaafedf0 (gen_server:loop/6 + 288)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002b741331f738 Return addr 0x00002b74132f1110 (proc_lib:init_p_do_apply/3 + 56)", "y(0) []","y(1) infinity", "y(2) kernel_config","y(3) []", "y(4) <0.26.0>","y(5) <0.11.0>", "0x00002b741331f770 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002b74132f1130 (proc_lib:init_p_do_apply/3 + 88)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,0}]}, {heap_size,233}, {total_heap_size,233}, {links,[<0.11.0>]}, {memory,2712}, {message_queue_len,0}, {reductions,176}, {trap_exit,true}]}, {<0.27.0>, [{registered_name,kernel_safe_sup}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x00002aaaaaafedf0 (gen_server:loop/6 + 288)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002b7413344528 Return addr 0x00002b74132f1110 (proc_lib:init_p_do_apply/3 + 56)", "y(0) []","y(1) infinity", "y(2) supervisor", "y(3) {state,{local,kernel_safe_sup},one_for_one,[{child,<0.304.0>,inet_gethost_native_", "y(4) kernel_safe_sup","y(5) <0.11.0>", "0x00002b7413344560 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002b74132f1130 (proc_lib:init_p_do_apply/3 + 88)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,4}]}, {heap_size,377}, {total_heap_size,754}, {links, [<0.132.0>,<0.136.0>,<0.304.0>,<0.135.0>,<0.53.0>, <0.131.0>,<0.11.0>]}, {memory,7120}, {message_queue_len,0}, {reductions,442}, {trap_exit,true}]}, {<0.32.0>, [{registered_name,[]}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x00002aaaaaaf6d08 (application_master:main_loop/2 + 64)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002aaaaac72380 Return addr 0x00002b74132f1110 (proc_lib:init_p_do_apply/3 + 56)", "y(0) []", "y(1) {state,<0.33.0>,{appl_data,sasl,[sasl_sup,alarm_handler,overload,release_handler]", "y(2) <0.7.0>", "0x00002aaaaac723a0 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002b74132f1130 (proc_lib:init_p_do_apply/3 + 88)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,1}]}, {heap_size,233}, {total_heap_size,610}, {links,[<0.7.0>,<0.33.0>]}, {memory,5768}, {message_queue_len,0}, {reductions,35}, {trap_exit,true}]}, {<0.33.0>, [{registered_name,[]}, {status,waiting}, {initial_call,{application_master,start_it,4}}, {backtrace, ["Program counter: 0x00002aaaaaaf9390 (application_master:loop_it/4 + 80)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002b7413326560 Return addr 0x0000000000846c78 ()", "y(0) {state,tty,undefined}","y(1) sasl", "y(2) <0.34.0>","y(3) <0.32.0>"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,0}]}, {heap_size,233}, {total_heap_size,233}, {links,[<0.32.0>,<0.34.0>]}, {memory,2648}, {message_queue_len,0}, {reductions,70}, {trap_exit,true}]}, {<0.34.0>, [{registered_name,sasl_sup}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x00002aaaaaafedf0 (gen_server:loop/6 + 288)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002b741334bc20 Return addr 0x00002b74132f1110 (proc_lib:init_p_do_apply/3 + 56)", "y(0) []","y(1) infinity", "y(2) supervisor", "y(3) {state,{local,sasl_sup},one_for_one,[{child,<0.38.0>,release_handler,{release_han", "y(4) sasl_sup","y(5) <0.33.0>", "0x00002b741334bc58 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002b74132f1130 (proc_lib:init_p_do_apply/3 + 88)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,4}]}, {heap_size,610}, {total_heap_size,1220}, {links,[<0.35.0>,<0.38.0>,<0.33.0>]}, {memory,10688}, {message_queue_len,0}, {reductions,495}, {trap_exit,true}]}, {<0.35.0>, [{registered_name,sasl_safe_sup}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x00002aaaaaafedf0 (gen_server:loop/6 + 288)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002b741339e588 Return addr 0x00002b74132f1110 (proc_lib:init_p_do_apply/3 + 56)", "y(0) []","y(1) infinity", "y(2) supervisor", "y(3) {state,{local,sasl_safe_sup},one_for_one,[{child,<0.37.0>,overload,{overload,star", "y(4) sasl_safe_sup","y(5) <0.34.0>", "0x00002b741339e5c0 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002b74132f1130 (proc_lib:init_p_do_apply/3 + 88)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,1}]}, {heap_size,233}, {total_heap_size,610}, {links,[<0.36.0>,<0.37.0>,<0.34.0>]}, {memory,5808}, {message_queue_len,0}, {reductions,174}, {trap_exit,true}]}, {<0.36.0>, [{registered_name,alarm_handler}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x00002b74132c6c48 (gen_event:fetch_msg/5 + 88)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002b7413327b38 Return addr 0x00002b74132f1110 (proc_lib:init_p_do_apply/3 + 56)", "y(0) false","y(1) []", "y(2) [{handler,alarm_handler,false,[],false}]", "y(3) alarm_handler","y(4) <0.35.0>", "0x00002b7413327b68 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002b74132f1130 (proc_lib:init_p_do_apply/3 + 88)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,0}]}, {heap_size,233}, {total_heap_size,233}, {links,[<0.35.0>]}, {memory,2712}, {message_queue_len,0}, {reductions,28}, {trap_exit,true}]}, {<0.37.0>, [{registered_name,overload}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x00002aaaaaafedf0 (gen_server:loop/6 + 288)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002b741335f348 Return addr 0x00002b74132f1110 (proc_lib:init_p_do_apply/3 + 56)", "y(0) []","y(1) infinity", "y(2) overload", "y(3) {state,0,0,8.000000e-01,226,1.000000e-01,{0,0},clear}", "y(4) overload","y(5) <0.35.0>", "0x00002b741335f380 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002b74132f1130 (proc_lib:init_p_do_apply/3 + 88)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,0}]}, {heap_size,233}, {total_heap_size,233}, {links,[<0.35.0>]}, {memory,2712}, {message_queue_len,0}, {reductions,39}, {trap_exit,false}]}, {<0.38.0>, [{registered_name,release_handler}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x00002aaaaaafedf0 (gen_server:loop/6 + 288)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002b7413387948 Return addr 0x00002b74132f1110 (proc_lib:init_p_do_apply/3 + 56)", "y(0) []","y(1) infinity", "y(2) release_handler", "y(3) {state,[],\"/opt/membase/1.6.5.4r/erlang-13b03/lib/erlang\",\"/opt/membase/1.6.5.4r/", "y(4) release_handler","y(5) <0.34.0>", "0x00002b7413387980 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002b74132f1130 (proc_lib:init_p_do_apply/3 + 88)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,2}]}, {heap_size,2584}, {total_heap_size,5168}, {links,[<0.34.0>]}, {memory,42192}, {message_queue_len,0}, {reductions,2414}, {trap_exit,false}]}, {<0.41.0>, [{registered_name,[]}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x00002aaaaaaf6d08 (application_master:main_loop/2 + 64)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002b7413339f20 Return addr 0x00002b74132f1110 (proc_lib:init_p_do_apply/3 + 56)", "y(0) []", "y(1) {state,<0.42.0>,{appl_data,os_mon,[os_mon_sup,os_mon_sysinfo,disksup,memsup,cpu_s", "y(2) <0.7.0>", "0x00002b7413339f40 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002b74132f1130 (proc_lib:init_p_do_apply/3 + 88)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,0}]}, {heap_size,233}, {total_heap_size,233}, {links,[<0.7.0>,<0.42.0>]}, {memory,2752}, {message_queue_len,0}, {reductions,23}, {trap_exit,true}]}, {<0.42.0>, [{registered_name,[]}, {status,waiting}, {initial_call,{application_master,start_it,4}}, {backtrace, ["Program counter: 0x00002aaaaaaf9390 (application_master:loop_it/4 + 80)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002b741339c6b8 Return addr 0x0000000000846c78 ()", "y(0) []","y(1) os_mon","y(2) <0.43.0>", "y(3) <0.41.0>"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,0}]}, {heap_size,233}, {total_heap_size,233}, {links,[<0.41.0>,<0.43.0>]}, {memory,2648}, {message_queue_len,0}, {reductions,40}, {trap_exit,true}]}, {<0.43.0>, [{registered_name,os_mon_sup}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x00002aaaaaafedf0 (gen_server:loop/6 + 288)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002b74133304c8 Return addr 0x00002b74132f1110 (proc_lib:init_p_do_apply/3 + 56)", "y(0) []","y(1) infinity", "y(2) supervisor", "y(3) {state,{local,os_mon_sup},one_for_one,[{child,<0.47.0>,cpu_sup,{cpu_sup,start_lin", "y(4) os_mon_sup","y(5) <0.42.0>", "0x00002b7413330500 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002b74132f1130 (proc_lib:init_p_do_apply/3 + 88)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,1}]}, {heap_size,377}, {total_heap_size,754}, {links,[<0.44.0>,<0.45.0>,<0.47.0>,<0.42.0>]}, {memory,7000}, {message_queue_len,0}, {reductions,273}, {trap_exit,true}]}, {<0.44.0>, [{registered_name,disksup}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x00002aaaaaafedf0 (gen_server:loop/6 + 288)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002aaaaac99930 Return addr 0x00002b74132f1110 (proc_lib:init_p_do_apply/3 + 56)", "y(0) []","y(1) infinity","y(2) disksup", "y(3) {state,80,60000,{unix,linux},[{\"/\",11204616,49},{\"/boot\",101086,14},{\"/dev/shm\",1", "y(4) disksup","y(5) <0.43.0>", "0x00002aaaaac99968 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002b74132f1130 (proc_lib:init_p_do_apply/3 + 88)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,42}]}, {heap_size,2584}, {total_heap_size,5168}, {links,[<0.43.0>,#Port<0.988>]}, {memory,42232}, {message_queue_len,0}, {reductions,35621}, {trap_exit,true}]}, {<0.45.0>, [{registered_name,memsup}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x00002aaaaaafedf0 (gen_server:loop/6 + 288)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002aaaab0e8068 Return addr 0x00002b74132f1110 (proc_lib:init_p_do_apply/3 + 56)", "y(0) []","y(1) infinity","y(2) memsup", "y(3) {state,{unix,linux},true,{680501248,2058731520},{<0.12.0>,6657168},false,60000,30", "y(4) memsup","y(5) <0.43.0>", "0x00002aaaab0e80a0 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002b74132f1130 (proc_lib:init_p_do_apply/3 + 88)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,54}]}, {heap_size,987}, {total_heap_size,1597}, {links,[<0.43.0>,<0.46.0>]}, {memory,13664}, {message_queue_len,0}, {reductions,59079}, {trap_exit,true}]}, {<0.46.0>, [{registered_name,[]}, {status,waiting}, {initial_call,{erlang,apply,2}}, {backtrace, ["Program counter: 0x00002aaaaaf49880 (memsup:port_idle/1 + 40)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002b7413396bd0 Return addr 0x0000000000846c78 ()", "y(0) []","y(1) #Port<0.1020>"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,43}]}, {heap_size,2584}, {total_heap_size,2961}, {links,[<0.45.0>,#Port<0.1020>]}, {memory,24472}, {message_queue_len,0}, {reductions,16154}, {trap_exit,true}]}, {<0.47.0>, [{registered_name,cpu_sup}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x00002aaaaaafedf0 (gen_server:loop/6 + 288)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002b741335ae60 Return addr 0x00002b74132f1110 (proc_lib:init_p_do_apply/3 + 56)", "y(0) []","y(1) infinity","y(2) cpu_sup", "y(3) {state,<0.48.0>,{unix,linux}}", "y(4) cpu_sup","y(5) <0.43.0>", "0x00002b741335ae98 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002b74132f1130 (proc_lib:init_p_do_apply/3 + 88)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,0}]}, {heap_size,233}, {total_heap_size,233}, {links,[<0.43.0>]}, {memory,2712}, {message_queue_len,0}, {reductions,33}, {trap_exit,true}]}, {<0.48.0>, [{registered_name,[]}, {status,waiting}, {initial_call,{erlang,apply,2}}, {backtrace, ["Program counter: 0x00002aaaaaf3a558 (cpu_sup:measurement_server_loop/1 + 40)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002b7413330c10 Return addr 0x0000000000846c78 ()", "y(0) []","y(1) []","y(2) []", "y(3) []","y(4) []","y(5) []", "y(6) []","y(7) []", "y(8) {internal,<0.49.0>,[],{unix,linux}}"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,0}]}, {heap_size,233}, {total_heap_size,233}, {links,[<0.49.0>]}, {memory,2608}, {message_queue_len,0}, {reductions,11}, {trap_exit,true}]}, {<0.49.0>, [{registered_name,[]}, {status,waiting}, {initial_call,{erlang,apply,2}}, {backtrace, ["Program counter: 0x00002aaaaaf3b730 (cpu_sup:port_server_loop/2 + 64)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002b7413345868 Return addr 0x0000000000846c78 ()", "y(0) []","y(1) 6000", "y(2) #Port<0.1052>"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,3}]}, {heap_size,610}, {total_heap_size,987}, {links,[<0.48.0>,#Port<0.1052>]}, {memory,8680}, {message_queue_len,0}, {reductions,335}, {trap_exit,false}]}, {<0.51.0>, [{registered_name,[]}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x00002aaaaaaf6d08 (application_master:main_loop/2 + 64)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002aaaaac570b8 Return addr 0x00002b74132f1110 (proc_lib:init_p_do_apply/3 + 56)", "y(0) []", "y(1) {state,<0.52.0>,{appl_data,ns_server,[ns_server_sup,ns_config,ns_config_sup,ns_co", "y(2) <0.7.0>", "0x00002aaaaac570d8 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002b74132f1130 (proc_lib:init_p_do_apply/3 + 88)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,0}]}, {heap_size,233}, {total_heap_size,233}, {links,[<0.7.0>,<0.52.0>]}, {memory,2752}, {message_queue_len,0}, {reductions,27}, {trap_exit,true}]}, {<0.52.0>, [{registered_name,[]}, {status,waiting}, {initial_call,{application_master,start_it,4}}, {backtrace, ["Program counter: 0x00002aaaaaaf9390 (application_master:loop_it/4 + 80)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002b7413331388 Return addr 0x0000000000846c78 ()", "y(0) []","y(1) ns_server", "y(2) <0.54.0>","y(3) <0.51.0>"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,0}]}, {heap_size,233}, {total_heap_size,233}, {links,[<0.51.0>,<0.54.0>]}, {memory,2648}, {message_queue_len,0}, {reductions,50}, {trap_exit,true}]}, {<0.53.0>, [{registered_name,timer_server}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x00002aaaaaafedf0 (gen_server:loop/6 + 288)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002aaaab092f38 Return addr 0x00002b74132f1110 (proc_lib:init_p_do_apply/3 + 56)", "y(0) []","y(1) 273","y(2) timer", "y(3) []","y(4) timer_server", "y(5) <0.27.0>", "0x00002aaaab092f70 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002b74132f1130 (proc_lib:init_p_do_apply/3 + 88)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,3782}]}, {heap_size,2584}, {total_heap_size,2961}, {links, [<0.109.0>,<0.460.0>,<0.473.0>,<0.474.0>,<0.470.0>, <0.349.0>,<0.420.0>,<0.303.0>,<0.71.0>,<0.90.0>, <0.78.0>,<0.27.0>]}, {memory,24976}, {message_queue_len,0}, {reductions,2529362}, {trap_exit,true}]}, {<0.54.0>, [{registered_name,ns_server_cluster_sup}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x00002aaaaaafedf0 (gen_server:loop/6 + 288)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002aaaaabb4290 Return addr 0x00002b74132f1110 (proc_lib:init_p_do_apply/3 + 56)", "y(0) []","y(1) infinity", "y(2) supervisor", "y(3) {state,{local,ns_server_cluster_sup},one_for_one,[{child,<0.70.0>,ns_server_sup,{", "y(4) ns_server_cluster_sup","y(5) <0.52.0>", "0x00002aaaaabb42c8 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002b74132f1130 (proc_lib:init_p_do_apply/3 + 88)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,0}]}, {heap_size,2584}, {total_heap_size,2584}, {links,[<0.55.0>,<0.63.0>,<0.70.0>,<0.62.0>,<0.52.0>]}, {memory,21680}, {message_queue_len,0}, {reductions,1840}, {trap_exit,true}]}, {<0.55.0>, [{registered_name,dist_manager}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x00002aaaaaafedf0 (gen_server:loop/6 + 288)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002aaaaac6c6e8 Return addr 0x00002b74132f1110 (proc_lib:init_p_do_apply/3 + 56)", "y(0) []","y(1) infinity", "y(2) dist_manager", "y(3) {state,true,\"10.1.5.227\"}", "y(4) dist_manager","y(5) <0.54.0>", "0x00002aaaaac6c720 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002b74132f1130 (proc_lib:init_p_do_apply/3 + 88)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,6}]}, {heap_size,377}, {total_heap_size,754}, {links,[<0.54.0>]}, {memory,6880}, {message_queue_len,0}, {reductions,601}, {trap_exit,false}]}, {<0.62.0>, [{registered_name,ns_cluster}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x00002aaaaaafedf0 (gen_server:loop/6 + 288)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002aaaaabc11a0 Return addr 0x00002b74132f1110 (proc_lib:init_p_do_apply/3 + 56)", "y(0) []","y(1) infinity", "y(2) ns_cluster","y(3) {state}", "y(4) ns_cluster","y(5) <0.54.0>", "0x00002aaaaabc11d8 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002b74132f1130 (proc_lib:init_p_do_apply/3 + 88)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,8}]}, {heap_size,4181}, {total_heap_size,32838}, {links,[<0.54.0>]}, {memory,263552}, {message_queue_len,0}, {reductions,13561}, {trap_exit,false}]}, {<0.63.0>, [{registered_name,ns_config_sup}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x00002aaaaaafedf0 (gen_server:loop/6 + 288)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002aaaaabe7178 Return addr 0x00002b74132f1110 (proc_lib:init_p_do_apply/3 + 56)", "y(0) []","y(1) infinity", "y(2) supervisor", "y(3) {state,{local,ns_config_sup},rest_for_one,[{child,undefined,ns_config_log,{ns_con", "y(4) ns_config_sup","y(5) <0.54.0>", "0x00002aaaaabe71b0 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002b74132f1130 (proc_lib:init_p_do_apply/3 + 88)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,0}]}, {heap_size,6765}, {total_heap_size,6765}, {links,[<0.64.0>,<0.65.0>,<0.54.0>]}, {memory,55048}, {message_queue_len,0}, {reductions,582}, {trap_exit,true}]}, {<0.64.0>, [{registered_name,ns_config_events}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x00002b74132c6c48 (gen_event:fetch_msg/5 + 88)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002aaaab065cf0 Return addr 0x00002b74132f11f0 (proc_lib:wake_up/3 + 120)", "y(0) false","y(1) []", "y(2) [{handler,ns_pubsub,#Ref<0.0.0.6805>,{state,#Fun,ok},<0.4", "y(3) ns_config_events","y(4) <0.63.0>", "0x00002aaaab065d20 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002b74132f1210 (proc_lib:wake_up/3 + 152)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,9}]}, {heap_size,2584}, {total_heap_size,20295}, {links,[<0.108.0>,<0.468.0>,<0.472.0>,<0.63.0>]}, {memory,163544}, {message_queue_len,0}, {reductions,143460}, {trap_exit,true}]}, {<0.65.0>, [{registered_name,ns_config}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x00002aaaaaafedf0 (gen_server:loop/6 + 288)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002aaaabf0d870 Return addr 0x00002b74132f1110 (proc_lib:init_p_do_apply/3 + 56)", "y(0) []","y(1) infinity", "y(2) ns_config", "y(3) {config,{full,\"/etc/opt/membase/1.6.5.4r/config\",undefined,ns_config_default},[[{", "y(4) ns_config","y(5) <0.63.0>", "0x00002aaaabf0d8a8 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002b74132f1130 (proc_lib:init_p_do_apply/3 + 88)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,25}]}, {heap_size,46368}, {total_heap_size,167761}, {links,[<0.63.0>]}, {memory,1342936}, {message_queue_len,0}, {reductions,228569}, {trap_exit,true}]}, {<0.70.0>, [{registered_name,ns_server_sup}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x00002aaaaaafedf0 (gen_server:loop/6 + 288)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002aaaab416168 Return addr 0x00002b74132f1110 (proc_lib:init_p_do_apply/3 + 56)", "y(0) []","y(1) infinity", "y(2) supervisor", "y(3) {state,{local,ns_server_sup},one_for_one,[{child,<0.473.0>,ns_tick,{ns_tick,start", "y(4) ns_server_sup","y(5) <0.54.0>", "0x00002aaaab4161a0 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002b74132f1130 (proc_lib:init_p_do_apply/3 + 88)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,0}]}, {heap_size,121393}, {total_heap_size,121393}, {links, [<0.94.0>,<0.467.0>,<0.472.0>,<0.473.0>,<0.474.0>, <0.468.0>,<0.107.0>,<0.109.0>,<0.408.0>,<0.108.0>, <0.105.0>,<0.106.0>,<0.99.0>,<0.73.0>,<0.90.0>, <0.76.0>,<0.71.0>,<0.72.0>,<0.54.0>]}, {memory,972712}, {message_queue_len,0}, {reductions,107693}, {trap_exit,true}]}, {<0.71.0>, [{registered_name,ns_log}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x00002aaaaaafedf0 (gen_server:loop/6 + 288)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002aaaaafbb348 Return addr 0x00002b74132f1110 (proc_lib:init_p_do_apply/3 + 56)", "y(0) []","y(1) infinity","y(2) ns_log", "y(3) {state,[{log_entry,{1305,60608,183912},'ns_1@127.0.0.1',ns_node_disco,3,\"Initial ", "y(4) ns_log","y(5) <0.70.0>", "0x00002aaaaafbb380 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002b74132f1130 (proc_lib:init_p_do_apply/3 + 88)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,3}]}, {heap_size,6765}, {total_heap_size,17711}, {links,[<0.53.0>,<0.70.0>]}, {memory,142576}, {message_queue_len,0}, {reductions,3735}, {trap_exit,false}]}, {<0.72.0>, [{registered_name,ns_log_events}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x0000000000846c70 (unknown function)", "CP: 0x0000000000846c78 ()", "arity = 3"," proc_lib"," wake_up", " [gen_event,wake_hib,[<0.70.0>,ns_log_events,[{handler,ns_mail_log,false,{state},false}]"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,0}]}, {heap_size,34}, {total_heap_size,34}, {links,[<0.70.0>]}, {memory,1120}, {message_queue_len,0}, {reductions,381}, {trap_exit,true}]}, {<0.73.0>, [{registered_name,ns_mail_sup}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x00002aaaaaafedf0 (gen_server:loop/6 + 288)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002aaaaabb8ef0 Return addr 0x00002b74132f1110 (proc_lib:init_p_do_apply/3 + 56)", "y(0) []","y(1) infinity", "y(2) supervisor", "y(3) {state,{local,ns_mail_sup},one_for_all,[{child,undefined,ns_mail_log,{ns_mail_log", "y(4) ns_mail_sup","y(5) <0.70.0>", "0x00002aaaaabb8f28 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002b74132f1130 (proc_lib:init_p_do_apply/3 + 88)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,1}]}, {heap_size,233}, {total_heap_size,610}, {links,[<0.74.0>,<0.70.0>]}, {memory,5768}, {message_queue_len,0}, {reductions,661}, {trap_exit,true}]}, {<0.74.0>, [{registered_name,ns_mail}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x00002aaaaaafedf0 (gen_server:loop/6 + 288)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002aaaaabe8498 Return addr 0x00002b74132f1110 (proc_lib:init_p_do_apply/3 + 56)", "y(0) []","y(1) infinity","y(2) ns_mail", "y(3) empty_state","y(4) ns_mail", "y(5) <0.73.0>", "0x00002aaaaabe84d0 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002b74132f1130 (proc_lib:init_p_do_apply/3 + 88)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,0}]}, {heap_size,233}, {total_heap_size,233}, {links,[<0.73.0>]}, {memory,2712}, {message_queue_len,0}, {reductions,27}, {trap_exit,true}]}, {<0.76.0>, [{registered_name,ns_node_disco_sup}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x00002aaaaaafedf0 (gen_server:loop/6 + 288)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002aaaaac56950 Return addr 0x00002b74132f1110 (proc_lib:init_p_do_apply/3 + 56)", "y(0) []","y(1) infinity", "y(2) supervisor", "y(3) {state,{local,ns_node_disco_sup},rest_for_one,[{child,<0.85.0>,ns_config_rep,{ns_", "y(4) ns_node_disco_sup","y(5) <0.70.0>", "0x00002aaaaac56988 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002b74132f1130 (proc_lib:init_p_do_apply/3 + 88)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,3}]}, {heap_size,377}, {total_heap_size,754}, {links,[<0.77.0>,<0.78.0>,<0.85.0>,<0.70.0>]}, {memory,7000}, {message_queue_len,0}, {reductions,832}, {trap_exit,true}]}, {<0.77.0>, [{registered_name,ns_node_disco_events}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x00002b74132c6c48 (gen_event:fetch_msg/5 + 88)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002b7413328fa8 Return addr 0x00002b74132f11f0 (proc_lib:wake_up/3 + 120)", "y(0) false","y(1) []", "y(2) [{handler,menelaus_event,ns_node_disco_events,{state,ns_node_disco_events,undefin", "y(3) ns_node_disco_events","y(4) <0.76.0>", "0x00002b7413328fd8 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002b74132f1210 (proc_lib:wake_up/3 + 152)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,101}]}, {heap_size,233}, {total_heap_size,377}, {links,[<0.76.0>]}, {memory,4080}, {message_queue_len,0}, {reductions,4882}, {trap_exit,true}]}, {<0.78.0>, [{registered_name,ns_node_disco}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x00002aaaaaafedf0 (gen_server:loop/6 + 288)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002aaaabf301f0 Return addr 0x00002b74132f1110 (proc_lib:init_p_do_apply/3 + 56)", "y(0) []","y(1) infinity", "y(2) ns_node_disco", "y(3) {state,['ns_1@10.1.5.227','ns_1@10.1.5.229'],{interval,#Ref<0.0.0.149>}}", "y(4) ns_node_disco","y(5) <0.76.0>", "0x00002aaaabf30228 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002b74132f1130 (proc_lib:init_p_do_apply/3 + 88)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,1181}]}, {heap_size,17711}, {total_heap_size,46368}, {links,[<0.76.0>,<0.53.0>]}, {memory,371832}, {message_queue_len,0}, {reductions,39052}, {trap_exit,false}]}, {<0.85.0>, [{registered_name,ns_config_rep}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x00002aaaaaafedf0 (gen_server:loop/6 + 288)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002aaaad4028e0 Return addr 0x00002b74132f1110 (proc_lib:init_p_do_apply/3 + 56)", "y(0) []","y(1) infinity", "y(2) ns_config_rep","y(3) {state}", "y(4) ns_config_rep","y(5) <0.76.0>", "0x00002aaaad402918 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002b74132f1130 (proc_lib:init_p_do_apply/3 + 88)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,15}]}, {heap_size,46368}, {total_heap_size,64079}, {links,[<0.76.0>]}, {memory,513480}, {message_queue_len,0}, {reductions,22132}, {trap_exit,false}]}, {<0.90.0>, [{registered_name,ns_heart}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x00002aaaaaafedf0 (gen_server:loop/6 + 288)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002aaaaaff83b8 Return addr 0x00002b74132f1110 (proc_lib:init_p_do_apply/3 + 56)", "y(0) []","y(1) infinity", "y(2) ns_heart", "y(3) [{meminfo,<<777 bytes>>},{replication,[{\"default\",1.000000e+00}]},{system_memory_", "y(4) ns_heart","y(5) <0.70.0>", "0x00002aaaaaff83f0 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002b74132f1130 (proc_lib:init_p_do_apply/3 + 88)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,2}]}, {heap_size,4181}, {total_heap_size,32838}, {links,[<0.70.0>,<0.53.0>]}, {memory,263592}, {message_queue_len,0}, {reductions,2216836}, {trap_exit,false}]}, {<0.94.0>, [{registered_name,menelaus_sup}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x00002aaaaaafedf0 (gen_server:loop/6 + 288)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002aaaab0385f8 Return addr 0x00002b74132f1110 (proc_lib:init_p_do_apply/3 + 56)", "y(0) []","y(1) infinity", "y(2) supervisor", "y(3) {state,{local,menelaus_sup},one_for_one,[{child,<0.349.0>,hot_keys_keeper,{hot_ke", "y(4) menelaus_sup","y(5) <0.70.0>", "0x00002aaaab038630 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002b74132f1130 (proc_lib:init_p_do_apply/3 + 88)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,1}]}, {heap_size,6765}, {total_heap_size,17711}, {links,[<0.95.0>,<0.349.0>,<0.70.0>]}, {memory,142616}, {message_queue_len,0}, {reductions,1936}, {trap_exit,true}]}, {<0.95.0>, [{registered_name,menelaus_web}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x00002aaaaaafedf0 (gen_server:loop/6 + 288)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002aaaab076260 Return addr 0x00002b74132f1110 (proc_lib:init_p_do_apply/3 + 56)", "y(0) []","y(1) infinity", "y(2) mochiweb_socket_server", "y(3) {mochiweb_socket_server,8091,#Fun,{local,menelaus_web}", "y(4) menelaus_web","y(5) <0.94.0>", "0x00002aaaab076298 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002b74132f1130 (proc_lib:init_p_do_apply/3 + 88)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,6}]}, {heap_size,1597}, {total_heap_size,2207}, {links, [<0.360.0>,<0.6480.0>,<0.6680.0>,<0.6806.0>, <0.2148.0>,<0.94.0>,<0.160.0>,#Port<0.1399>]}, {memory,18784}, {message_queue_len,0}, {reductions,1625}, {trap_exit,true}]}, {<0.99.0>, [{registered_name,ns_port_sup}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x00002aaaaaafedf0 (gen_server:loop/6 + 288)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002aaaaafc86b8 Return addr 0x00002b74132f1110 (proc_lib:init_p_do_apply/3 + 56)", "y(0) []","y(1) infinity", "y(2) supervisor", "y(3) {state,{local,ns_port_sup},one_for_one,[{child,<0.158.0>,{moxi,\"./bin/moxi/moxi\",", "y(4) ns_port_sup","y(5) <0.70.0>", "0x00002aaaaafc86f0 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002b74132f1130 (proc_lib:init_p_do_apply/3 + 88)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,7}]}, {heap_size,6765}, {total_heap_size,17711}, {links,[<0.103.0>,<0.158.0>,<0.70.0>]}, {memory,142616}, {message_queue_len,0}, {reductions,4100}, {trap_exit,true}]}, {<0.103.0>, [{registered_name,[]}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x00002aaaaaafedf0 (gen_server:loop/6 + 288)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002aaaaac74450 Return addr 0x00002b74132f1110 (proc_lib:init_p_do_apply/3 + 56)", "y(0) []","y(1) infinity", "y(2) supervisor_cushion", "y(3) {state,memcached,5000,{1305,60608,515093},<0.104.0>}", "y(4) <0.103.0>","y(5) <0.99.0>", "0x00002aaaaac74488 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002b74132f1130 (proc_lib:init_p_do_apply/3 + 88)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,0}]}, {heap_size,987}, {total_heap_size,987}, {links,[<0.99.0>,<0.104.0>]}, {memory,8784}, {message_queue_len,0}, {reductions,55}, {trap_exit,true}]}, {<0.104.0>, [{registered_name,[]}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x00002aaaaaafedf0 (gen_server:loop/6 + 288)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002aaaab015548 Return addr 0x00002b74132f1110 (proc_lib:init_p_do_apply/3 + 56)", "y(0) []","y(1) infinity", "y(2) ns_port_server", "y(3) {state,#Port<0.1445>,memcached,{[\"Vbucket ","y(5) <0.103.0>", "0x00002aaaab015580 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002b74132f1130 (proc_lib:init_p_do_apply/3 + 88)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,28}]}, {heap_size,4181}, {total_heap_size,8362}, {links,[<0.103.0>,#Port<0.1445>]}, {memory,67784}, {message_queue_len,0}, {reductions,28783}, {trap_exit,true}]}, {<0.105.0>, [{registered_name,ns_tick_event}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x00002b74132c6c48 (gen_event:fetch_msg/5 + 88)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002aaaab07f808 Return addr 0x00002b74132f1110 (proc_lib:init_p_do_apply/3 + 56)", "y(0) false","y(1) []", "y(2) [{handler,ns_pubsub,#Ref<0.0.0.6772>,{state,#Fun,ignored},<", "y(3) ns_tick_event","y(4) <0.70.0>", "0x00002aaaab07f838 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002b74132f1130 (proc_lib:init_p_do_apply/3 + 88)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,105}]}, {heap_size,377}, {total_heap_size,754}, {links,[<0.70.0>,<0.469.0>]}, {memory,6920}, {message_queue_len,0}, {reductions,15028}, {trap_exit,true}]}, {<0.106.0>, [{registered_name,ns_stats_event}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x00002b74132c6c48 (gen_event:fetch_msg/5 + 88)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002aaaabba8e20 Return addr 0x00002b74132f1110 (proc_lib:init_p_do_apply/3 + 56)", "y(0) false","y(1) []", "y(2) [{handler,ns_pubsub,#Ref<0.0.0.6802>,{state,#Fun,ignored},<", "y(3) ns_stats_event","y(4) <0.70.0>", "0x00002aaaabba8e50 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002b74132f1130 (proc_lib:init_p_do_apply/3 + 88)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,2550}]}, {heap_size,377}, {total_heap_size,987}, {links,[<0.70.0>,<0.470.0>]}, {memory,8784}, {message_queue_len,0}, {reductions,28181}, {trap_exit,true}]}, {<0.107.0>, [{registered_name,ns_good_bucket_worker}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x00002aaaaaafedf0 (gen_server:loop/6 + 288)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002aaaaac77398 Return addr 0x00002b74132f1110 (proc_lib:init_p_do_apply/3 + 56)", "y(0) []","y(1) infinity", "y(2) work_queue","y(3) []", "y(4) ns_good_bucket_worker","y(5) <0.70.0>", "0x00002aaaaac773d0 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002b74132f1130 (proc_lib:init_p_do_apply/3 + 88)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,4}]}, {heap_size,233}, {total_heap_size,610}, {links,[<0.70.0>]}, {memory,5728}, {message_queue_len,0}, {reductions,238}, {trap_exit,false}]}, {<0.108.0>, [{registered_name,ns_good_bucket_sup}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x00002aaaaaafedf0 (gen_server:loop/6 + 288)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002aaaab08de70 Return addr 0x00002b74132f1110 (proc_lib:init_p_do_apply/3 + 56)", "y(0) []","y(1) infinity", "y(2) supervisor", "y(3) {state,{local,ns_good_bucket_sup},one_for_one,[{child,<0.303.0>,{ns_memcached,\"de", "y(4) ns_good_bucket_sup","y(5) <0.70.0>", "0x00002aaaab08dea8 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002b74132f1130 (proc_lib:init_p_do_apply/3 + 88)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,0}]}, {heap_size,6765}, {total_heap_size,6765}, {links,[<0.70.0>,<0.302.0>,<0.303.0>,<0.64.0>]}, {memory,55088}, {message_queue_len,0}, {reductions,247}, {trap_exit,true}]}, {<0.109.0>, [{registered_name,[]}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x00002aaaab271f30 (gen_fsm:loop/7 + 312)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002aaaaac94860 Return addr 0x00002b74132f1110 (proc_lib:init_p_do_apply/3 + 56)", "y(0) []","y(1) infinity", "y(2) ns_orchestrator", "y(3) {rebalancing_state,<0.577.0>,{dict,2,16,16,8,80,48,{[],[],[],[],[],[],[],[],[],[]", "y(4) rebalancing","y(5) ns_orchestrator", "y(6) <0.70.0>", "0x00002aaaaac948a0 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002b74132f1130 (proc_lib:init_p_do_apply/3 + 88)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,93}]}, {heap_size,6765}, {total_heap_size,35422}, {links,[<0.70.0>,<9352.387.0>,<0.577.0>,<0.53.0>]}, {memory,284568}, {message_queue_len,0}, {reductions,482933}, {trap_exit,true}]}, {<0.131.0>, [{registered_name,disk_log_sup}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x00002aaaaaafedf0 (gen_server:loop/6 + 288)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002aaaabb920a0 Return addr 0x00002b74132f1110 (proc_lib:init_p_do_apply/3 + 56)", "y(0) []","y(1) infinity", "y(2) supervisor", "y(3) {state,{local,disk_log_sup},simple_one_for_one,[{child,undefined,disk_log,{disk_l", "y(4) disk_log_sup","y(5) <0.27.0>", "0x00002aaaabb920d8 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002b74132f1130 (proc_lib:init_p_do_apply/3 + 88)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,23}]}, {heap_size,1597}, {total_heap_size,2207}, {links,[<0.456.0>,<0.27.0>]}, {memory,18544}, {message_queue_len,0}, {reductions,5206}, {trap_exit,true}]}, {<0.132.0>, [{registered_name,disk_log_server}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x00002aaaaaafedf0 (gen_server:loop/6 + 288)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002b741334ee18 Return addr 0x00002b74132f1110 (proc_lib:init_p_do_apply/3 + 56)", "y(0) []","y(1) infinity", "y(2) disk_log_server","y(3) {state,[]}", "y(4) disk_log_server","y(5) <0.27.0>", "0x00002b741334ee50 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002b74132f1130 (proc_lib:init_p_do_apply/3 + 88)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,41}]}, {heap_size,987}, {total_heap_size,2584}, {links,[<0.456.0>,<0.27.0>]}, {memory,21560}, {message_queue_len,0}, {reductions,5271}, {trap_exit,true}]}, {<0.135.0>, [{registered_name,dets_sup}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x00002aaaaaafedf0 (gen_server:loop/6 + 288)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002aaaaaca62b8 Return addr 0x00002b74132f1110 (proc_lib:init_p_do_apply/3 + 56)", "y(0) []","y(1) infinity", "y(2) supervisor", "y(3) {state,{local,dets_sup},simple_one_for_one,[{child,undefined,dets,{dets,istart_li", "y(4) dets_sup","y(5) <0.27.0>", "0x00002aaaaaca62f0 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002b74132f1130 (proc_lib:init_p_do_apply/3 + 88)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,6}]}, {heap_size,377}, {total_heap_size,987}, {links,[<0.27.0>]}, {memory,8744}, {message_queue_len,0}, {reductions,658}, {trap_exit,true}]}, {<0.136.0>, [{registered_name,dets}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x00002aaaaaafedf0 (gen_server:loop/6 + 288)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002aaaaac9d438 Return addr 0x00002b74132f1110 (proc_lib:init_p_do_apply/3 + 56)", "y(0) []","y(1) infinity", "y(2) dets_server", "y(3) {state,32802,[<0.27.0>],[]}", "y(4) dets","y(5) <0.27.0>", "0x00002aaaaac9d470 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002b74132f1130 (proc_lib:init_p_do_apply/3 + 88)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,12}]}, {heap_size,377}, {total_heap_size,754}, {links,[<0.27.0>]}, {memory,6880}, {message_queue_len,0}, {reductions,1003}, {trap_exit,true}]}, {<0.158.0>, [{registered_name,[]}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x00002aaaaaafedf0 (gen_server:loop/6 + 288)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002aaaaac54a78 Return addr 0x00002b74132f1110 (proc_lib:init_p_do_apply/3 + 56)", "y(0) []","y(1) infinity", "y(2) supervisor_cushion", "y(3) {state,moxi,5000,{1305,60613,122584},<0.159.0>}", "y(4) <0.158.0>","y(5) <0.99.0>", "0x00002aaaaac54ab0 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002b74132f1130 (proc_lib:init_p_do_apply/3 + 88)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,1}]}, {heap_size,987}, {total_heap_size,2584}, {links,[<0.99.0>,<0.159.0>]}, {memory,21560}, {message_queue_len,0}, {reductions,150}, {trap_exit,true}]}, {<0.159.0>, [{registered_name,[]}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x00002aaaaaafedf0 (gen_server:loop/6 + 288)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002aaaaabd67c0 Return addr 0x00002b74132f1110 (proc_lib:init_p_do_apply/3 + 56)", "y(0) []","y(1) infinity", "y(2) ns_port_server", "y(3) {state,#Port<0.2484>,moxi,{[\"\t\t\\\"vBucketMap\\\":\t[[-1, -1], [-1, -1], [-1, -1], [-1", "y(4) <0.159.0>","y(5) <0.158.0>", "0x00002aaaaabd67f8 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002b74132f1130 (proc_lib:init_p_do_apply/3 + 88)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,2}]}, {heap_size,4181}, {total_heap_size,5778}, {links,[<0.158.0>,#Port<0.2484>]}, {memory,47112}, {message_queue_len,0}, {reductions,716}, {trap_exit,true}]}, {<0.160.0>, [{registered_name,[]}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x00002aaaab230ea8 (menelaus_web:handle_streaming/4 + 392)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002aaaac19cd78 Return addr 0x00002aaaab22d558 (menelaus_web:loop/3 + 24272)", "y(0) {struct,[{buckets,[{struct,[{name,<<7 bytes>>},{nodeLocator,vbucket},{saslPasswor", "y(1) {struct,[{buckets,[{struct,[{name,<<7 bytes>>},{nodeLocator,vbucket},{saslPasswor", "y(2) {mochiweb_response,{mochiweb_request,#Port<0.2496>,'GET',\"/pools/default/saslBuck", "y(3) {mochiweb_request,#Port<0.2496>,'GET',\"/pools/default/saslBucketsStreaming\",{1,1}", "y(4) #Fun", "0x00002aaaac19cda8 Return addr 0x00002aaaab246680 (mochiweb_http:headers/5 + 1360)", "y(0) []","y(1) []","y(2) []", "y(3) []", "y(4) {mochiweb_request,#Port<0.2496>,'GET',\"/pools/default/saslBucketsStreaming\",{1,1}", "y(5) Catch 0x00002aaaab22d578 (menelaus_web:loop/3 + 24304)", "0x00002aaaac19cde0 Return addr 0x00002b74132f1110 (proc_lib:init_p_do_apply/3 + 56)", "y(0) #Fun", "y(1) []","y(2) []", "y(3) {mochiweb_request,#Port<0.2496>,'GET',\"/pools/default/saslBucketsStreaming\",{1,1}", "0x00002aaaac19ce08 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002b74132f1130 (proc_lib:init_p_do_apply/3 + 88)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,4}]}, {heap_size,46368}, {total_heap_size,121393}, {links,[<0.95.0>,#Port<0.2496>]}, {memory,972176}, {message_queue_len,0}, {reductions,4084284}, {trap_exit,false}]}, {<0.302.0>, [{registered_name,'ns_vbm_sup-default'}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x00002aaaaaafedf0 (gen_server:loop/6 + 288)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002b741331fe88 Return addr 0x00002b74132f1110 (proc_lib:init_p_do_apply/3 + 56)", "y(0) []","y(1) infinity", "y(2) supervisor", "y(3) {state,{local,'ns_vbm_sup-default'},one_for_one,[],{dict,0,16,16,8,80,48,{[],[],[", "y(4) 'ns_vbm_sup-default'","y(5) <0.108.0>", "0x00002b741331fec0 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002b74132f1130 (proc_lib:init_p_do_apply/3 + 88)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,0}]}, {heap_size,233}, {total_heap_size,233}, {links,[<0.108.0>]}, {memory,2712}, {message_queue_len,0}, {reductions,581}, {trap_exit,true}]}, {<0.303.0>, [{registered_name,'ns_memcached-default'}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x00002aaaaaafedf0 (gen_server:loop/6 + 288)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002aaaabe905f0 Return addr 0x00002b74132f1110 (proc_lib:init_p_do_apply/3 + 56)", "y(0) []","y(1) infinity", "y(2) ns_memcached", "y(3) {state,\"default\",#Port<0.2567>}", "y(4) <0.303.0>","y(5) <0.108.0>", "0x00002aaaabe90628 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002b74132f1130 (proc_lib:init_p_do_apply/3 + 88)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,62}]}, {heap_size,10946}, {total_heap_size,85971}, {links,[<0.53.0>,<0.108.0>,#Port<0.2567>]}, {memory,688696}, {message_queue_len,0}, {reductions,41518416}, {trap_exit,true}]}, {<0.304.0>, [{registered_name,inet_gethost_native_sup}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x00002aaaaaafedf0 (gen_server:loop/6 + 288)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002aaaaabaed60 Return addr 0x00002b74132f1110 (proc_lib:init_p_do_apply/3 + 56)", "y(0) []","y(1) infinity", "y(2) supervisor_bridge", "y(3) {state,inet_gethost_native,<0.305.0>,<0.305.0>,{local,inet_gethost_native_sup}}", "y(4) inet_gethost_native_sup", "y(5) <0.27.0>", "0x00002aaaaabaed98 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002b74132f1130 (proc_lib:init_p_do_apply/3 + 88)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,0}]}, {heap_size,233}, {total_heap_size,233}, {links,[<0.27.0>,<0.305.0>]}, {memory,2752}, {message_queue_len,0}, {reductions,41}, {trap_exit,true}]}, {<0.305.0>, [{registered_name,inet_gethost_native}, {status,waiting}, {initial_call,{inet_gethost_native,server_init,2}}, {backtrace, ["Program counter: 0x00002aaaaacd6440 (inet_gethost_native:main_loop/1 + 40)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002b7413341638 Return addr 0x0000000000846c78 ()", "y(0) {state,#Port<0.2556>,8000,40989,45091,<0.304.0>,4,{statistics,0,0,0,0,0,0,0,0}}"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,5}]}, {heap_size,233}, {total_heap_size,610}, {links,[<0.304.0>,#Port<0.2556>]}, {memory,5768}, {message_queue_len,0}, {reductions,232}, {trap_exit,true}]}, {<0.349.0>, [{registered_name,hot_keys_keeper}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x00002aaaaaafedf0 (gen_server:loop/6 + 288)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002aaaabf68180 Return addr 0x00002b74132f1110 (proc_lib:init_p_do_apply/3 + 56)", "y(0) []","y(1) infinity", "y(2) hot_keys_keeper", "y(3) {state,[{\"default\",[{\"b7aedb5d-898e-4666-9727-1265f77c4d24-10-0\",[{ops,7.923930e-", "y(4) hot_keys_keeper","y(5) <0.94.0>", "0x00002aaaabf681b8 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002b74132f1130 (proc_lib:init_p_do_apply/3 + 88)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,253}]}, {heap_size,28657}, {total_heap_size,75025}, {links,[<0.94.0>,<0.53.0>]}, {memory,601088}, {message_queue_len,0}, {reductions,33503}, {trap_exit,false}]}, {<0.360.0>, [{registered_name,[]}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x00002aaaab22eb10 (menelaus_web:handle_pool_info_wait/6 + 280)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002aaaab95f250 Return addr 0x00002aaaab22e590 (menelaus_web:check_and_handle_pool_info", "y(0) \"43798496\"","y(1) \"43798496\"", "y(2) 20000","y(3) \"10.1.5.227\"", "y(4) {\"Administrator\",\"password\"}", "y(5) \"default\"", "y(6) {mochiweb_request,#Port<0.2676>,'GET',\"/pools/default?waitChange=20000&etag=43798", "0x00002aaaab95f290 Return addr 0x00002aaaab22d4e8 (menelaus_web:loop/3 + 24160)", "y(0) Catch 0x00002aaaab22e5e8 (menelaus_web:check_and_handle_pool_info/2 + 336)", "y(1) []","y(2) <0.6840.0>", "0x00002aaaab95f2b0 Return addr 0x00002aaaab246680 (mochiweb_http:headers/5 + 1360)", "y(0) []","y(1) []","y(2) []", "y(3) []", "y(4) {mochiweb_request,#Port<0.2676>,'GET',\"/pools/default?waitChange=20000&etag=43798", "y(5) Catch 0x00002aaaab22d578 (menelaus_web:loop/3 + 24304)", "0x00002aaaab95f2e8 Return addr 0x00002b74132f1110 (proc_lib:init_p_do_apply/3 + 56)", "y(0) #Fun", "y(1) []","y(2) []", "y(3) {mochiweb_request,#Port<0.2676>,'GET',\"/pools/default?waitChange=20000&etag=43798", "0x00002aaaab95f310 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002b74132f1130 (proc_lib:init_p_do_apply/3 + 88)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,6}]}, {heap_size,17711}, {total_heap_size,46368}, {links,[<0.95.0>,<0.6840.0>,#Port<0.2676>]}, {memory,372016}, {message_queue_len,0}, {reductions,2357489}, {trap_exit,false}]}, {<0.388.0>, [{registered_name,net_sup}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x00002aaaaaafedf0 (gen_server:loop/6 + 288)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002aaaaabe7d48 Return addr 0x00002b74132f1110 (proc_lib:init_p_do_apply/3 + 56)", "y(0) []","y(1) infinity", "y(2) supervisor", "y(3) {state,{local,net_sup},one_for_all,[{child,<0.391.0>,net_kernel,{net_kernel,start", "y(4) net_sup","y(5) <0.11.0>", "0x00002aaaaabe7d80 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002b74132f1130 (proc_lib:init_p_do_apply/3 + 88)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,1}]}, {heap_size,377}, {total_heap_size,987}, {links,[<0.389.0>,<0.390.0>,<0.391.0>,<0.11.0>]}, {memory,8864}, {message_queue_len,0}, {reductions,244}, {trap_exit,true}]}, {<0.389.0>, [{registered_name,erl_epmd}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x00002aaaaaafedf0 (gen_server:loop/6 + 288)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002aaaabbe6ce8 Return addr 0x00002b74132f1110 (proc_lib:init_p_do_apply/3 + 56)", "y(0) []","y(1) infinity", "y(2) erl_epmd", "y(3) {state,#Port<0.2737>,21100,ns_1}", "y(4) erl_epmd","y(5) <0.388.0>", "0x00002aaaabbe6d20 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002b74132f1130 (proc_lib:init_p_do_apply/3 + 88)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,0}]}, {heap_size,233}, {total_heap_size,233}, {links,[<0.388.0>,#Port<0.2737>]}, {memory,2752}, {message_queue_len,0}, {reductions,127}, {trap_exit,false}]}, {<0.390.0>, [{registered_name,auth}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x00002aaaaaafedf0 (gen_server:loop/6 + 288)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002b7413352010 Return addr 0x00002b74132f1110 (proc_lib:init_p_do_apply/3 + 56)", "y(0) []","y(1) infinity","y(2) auth", "y(3) {state,bujipmggocppuqmx,241681}", "y(4) auth","y(5) <0.388.0>", "0x00002b7413352048 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002b74132f1130 (proc_lib:init_p_do_apply/3 + 88)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,1}]}, {heap_size,233}, {total_heap_size,610}, {links,[<0.388.0>]}, {memory,5728}, {message_queue_len,0}, {reductions,154}, {trap_exit,true}]}, {<0.391.0>, [{registered_name,net_kernel}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x00002aaaaaafedf0 (gen_server:loop/6 + 288)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002aaaabba8248 Return addr 0x00002b74132f1110 (proc_lib:init_p_do_apply/3 + 56)", "y(0) []","y(1) infinity", "y(2) net_kernel", "y(3) {state,'ns_1@10.1.5.227','ns_1@10.1.5.227',longnames,{tick,<0.393.0>,15000},7000,", "y(4) net_kernel","y(5) <0.388.0>", "0x00002aaaabba8280 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002b74132f1130 (proc_lib:init_p_do_apply/3 + 88)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,12}]}, {heap_size,987}, {total_heap_size,1364}, {links, [<0.388.0>,<0.393.0>,<0.541.0>,<0.392.0>, #Port<0.2735>]}, {memory,11952}, {message_queue_len,0}, {reductions,1942}, {trap_exit,true}]}, {<0.392.0>, [{registered_name,[]}, {status,waiting}, {initial_call,{inet_tcp_dist,accept_loop,2}}, {backtrace, ["Program counter: 0x00002b74132a8d78 (prim_inet:accept0/2 + 184)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002b741332e0c8 Return addr 0x00002aaaaaf7d900 (inet_tcp:accept/1 + 40)", "y(0) 10690","y(1) #Port<0.2735>", "0x00002b741332e0e0 Return addr 0x00002aaaaaf81d48 (inet_tcp_dist:accept_loop/2 + 96)", "y(0) []", "0x00002b741332e0f0 Return addr 0x0000000000846c78 ()", "y(0) []","y(1) #Port<0.2735>", "y(2) <0.391.0>"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,0}]}, {heap_size,233}, {total_heap_size,233}, {links,[<0.391.0>]}, {memory,2680}, {message_queue_len,0}, {reductions,268}, {trap_exit,false}]}, {<0.393.0>, [{registered_name,[]}, {status,waiting}, {initial_call,{net_kernel,ticker,2}}, {backtrace, ["Program counter: 0x00002aaaaad42130 (net_kernel:ticker_loop/2 + 56)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002aaaaabfe230 Return addr 0x0000000000846c78 ()", "y(0) 15000","y(1) <0.391.0>"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,0}]}, {heap_size,233}, {total_heap_size,233}, {links,[<0.391.0>]}, {memory,2608}, {message_queue_len,0}, {reductions,171}, {trap_exit,false}]}, {<0.408.0>, [{registered_name,ns_mnesia}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x00002aaaaaafedf0 (gen_server:loop/6 + 288)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002aaaab073070 Return addr 0x00002b74132f1110 (proc_lib:init_p_do_apply/3 + 56)", "y(0) []","y(1) infinity", "y(2) ns_mnesia","y(3) {state}", "y(4) ns_mnesia","y(5) <0.70.0>", "0x00002aaaab0730a8 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002b74132f1130 (proc_lib:init_p_do_apply/3 + 88)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,4}]}, {heap_size,1597}, {total_heap_size,1974}, {links,[<0.70.0>,<0.418.0>]}, {memory,16680}, {message_queue_len,0}, {reductions,1148}, {trap_exit,true}]}, {<0.410.0>, [{registered_name,[]}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x00002aaaaaaf6d08 (application_master:main_loop/2 + 64)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002aaaaaca9e88 Return addr 0x00002b74132f1110 (proc_lib:init_p_do_apply/3 + 56)", "y(0) []", "y(1) {state,<0.411.0>,{appl_data,mnesia,[mnesia_dumper_load_regulator,mnesia_event,mne", "y(2) <0.7.0>", "0x00002aaaaaca9ea8 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002b74132f1130 (proc_lib:init_p_do_apply/3 + 88)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,7}]}, {heap_size,233}, {total_heap_size,610}, {links,[<0.7.0>,<0.411.0>]}, {memory,5768}, {message_queue_len,0}, {reductions,49}, {trap_exit,true}]}, {<0.411.0>, [{registered_name,[]}, {status,waiting}, {initial_call,{application_master,start_it,4}}, {backtrace, ["Program counter: 0x00002aaaaaaf9390 (application_master:loop_it/4 + 80)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002aaaab078420 Return addr 0x0000000000846c78 ()", "y(0) {normal,[]}","y(1) mnesia_sup", "y(2) <0.412.0>","y(3) <0.410.0>"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,0}]}, {heap_size,233}, {total_heap_size,233}, {links,[<0.410.0>,<0.412.0>]}, {memory,2648}, {message_queue_len,0}, {reductions,32}, {trap_exit,true}]}, {<0.412.0>, [{registered_name,mnesia_sup}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x00002aaaaaafedf0 (gen_server:loop/6 + 288)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002aaaaac76c48 Return addr 0x00002b74132f1110 (proc_lib:init_p_do_apply/3 + 56)", "y(0) []","y(1) infinity", "y(2) supervisor", "y(3) {state,{local,mnesia_sup},one_for_all,[{child,<0.414.0>,mnesia_kernel_sup,{mnesia", "y(4) mnesia_sup","y(5) <0.411.0>", "0x00002aaaaac76c80 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002b74132f1130 (proc_lib:init_p_do_apply/3 + 88)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,1}]}, {heap_size,377}, {total_heap_size,754}, {links,[<0.413.0>,<0.414.0>,<0.411.0>]}, {memory,6960}, {message_queue_len,0}, {reductions,191}, {trap_exit,true}]}, {<0.413.0>, [{registered_name,mnesia_event}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x00002b74132c6c48 (gen_event:fetch_msg/5 + 88)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002aaaaaffff30 Return addr 0x00002b74132f1110 (proc_lib:init_p_do_apply/3 + 56)", "y(0) false","y(1) []", "y(2) [{handler,mnesia_event,false,{state,[],false,[]},false}]", "y(3) mnesia_event","y(4) <0.412.0>", "0x00002aaaaaffff60 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002b74132f1130 (proc_lib:init_p_do_apply/3 + 88)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,8}]}, {heap_size,377}, {total_heap_size,754}, {links,[<0.412.0>,<0.418.0>]}, {memory,6920}, {message_queue_len,0}, {reductions,448}, {trap_exit,true}]}, {<0.414.0>, [{registered_name,mnesia_kernel_sup}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x00002aaaaaafedf0 (gen_server:loop/6 + 288)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002aaaaabad2f0 Return addr 0x00002b74132f1110 (proc_lib:init_p_do_apply/3 + 56)", "y(0) []","y(1) infinity", "y(2) supervisor", "y(3) {state,{local,mnesia_kernel_sup},one_for_all,[{child,<0.461.0>,mnesia_late_loader", "y(4) mnesia_kernel_sup","y(5) <0.412.0>", "0x00002aaaaabad328 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002b74132f1130 (proc_lib:init_p_do_apply/3 + 88)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,4}]}, {heap_size,610}, {total_heap_size,987}, {links, [<0.419.0>,<0.459.0>,<0.460.0>,<0.461.0>,<0.421.0>, <0.458.0>,<0.420.0>,<0.415.0>,<0.418.0>,<0.412.0>]}, {memory,9104}, {message_queue_len,0}, {reductions,545}, {trap_exit,true}]}, {<0.415.0>, [{registered_name,mnesia_monitor}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x00002aaaaaafedf0 (gen_server:loop/6 + 288)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002b741334a908 Return addr 0x00002b74132f1110 (proc_lib:init_p_do_apply/3 + 56)", "y(0) []","y(1) infinity", "y(2) mnesia_monitor", "y(3) {state,<0.414.0>,[],[],true,[],undefined,[]}", "y(4) mnesia_monitor","y(5) <0.414.0>", "0x00002b741334a940 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002b74132f1130 (proc_lib:init_p_do_apply/3 + 88)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,22}]}, {heap_size,2584}, {total_heap_size,3194}, {links,[<0.456.0>,<0.414.0>]}, {memory,26440}, {message_queue_len,0}, {reductions,5317}, {trap_exit,true}]}, {<0.418.0>, [{registered_name,mnesia_subscr}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x00002aaaaaafedf0 (gen_server:loop/6 + 288)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002aaaaac2fda0 Return addr 0x00002b74132f1110 (proc_lib:init_p_do_apply/3 + 56)", "y(0) []","y(1) infinity", "y(2) mnesia_subscr", "y(3) {state,<0.414.0>,245787}", "y(4) mnesia_subscr","y(5) <0.414.0>", "0x00002aaaaac2fdd8 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002b74132f1130 (proc_lib:init_p_do_apply/3 + 88)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,0}]}, {heap_size,233}, {total_heap_size,233}, {links,[<0.413.0>,<0.414.0>,<0.408.0>]}, {memory,2792}, {message_queue_len,0}, {reductions,111}, {trap_exit,true}]}, {<0.419.0>, [{registered_name,mnesia_locker}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x00002aaaab484c68 (mnesia_locker:loop/1 + 40)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002b7413359b10 Return addr 0x00002aaaab443770 (mnesia_sp:init_proc/4 + 264)", "y(0) []","y(1) []","y(2) []", "y(3) []","y(4) []", "y(5) {state,<0.414.0>}", "0x00002b7413359b48 Return addr 0x00002b74132f1110 (proc_lib:init_p_do_apply/3 + 56)", "y(0) Catch 0x00002aaaab443770 (mnesia_sp:init_proc/4 + 264)", "y(1) mnesia_locker","y(2) []","y(3) []", "y(4) [<0.414.0>]", "0x00002b7413359b78 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002b74132f1130 (proc_lib:init_p_do_apply/3 + 88)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,472}]}, {heap_size,987}, {total_heap_size,1364}, {links,[<0.414.0>]}, {memory,11760}, {message_queue_len,0}, {reductions,189991}, {trap_exit,true}]}, {<0.420.0>, [{registered_name,mnesia_recover}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x00002aaaaaafedf0 (gen_server:loop/6 + 288)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002aaaaaffd480 Return addr 0x00002b74132f1110 (proc_lib:init_p_do_apply/3 + 56)", "y(0) []","y(1) infinity", "y(2) mnesia_recover", "y(3) {state,<0.414.0>,undefined,undefined,undefined,0,true,[]}", "y(4) mnesia_recover","y(5) <0.414.0>", "0x00002aaaaaffd4b8 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002b74132f1130 (proc_lib:init_p_do_apply/3 + 88)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,9}]}, {heap_size,2584}, {total_heap_size,2961}, {links,[<0.414.0>,<0.53.0>]}, {memory,24576}, {message_queue_len,0}, {reductions,3713}, {trap_exit,true}]}, {<0.421.0>, [{registered_name,mnesia_tm}, {status,runnable}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x00002aaaab4d47d0 (mnesia_tm:doit_loop/1 + 216)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002aaaab96c608 Return addr 0x00002aaaab443770 (mnesia_sp:init_proc/4 + 264)", "y(0) []","y(1) []", "y(2) {state,{1,{{tid,3179,<0.470.0>},[13312044],nil,nil}},{0,nil},<0.414.0>,[],[],[]}", "y(3) []","y(4) []","y(5) <0.414.0>", "y(6) {0,nil}", "y(7) {1,{{tid,3179,<0.470.0>},[13312044],nil,nil}}", "0x00002aaaab96c650 Return addr 0x00002b74132f1110 (proc_lib:init_p_do_apply/3 + 56)", "y(0) Catch 0x00002aaaab443770 (mnesia_sp:init_proc/4 + 264)", "y(1) mnesia_tm","y(2) []","y(3) []", "y(4) [<0.414.0>]", "0x00002aaaab96c680 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002b74132f1130 (proc_lib:init_p_do_apply/3 + 88)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,56}]}, {heap_size,6765}, {total_heap_size,24476}, {links,[<0.414.0>]}, {memory,196688}, {message_queue_len,1}, {reductions,453962}, {trap_exit,true}]}, {<0.456.0>, [{registered_name,[]}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x00002aaaab7985d0 (disk_log:loop/1 + 168)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002b741335e078 Return addr 0x00002b74132f1110 (proc_lib:init_p_do_apply/3 + 56)", "y(0) {state,[],[],<0.131.0>,<0.132.0>,17,{arg,latest_log,undefined,\"/var/opt/membase/1", "0x00002b741335e088 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002b74132f1130 (proc_lib:init_p_do_apply/3 + 88)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,6}]}, {heap_size,1597}, {total_heap_size,3194}, {links,[<0.132.0>,<0.415.0>,<0.131.0>,#Port<0.8915>]}, {memory,26520}, {message_queue_len,0}, {reductions,137631}, {trap_exit,true}]}, {<0.458.0>, [{registered_name,mnesia_checkpoint_sup}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x00002aaaaaafedf0 (gen_server:loop/6 + 288)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002b741338c5a8 Return addr 0x00002b74132f1110 (proc_lib:init_p_do_apply/3 + 56)", "y(0) []","y(1) infinity", "y(2) supervisor", "y(3) {state,{local,mnesia_checkpoint_sup},simple_one_for_one,[{child,undefined,mnesia_", "y(4) mnesia_checkpoint_sup", "y(5) <0.414.0>", "0x00002b741338c5e0 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002b74132f1130 (proc_lib:init_p_do_apply/3 + 88)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,0}]}, {heap_size,233}, {total_heap_size,233}, {links,[<0.414.0>]}, {memory,2712}, {message_queue_len,0}, {reductions,61}, {trap_exit,true}]}, {<0.459.0>, [{registered_name,mnesia_snmp_sup}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x00002aaaaaafedf0 (gen_server:loop/6 + 288)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002b74133211a8 Return addr 0x00002b74132f1110 (proc_lib:init_p_do_apply/3 + 56)", "y(0) []","y(1) infinity", "y(2) supervisor", "y(3) {state,{local,mnesia_snmp_sup},simple_one_for_one,[{child,undefined,mnesia_snmp_s", "y(4) mnesia_snmp_sup","y(5) <0.414.0>", "0x00002b74133211e0 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002b74132f1130 (proc_lib:init_p_do_apply/3 + 88)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,0}]}, {heap_size,233}, {total_heap_size,233}, {links,[<0.414.0>]}, {memory,2712}, {message_queue_len,0}, {reductions,61}, {trap_exit,true}]}, {<0.460.0>, [{registered_name,mnesia_controller}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x00002aaaaaafedf0 (gen_server:loop/6 + 288)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002b7413340eb0 Return addr 0x00002b74132f1110 (proc_lib:init_p_do_apply/3 + 56)", "y(0) []","y(1) infinity", "y(2) mnesia_controller", "y(3) {state,<0.414.0>,true,[],[],{0,nil},[],[],{0,nil},undefined,[],[],{interval,#Ref<", "y(4) mnesia_controller","y(5) <0.414.0>", "0x00002b7413340ee8 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002b74132f1130 (proc_lib:init_p_do_apply/3 + 88)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,8}]}, {heap_size,987}, {total_heap_size,1364}, {links,[<0.414.0>,<0.53.0>]}, {memory,11800}, {message_queue_len,0}, {reductions,1712}, {trap_exit,true}]}, {<0.461.0>, [{registered_name,mnesia_late_loader}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x00002aaaab50f5e8 (mnesia_late_loader:loop/1 + 40)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002aaaab094e00 Return addr 0x00002aaaab443770 (mnesia_sp:init_proc/4 + 264)", "y(0) []","y(1) []", "y(2) {state,<0.414.0>}", "0x00002aaaab094e20 Return addr 0x00002b74132f1110 (proc_lib:init_p_do_apply/3 + 56)", "y(0) Catch 0x00002aaaab443770 (mnesia_sp:init_proc/4 + 264)", "y(1) mnesia_late_loader","y(2) []", "y(3) []","y(4) [<0.414.0>]", "0x00002aaaab094e50 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002b74132f1130 (proc_lib:init_p_do_apply/3 + 88)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,2}]}, {heap_size,987}, {total_heap_size,1597}, {links,[<0.414.0>]}, {memory,13624}, {message_queue_len,0}, {reductions,509}, {trap_exit,false}]}, {<0.467.0>, [{registered_name,ns_bad_bucket_worker}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x00002aaaaaafedf0 (gen_server:loop/6 + 288)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002b74133243a8 Return addr 0x00002b74132f1110 (proc_lib:init_p_do_apply/3 + 56)", "y(0) []","y(1) infinity", "y(2) work_queue","y(3) []", "y(4) ns_bad_bucket_worker","y(5) <0.70.0>", "0x00002b74133243e0 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002b74132f1130 (proc_lib:init_p_do_apply/3 + 88)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,1}]}, {heap_size,233}, {total_heap_size,610}, {links,[<0.70.0>]}, {memory,5728}, {message_queue_len,0}, {reductions,61}, {trap_exit,false}]}, {<0.468.0>, [{registered_name,ns_bad_bucket_sup}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x00002aaaaaafedf0 (gen_server:loop/6 + 288)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002aaaabfc2a88 Return addr 0x00002b74132f1110 (proc_lib:init_p_do_apply/3 + 56)", "y(0) []","y(1) infinity", "y(2) supervisor", "y(3) {state,{local,ns_bad_bucket_sup},one_for_one,[{child,<0.477.0>,{stats_reader,\"def", "y(4) ns_bad_bucket_sup","y(5) <0.70.0>", "0x00002aaaabfc2ac0 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002b74132f1130 (proc_lib:init_p_do_apply/3 + 88)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,0}]}, {heap_size,17711}, {total_heap_size,17711}, {links, [<0.70.0>,<0.470.0>,<0.477.0>,<0.469.0>,<0.64.0>]}, {memory,142696}, {message_queue_len,0}, {reductions,335}, {trap_exit,true}]}, {<0.469.0>, [{registered_name,[]}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x00002aaaaaafedf0 (gen_server:loop/6 + 288)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002aaaaabce510 Return addr 0x00002b74132f1110 (proc_lib:init_p_do_apply/3 + 56)", "y(0) []","y(1) infinity", "y(2) stats_collector", "y(3) {state,\"default\",[1660218,18122129,0,0,0,0,16004,0,0,0,0,0,0,0,0,1708,15741,0,0,0", "y(4) <0.469.0>","y(5) <0.468.0>", "0x00002aaaaabce548 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002b74132f1130 (proc_lib:init_p_do_apply/3 + 88)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,48}]}, {heap_size,6765}, {total_heap_size,17711}, {links,[<0.468.0>,<0.105.0>]}, {memory,142576}, {message_queue_len,0}, {reductions,8922968}, {trap_exit,false}]}, {<0.470.0>, [{registered_name,'stats_archiver-default'}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x00002aaaaaafedf0 (gen_server:loop/6 + 288)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002aaaaac65bb0 Return addr 0x00002b74132f1110 (proc_lib:init_p_do_apply/3 + 56)", "y(0) []","y(1) infinity", "y(2) stats_archiver", "y(3) {state,\"default\"}", "y(4) 'stats_archiver-default'", "y(5) <0.468.0>", "0x00002aaaaac65be8 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002b74132f1130 (proc_lib:init_p_do_apply/3 + 88)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,110}]}, {heap_size,6765}, {total_heap_size,24476}, {links,[<0.106.0>,<0.468.0>,<0.53.0>]}, {memory,196736}, {message_queue_len,0}, {reductions,3998504}, {trap_exit,false}]}, {<0.472.0>, [{registered_name,ns_moxi_sup}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x00002aaaaaafedf0 (gen_server:loop/6 + 288)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002aaaac01d398 Return addr 0x00002b74132f1110 (proc_lib:init_p_do_apply/3 + 56)", "y(0) []","y(1) infinity", "y(2) supervisor", "y(3) {state,{local,ns_moxi_sup},one_for_one,[],{dict,0,16,16,8,80,48,{[],[],[],[],[],[", "y(4) ns_moxi_sup","y(5) <0.70.0>", "0x00002aaaac01d3d0 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002b74132f1130 (proc_lib:init_p_do_apply/3 + 88)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,0}]}, {heap_size,17711}, {total_heap_size,17711}, {links,[<0.70.0>,<0.64.0>]}, {memory,142576}, {message_queue_len,0}, {reductions,747}, {trap_exit,true}]}, {<0.473.0>, [{registered_name,[]}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x00002aaaaaafedf0 (gen_server:loop/6 + 288)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002aaaabb7e950 Return addr 0x00002b74132f1110 (proc_lib:init_p_do_apply/3 + 56)", "y(0) []","y(1) infinity","y(2) ns_tick", "y(3) {state,1305062088207}","y(4) ns_tick", "y(5) <0.70.0>", "0x00002aaaabb7e988 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002b74132f1130 (proc_lib:init_p_do_apply/3 + 88)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,79}]}, {heap_size,2584}, {total_heap_size,2961}, {links,[<0.70.0>,<9352.423.0>,<0.53.0>]}, {memory,24840}, {message_queue_len,0}, {reductions,43200}, {trap_exit,false}]}, {<0.474.0>, [{registered_name,ns_doctor}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x00002aaaaaafedf0 (gen_server:loop/6 + 288)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002aaaab9becb8 Return addr 0x00002b74132f1110 (proc_lib:init_p_do_apply/3 + 56)", "y(0) []","y(1) infinity", "y(2) ns_doctor", "y(3) {state,{dict,2,16,16,8,80,48,{[],[],[],[],[],[],[],[],[],[],[],[],[],[],[],[]},{{", "y(4) ns_doctor","y(5) <0.70.0>", "0x00002aaaab9becf0 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002b74132f1130 (proc_lib:init_p_do_apply/3 + 88)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,627}]}, {heap_size,6765}, {total_heap_size,13530}, {links,[<0.70.0>,<0.53.0>]}, {memory,109128}, {message_queue_len,0}, {reductions,96942}, {trap_exit,false}]}, {<0.477.0>, [{registered_name,'stats_reader-default'}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x00002aaaaaafedf0 (gen_server:loop/6 + 288)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002aaaab07d4c8 Return addr 0x00002b74132f1110 (proc_lib:init_p_do_apply/3 + 56)", "y(0) []","y(1) infinity", "y(2) stats_reader", "y(3) {state,\"default\"}", "y(4) 'stats_reader-default'", "y(5) <0.468.0>", "0x00002aaaab07d500 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002b74132f1130 (proc_lib:init_p_do_apply/3 + 88)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,75}]}, {heap_size,2584}, {total_heap_size,5168}, {links,[<0.468.0>]}, {memory,42192}, {message_queue_len,0}, {reductions,73475}, {trap_exit,false}]}, {<0.513.0>, [{registered_name,[]}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x00002aaaaaaf6d08 (application_master:main_loop/2 + 64)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002aaaaaffe338 Return addr 0x00002b74132f1110 (proc_lib:init_p_do_apply/3 + 56)", "y(0) []", "y(1) {state,<0.514.0>,{appl_data,inets,[inets_sup,httpc_manager],undefined,{inets_app,", "y(2) <0.7.0>", "0x00002aaaaaffe358 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002b74132f1130 (proc_lib:init_p_do_apply/3 + 88)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,0}]}, {heap_size,233}, {total_heap_size,233}, {links,[<0.7.0>,<0.514.0>]}, {memory,2752}, {message_queue_len,0}, {reductions,23}, {trap_exit,true}]}, {<0.514.0>, [{registered_name,[]}, {status,waiting}, {initial_call,{application_master,start_it,4}}, {backtrace, ["Program counter: 0x00002aaaaaaf9390 (application_master:loop_it/4 + 80)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002aaaaac7a1a8 Return addr 0x0000000000846c78 ()", "y(0) []","y(1) inets_app", "y(2) <0.515.0>","y(3) <0.513.0>"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,0}]}, {heap_size,233}, {total_heap_size,233}, {links,[<0.513.0>,<0.515.0>]}, {memory,2648}, {message_queue_len,0}, {reductions,42}, {trap_exit,true}]}, {<0.515.0>, [{registered_name,inets_sup}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x00002aaaaaafedf0 (gen_server:loop/6 + 288)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002aaaaaffeef0 Return addr 0x00002b74132f1110 (proc_lib:init_p_do_apply/3 + 56)", "y(0) []","y(1) infinity", "y(2) supervisor", "y(3) {state,{local,inets_sup},one_for_one,[{child,<0.522.0>,tftp_sup,{tftp_sup,start_l", "y(4) inets_sup","y(5) <0.514.0>", "0x00002aaaaaffef28 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002b74132f1130 (proc_lib:init_p_do_apply/3 + 88)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,4}]}, {heap_size,377}, {total_heap_size,754}, {links, [<0.516.0>,<0.521.0>,<0.522.0>,<0.517.0>,<0.514.0>]}, {memory,7040}, {message_queue_len,0}, {reductions,320}, {trap_exit,true}]}, {<0.516.0>, [{registered_name,ftp_sup}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x00002aaaaaafedf0 (gen_server:loop/6 + 288)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002b7413341d50 Return addr 0x00002b74132f1110 (proc_lib:init_p_do_apply/3 + 56)", "y(0) []","y(1) infinity", "y(2) supervisor", "y(3) {state,{local,ftp_sup},simple_one_for_one,[{child,undefined,undefined,{ftp,start_", "y(4) ftp_sup","y(5) <0.515.0>", "0x00002b7413341d88 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002b74132f1130 (proc_lib:init_p_do_apply/3 + 88)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,0}]}, {heap_size,233}, {total_heap_size,233}, {links,[<0.515.0>]}, {memory,2712}, {message_queue_len,0}, {reductions,54}, {trap_exit,true}]}, {<0.517.0>, [{registered_name,httpc_sup}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x00002aaaaaafedf0 (gen_server:loop/6 + 288)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002aaaab077580 Return addr 0x00002b74132f1110 (proc_lib:init_p_do_apply/3 + 56)", "y(0) []","y(1) infinity", "y(2) supervisor", "y(3) {state,{local,httpc_sup},one_for_one,[{child,<0.520.0>,httpc_handler_sup,{httpc_h", "y(4) httpc_sup","y(5) <0.515.0>", "0x00002aaaab0775b8 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002b74132f1130 (proc_lib:init_p_do_apply/3 + 88)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,1}]}, {heap_size,233}, {total_heap_size,610}, {links,[<0.518.0>,<0.520.0>,<0.515.0>]}, {memory,5808}, {message_queue_len,0}, {reductions,173}, {trap_exit,true}]}, {<0.518.0>, [{registered_name,httpc_profile_sup}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x00002aaaaaafedf0 (gen_server:loop/6 + 288)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002b7413352760 Return addr 0x00002b74132f1110 (proc_lib:init_p_do_apply/3 + 56)", "y(0) []","y(1) infinity", "y(2) supervisor", "y(3) {state,{local,httpc_profile_sup},one_for_one,[{child,<0.519.0>,httpc_manager,{htt", "y(4) httpc_profile_sup","y(5) <0.517.0>", "0x00002b7413352798 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002b74132f1130 (proc_lib:init_p_do_apply/3 + 88)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,0}]}, {heap_size,233}, {total_heap_size,233}, {links,[<0.517.0>,<0.519.0>]}, {memory,2752}, {message_queue_len,0}, {reductions,124}, {trap_exit,true}]}, {<0.519.0>, [{registered_name,httpc_manager}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x00002aaaaaafedf0 (gen_server:loop/6 + 288)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002aaaab06cc98 Return addr 0x00002b74132f1110 (proc_lib:init_p_do_apply/3 + 56)", "y(0) []","y(1) infinity", "y(2) httpc_manager", "y(3) {state,[],266289,{undefined,270384},httpc_manager_session_db,httpc_manager,{optio", "y(4) httpc_manager","y(5) <0.518.0>", "0x00002aaaab06ccd0 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002b74132f1130 (proc_lib:init_p_do_apply/3 + 88)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,1}]}, {heap_size,2584}, {total_heap_size,6765}, {links,[<0.518.0>]}, {memory,54968}, {message_queue_len,0}, {reductions,318}, {trap_exit,true}]}, {<0.520.0>, [{registered_name,httpc_handler_sup}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x00002aaaaaafedf0 (gen_server:loop/6 + 288)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002aaaabba6368 Return addr 0x00002b74132f1110 (proc_lib:init_p_do_apply/3 + 56)", "y(0) []","y(1) infinity", "y(2) supervisor", "y(3) {state,{local,httpc_handler_sup},simple_one_for_one,[{child,undefined,undefined,{", "y(4) httpc_handler_sup","y(5) <0.517.0>", "0x00002aaaabba63a0 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002b74132f1130 (proc_lib:init_p_do_apply/3 + 88)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,0}]}, {heap_size,2584}, {total_heap_size,2584}, {links,[<0.517.0>]}, {memory,21520}, {message_queue_len,0}, {reductions,120}, {trap_exit,true}]}, {<0.521.0>, [{registered_name,httpd_sup}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x00002aaaaaafedf0 (gen_server:loop/6 + 288)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002b74133518c0 Return addr 0x00002b74132f1110 (proc_lib:init_p_do_apply/3 + 56)", "y(0) []","y(1) infinity", "y(2) supervisor", "y(3) {state,{local,httpd_sup},one_for_one,[],{dict,0,16,16,8,80,48,{[],[],[],[],[],[],", "y(4) httpd_sup","y(5) <0.515.0>", "0x00002b74133518f8 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002b74132f1130 (proc_lib:init_p_do_apply/3 + 88)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,0}]}, {heap_size,233}, {total_heap_size,233}, {links,[<0.515.0>]}, {memory,2712}, {message_queue_len,0}, {reductions,45}, {trap_exit,true}]}, {<0.522.0>, [{registered_name,tftp_sup}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x00002aaaaaafedf0 (gen_server:loop/6 + 288)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002aaaaaffdbd0 Return addr 0x00002b74132f1110 (proc_lib:init_p_do_apply/3 + 56)", "y(0) []","y(1) infinity", "y(2) supervisor", "y(3) {state,{local,tftp_sup},one_for_one,[],{dict,0,16,16,8,80,48,{[],[],[],[],[],[],[", "y(4) tftp_sup","y(5) <0.515.0>", "0x00002aaaaaffdc08 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002b74132f1130 (proc_lib:init_p_do_apply/3 + 88)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,0}]}, {heap_size,233}, {total_heap_size,233}, {links,[<0.515.0>]}, {memory,2712}, {message_queue_len,0}, {reductions,46}, {trap_exit,true}]}, {<0.541.0>, [{registered_name,[]}, {status,waiting}, {initial_call,{inet_tcp_dist,do_accept,6}}, {backtrace, ["Program counter: 0x00002aaaab7d5f10 (dist_util:con_loop/9 + 144)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002b741332d968 Return addr 0x0000000000846c78 ()", "y(0) []", "y(1) #Fun", "y(2) #Fun", "y(3) {tick,3514,6565,0,4}","y(4) normal", "y(5) 'ns_1@10.1.5.227'", "y(6) {net_address,{{10,1,5,229},33986},\"10.1.5.229\",tcp,inet}", "y(7) #Port<0.3676>", "y(8) 'ns_1@10.1.5.229'","y(9) <0.391.0>"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,10}]}, {heap_size,1597}, {total_heap_size,1974}, {links,[<0.391.0>,#Port<0.3676>]}, {memory,16576}, {message_queue_len,0}, {reductions,2400}, {trap_exit,false}]}, {<0.577.0>, [{registered_name,[]}, {status,waiting}, {initial_call,{erlang,apply,2}}, {backtrace, ["Program counter: 0x00002aaaab8b5e48 (ns_rebalancer:wait_for_mover/1 + 40)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002aaaaaff0078 Return addr 0x00002aaaab8b5940 (ns_rebalancer:perform_moves/4 + 304)", "y(0) <0.614.0>", "0x00002aaaaaff0088 Return addr 0x00002aaaab8b3620 (ns_rebalancer:rebalance/5 + 784)", "y(0) []","y(1) []", "y(2) [{511,'ns_1@10.1.5.227','ns_1@10.1.5.229'},{510,'ns_1@10.1.5.227','ns_1@10.1.5.22", "y(3) [['ns_1@10.1.5.227',undefined],['ns_1@10.1.5.227',undefined],['ns_1@10.1.5.227',u", "0x00002aaaaaff00b0 Return addr 0x00002aaaaaad6470 (lists:foreach/2 + 128)", "y(0) []","y(1) []", "y(2) Catch 0x00002aaaab8b3980 (ns_rebalancer:rebalance/5 + 1648)", "y(3) []","y(4) []", "y(5) ['ns_1@10.1.5.227','ns_1@10.1.5.229']", "y(6) \"default\"", "0x00002aaaaaff00f0 Return addr 0x00002aaaab8b3138 (ns_rebalancer:rebalance/3 + 704)", "y(0) #Fun", "y(1) []", "0x00002aaaaaff0108 Return addr 0x0000000000846c78 ()", "y(0) []","y(1) []", "y(2) Catch 0x00002aaaab8b3158 (ns_rebalancer:rebalance/3 + 736)", "y(3) []","y(4) []","y(5) []", "y(6) []","y(7) []"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,134}]}, {heap_size,6765}, {total_heap_size,53133}, {links,[<0.109.0>,<0.614.0>]}, {memory,425848}, {message_queue_len,0}, {reductions,1416211}, {trap_exit,true}]}, {<0.614.0>, [{registered_name,[]}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x00002aaaaaafedf0 (gen_server:loop/6 + 288)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002aaaab9b1948 Return addr 0x00002b74132f1110 (proc_lib:init_p_do_apply/3 + 56)", "y(0) []","y(1) infinity", "y(2) ns_vbucket_mover", "y(3) {state,\"default\",{dict,2,16,16,8,80,48,{[],[],[],[],[],[],[],[],[],[],[],[],[],[]", "y(4) <0.614.0>","y(5) <0.577.0>", "0x00002aaaab9b1980 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002b74132f1130 (proc_lib:init_p_do_apply/3 + 88)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,183}]}, {heap_size,6765}, {total_heap_size,13530}, {links,[<0.577.0>,<0.6871.0>]}, {memory,109128}, {message_queue_len,0}, {reductions,1314810}, {trap_exit,false}]}, {<0.2148.0>, [{registered_name,[]}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x00002aaaab22eb10 (menelaus_web:handle_pool_info_wait/6 + 280)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002aaaaba91700 Return addr 0x00002aaaab22e590 (menelaus_web:check_and_handle_pool_info", "y(0) \"43798496\"","y(1) \"43798496\"", "y(2) 20000","y(3) \"10.1.5.227\"", "y(4) {\"Administrator\",\"password\"}", "y(5) \"default\"", "y(6) {mochiweb_request,#Port<0.6794>,'GET',\"/pools/default?waitChange=20000&etag=43798", "0x00002aaaaba91740 Return addr 0x00002aaaab22d4e8 (menelaus_web:loop/3 + 24160)", "y(0) Catch 0x00002aaaab22e5e8 (menelaus_web:check_and_handle_pool_info/2 + 336)", "y(1) []","y(2) <0.6844.0>", "0x00002aaaaba91760 Return addr 0x00002aaaab246680 (mochiweb_http:headers/5 + 1360)", "y(0) []","y(1) []","y(2) []", "y(3) []", "y(4) {mochiweb_request,#Port<0.6794>,'GET',\"/pools/default?waitChange=20000&etag=43798", "y(5) Catch 0x00002aaaab22d578 (menelaus_web:loop/3 + 24304)", "0x00002aaaaba91798 Return addr 0x00002b74132f1110 (proc_lib:init_p_do_apply/3 + 56)", "y(0) #Fun", "y(1) []","y(2) []", "y(3) {mochiweb_request,#Port<0.6794>,'GET',\"/pools/default?waitChange=20000&etag=43798", "0x00002aaaaba917c0 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002b74132f1130 (proc_lib:init_p_do_apply/3 + 88)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,6}]}, {heap_size,17711}, {total_heap_size,46368}, {links,[<0.95.0>,<0.6844.0>,#Port<0.6794>]}, {memory,372016}, {message_queue_len,0}, {reductions,1889149}, {trap_exit,false}]}, {<0.6480.0>, [{registered_name,[]}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x00002b74132aa1d8 (prim_inet:recv0/3 + 224)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002aaaaac38058 Return addr 0x00002aaaab245da8 (mochiweb_http:request/2 + 120)", "y(0) 2362","y(1) #Port<0.8875>", "0x00002aaaaac38070 Return addr 0x00002b74132f1110 (proc_lib:init_p_do_apply/3 + 56)", "y(0) #Fun", "y(1) #Port<0.8875>", "0x00002aaaaac38088 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002b74132f1130 (proc_lib:init_p_do_apply/3 + 88)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,1}]}, {heap_size,4181}, {total_heap_size,32838}, {links,[<0.95.0>,#Port<0.8875>]}, {memory,263592}, {message_queue_len,0}, {reductions,1407919}, {trap_exit,false}]}, {<0.6680.0>, [{registered_name,[]}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x00002aaaaaaff7d0 (gen_server:do_multi_call/4 + 768)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002aaaabb8bc08 Return addr 0x00002aaaaab5e6e0 (rpc:do_multicall/5 + 320)", "y(0) []","y(1) #Ref<0.0.0.150230>", "y(2) <0.6876.0>","y(3) #Ref<0.0.0.150229>", "0x00002aaaabb8bc30 Return addr 0x00002aaaab1d52f8 (diag_handler:diag_multicall/3 + 232)", "y(0) []","y(1) []", "0x00002aaaabb8bc48 Return addr 0x00002aaaab1d5f28 (diag_handler:handle_diag/1 + 384)", "y(0) ['ns_1@10.1.5.227','ns_1@10.1.5.229']", "y(1) []","y(2) []","y(3) []", "0x00002aaaabb8bc70 Return addr 0x00002aaaab22d000 (menelaus_web:loop/3 + 22904)", "y(0) []", "y(1) [[\"2011\",45,[\"0\",\"5\"],45,\"10\",32,\"13\",58,\"50\",58,[\"0\",\"8\"],46,\"183\"],32,\"ns_node_", "y(2) [{\"default\",[{num_replicas,1},{ram_quota,1096810496},{auth_type,sasl},{sasl_passw", "y(3) {mochiweb_request,#Port<0.8903>,'GET',\"/diag\",{1,1},{9,{\"host\",{'Host',\"10.1.5.22", "0x00002aaaabb8bc98 Return addr 0x00002aaaab246680 (mochiweb_http:headers/5 + 1360)", "y(0) []","y(1) []","y(2) []", "y(3) []", "y(4) {mochiweb_request,#Port<0.8903>,'GET',\"/diag\",{1,1},{9,{\"host\",{'Host',\"10.1.5.22", "y(5) Catch 0x00002aaaab22d578 (menelaus_web:loop/3 + 24304)", "0x00002aaaabb8bcd0 Return addr 0x00002b74132f1110 (proc_lib:init_p_do_apply/3 + 56)", "y(0) #Fun", "y(1) []","y(2) []", "y(3) {mochiweb_request,#Port<0.8903>,'GET',\"/diag\",{1,1},{9,{\"host\",{'Host',\"10.1.5.22", "0x00002aaaabb8bcf8 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002b74132f1130 (proc_lib:init_p_do_apply/3 + 88)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,4}]}, {heap_size,6765}, {total_heap_size,35422}, {links,[<0.95.0>,#Port<0.8903>]}, {memory,284408}, {message_queue_len,0}, {reductions,432905}, {trap_exit,false}]}, {<0.6806.0>, [{registered_name,[]}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x00002b74132a8d78 (prim_inet:accept0/2 + 184)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002aaaaabd96c8 Return addr 0x00002aaaaaf7d900 (inet_tcp:accept/1 + 40)", "y(0) 43394","y(1) #Port<0.1399>", "0x00002aaaaabd96e0 Return addr 0x00002aaaab254a90 (mochiweb_socket_server:acceptor_loop/1 ", "y(0) []", "0x00002aaaaabd96f0 Return addr 0x00002b74132f1110 (proc_lib:init_p_do_apply/3 + 56)", "y(0) #Fun", "y(1) <0.95.0>", "y(2) Catch 0x00002aaaab254a90 (mochiweb_socket_server:acceptor_loop/1 + 160)", "0x00002aaaaabd9710 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002b74132f1130 (proc_lib:init_p_do_apply/3 + 88)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,1}]}, {heap_size,377}, {total_heap_size,754}, {links,[<0.95.0>]}, {memory,6952}, {message_queue_len,0}, {reductions,39}, {trap_exit,false}]}, {<0.6840.0>, [{registered_name,[]}, {status,waiting}, {initial_call,{erlang,apply,2}}, {backtrace, ["Program counter: 0x00002aaaab1d6e50 (diag_handler:'-arm_timeout/2-fun-0-'/3 + 64)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002aaaaabada68 Return addr 0x0000000000846c78 ()", "y(0) <0.360.0>", "y(1) #Fun", "y(2) 23000"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,0}]}, {heap_size,233}, {total_heap_size,233}, {links,[<0.360.0>]}, {memory,2608}, {message_queue_len,0}, {reductions,2}, {trap_exit,false}]}, {<0.6844.0>, [{registered_name,[]}, {status,waiting}, {initial_call,{erlang,apply,2}}, {backtrace, ["Program counter: 0x00002aaaab1d6e50 (diag_handler:'-arm_timeout/2-fun-0-'/3 + 64)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002b7413397318 Return addr 0x0000000000846c78 ()", "y(0) <0.2148.0>", "y(1) #Fun", "y(2) 23000"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,0}]}, {heap_size,233}, {total_heap_size,233}, {links,[<0.2148.0>]}, {memory,2608}, {message_queue_len,0}, {reductions,2}, {trap_exit,false}]}, {<0.6871.0>, [{registered_name,[]}, {status,waiting}, {initial_call,{erlang,apply,2}}, {backtrace, ["Program counter: 0x00002aaaabe626f0 (ns_vbucket_mover:wait_for_mover/5 + 88)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002aaaabab4108 Return addr 0x00002aaaabe63420 (ns_vbucket_mover:'-spawn_workers/1-fun-", "y(0) 2","y(1) 'ns_1@10.1.5.229'", "y(2) 'ns_1@10.1.5.227'","y(3) 213", "y(4) \"default\"", "0x00002aaaabab4138 Return addr 0x0000000000846c78 ()", "y(0) 'ns_1@10.1.5.227'","y(1) <0.614.0>"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,2}]}, {heap_size,17711}, {total_heap_size,46368}, {links,[<0.614.0>,<0.6872.0>]}, {memory,371728}, {message_queue_len,0}, {reductions,492}, {trap_exit,true}]}, {<0.6872.0>, [{registered_name,[]}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x00002aaaaaafedf0 (gen_server:loop/6 + 288)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002aaaabbe5e78 Return addr 0x00002b74132f1110 (proc_lib:init_p_do_apply/3 + 56)", "y(0) []","y(1) infinity", "y(2) ns_port_server", "y(3) {state,#Port<0.8934>,vbucketmigrator,{[\"Bucket 213 moved to the next server\",\"Sta", "y(4) <0.6872.0>","y(5) <0.6871.0>", "0x00002aaaabbe5eb0 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002b74132f1130 (proc_lib:init_p_do_apply/3 + 88)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,1}]}, {heap_size,987}, {total_heap_size,1974}, {links,[<0.6871.0>,#Port<0.8934>]}, {memory,16680}, {message_queue_len,0}, {reductions,315}, {trap_exit,true}]}, {<0.6876.0>, [{registered_name,[]}, {status,waiting}, {initial_call,{erlang,apply,2}}, {backtrace, ["Program counter: 0x00002aaaaaaffdf0 (gen_server:rec_nodes/7 + 240)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002aaaadd669b8 Return addr 0x00002aaaaab05f28 (gen_server:'-do_multi_call/4-fun-0-'/6 ", "y(0) #Ref<0.0.0.150234>","y(1) 2000", "y(2) [{'ns_1@10.1.5.229',[{version,[{kernel,\"2.13.4\"},{mnesia,\"4.4.12\"},{stdlib,\"1.16.", "y(3) []","y(4) rex", "y(5) #Ref<0.0.0.150229>","y(6) []", "y(7) #Ref<0.0.0.150232>", "y(8) 'ns_1@10.1.5.227'", "0x00002aaaadd66a08 Return addr 0x0000000000846c78 ()", "y(0) #Ref<0.0.0.150229>","y(1) []", "y(2) []","y(3) []","y(4) []"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,0}]}, {heap_size,196418}, {total_heap_size,196418}, {links,[]}, {memory,1572264}, {message_queue_len,0}, {reductions,17543}, {trap_exit,true}]}, {<0.6877.0>, [{registered_name,[]}, {status,running}, {initial_call,{erlang,apply,2}}, {backtrace, ["Program counter: 0x00002b741322ec00 (unknown function)", "CP: 0x00002aaaab1d4da8 (diag_handler:grab_process_info/1 + 64)", "0x00002aaaaef10478 Return addr 0x00002aaaab1d7308 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) []", "0x00002aaaaef10488 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) []","y(1) <0.6877.0>", "0x00002aaaaef104a0 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,[]},{status,waiting},{initial_call,{erlang,apply,2}},{backtrace", "y(1) <0.6876.0>", "0x00002aaaaef104b8 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,[]},{status,waiting},{initial_call,{proc_lib,init_p,5}},{backtr", "y(1) <0.6872.0>", "0x00002aaaaef104d0 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,[]},{status,waiting},{initial_call,{erlang,apply,2}},{backtrace", "y(1) <0.6871.0>", "0x00002aaaaef104e8 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,[]},{status,waiting},{initial_call,{erlang,apply,2}},{backtrace", "y(1) <0.6844.0>", "0x00002aaaaef10500 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,[]},{status,waiting},{initial_call,{erlang,apply,2}},{backtrace", "y(1) <0.6840.0>", "0x00002aaaaef10518 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,[]},{status,waiting},{initial_call,{proc_lib,init_p,5}},{backtr", "y(1) <0.6806.0>", "0x00002aaaaef10530 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,[]},{status,waiting},{initial_call,{proc_lib,init_p,5}},{backtr", "y(1) <0.6680.0>", "0x00002aaaaef10548 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,[]},{status,waiting},{initial_call,{proc_lib,init_p,5}},{backtr", "y(1) <0.6480.0>", "0x00002aaaaef10560 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,[]},{status,waiting},{initial_call,{proc_lib,init_p,5}},{backtr", "y(1) <0.2148.0>", "0x00002aaaaef10578 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,[]},{status,waiting},{initial_call,{proc_lib,init_p,5}},{backtr", "y(1) <0.614.0>", "0x00002aaaaef10590 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,[]},{status,waiting},{initial_call,{erlang,apply,2}},{backtrace", "y(1) <0.577.0>", "0x00002aaaaef105a8 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,[]},{status,waiting},{initial_call,{inet_tcp_dist,do_accept,6}}", "y(1) <0.541.0>", "0x00002aaaaef105c0 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,tftp_sup},{status,waiting},{initial_call,{proc_lib,init_p,5}},{", "y(1) <0.522.0>", "0x00002aaaaef105d8 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,httpd_sup},{status,waiting},{initial_call,{proc_lib,init_p,5}},", "y(1) <0.521.0>", "0x00002aaaaef105f0 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,httpc_handler_sup},{status,waiting},{initial_call,{proc_lib,ini", "y(1) <0.520.0>", "0x00002aaaaef10608 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,httpc_manager},{status,waiting},{initial_call,{proc_lib,init_p,", "y(1) <0.519.0>", "0x00002aaaaef10620 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,httpc_profile_sup},{status,waiting},{initial_call,{proc_lib,ini", "y(1) <0.518.0>", "0x00002aaaaef10638 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,httpc_sup},{status,waiting},{initial_call,{proc_lib,init_p,5}},", "y(1) <0.517.0>", "0x00002aaaaef10650 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,ftp_sup},{status,waiting},{initial_call,{proc_lib,init_p,5}},{b", "y(1) <0.516.0>", "0x00002aaaaef10668 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,inets_sup},{status,waiting},{initial_call,{proc_lib,init_p,5}},", "y(1) <0.515.0>", "0x00002aaaaef10680 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,[]},{status,waiting},{initial_call,{application_master,start_it", "y(1) <0.514.0>", "0x00002aaaaef10698 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,[]},{status,waiting},{initial_call,{proc_lib,init_p,5}},{backtr", "y(1) <0.513.0>", "0x00002aaaaef106b0 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,'stats_reader-default'},{status,waiting},{initial_call,{proc_li", "y(1) <0.477.0>", "0x00002aaaaef106c8 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,ns_doctor},{status,waiting},{initial_call,{proc_lib,init_p,5}},", "y(1) <0.474.0>", "0x00002aaaaef106e0 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,[]},{status,waiting},{initial_call,{proc_lib,init_p,5}},{backtr", "y(1) <0.473.0>", "0x00002aaaaef106f8 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,ns_moxi_sup},{status,waiting},{initial_call,{proc_lib,init_p,5}", "y(1) <0.472.0>", "0x00002aaaaef10710 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,'stats_archiver-default'},{status,waiting},{initial_call,{proc_", "y(1) <0.470.0>", "0x00002aaaaef10728 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,[]},{status,waiting},{initial_call,{proc_lib,init_p,5}},{backtr", "y(1) <0.469.0>", "0x00002aaaaef10740 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,ns_bad_bucket_sup},{status,waiting},{initial_call,{proc_lib,ini", "y(1) <0.468.0>", "0x00002aaaaef10758 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,ns_bad_bucket_worker},{status,waiting},{initial_call,{proc_lib,", "y(1) <0.467.0>", "0x00002aaaaef10770 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,mnesia_late_loader},{status,waiting},{initial_call,{proc_lib,in", "y(1) <0.461.0>", "0x00002aaaaef10788 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,mnesia_controller},{status,waiting},{initial_call,{proc_lib,ini", "y(1) <0.460.0>", "0x00002aaaaef107a0 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,mnesia_snmp_sup},{status,waiting},{initial_call,{proc_lib,init_", "y(1) <0.459.0>", "0x00002aaaaef107b8 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,mnesia_checkpoint_sup},{status,waiting},{initial_call,{proc_lib", "y(1) <0.458.0>", "0x00002aaaaef107d0 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,[]},{status,waiting},{initial_call,{proc_lib,init_p,5}},{backtr", "y(1) <0.456.0>", "0x00002aaaaef107e8 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,mnesia_tm},{status,runnable},{initial_call,{proc_lib,init_p,5}}", "y(1) <0.421.0>", "0x00002aaaaef10800 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,mnesia_recover},{status,waiting},{initial_call,{proc_lib,init_p", "y(1) <0.420.0>", "0x00002aaaaef10818 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,mnesia_locker},{status,waiting},{initial_call,{proc_lib,init_p,", "y(1) <0.419.0>", "0x00002aaaaef10830 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,mnesia_subscr},{status,waiting},{initial_call,{proc_lib,init_p,", "y(1) <0.418.0>", "0x00002aaaaef10848 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,mnesia_monitor},{status,waiting},{initial_call,{proc_lib,init_p", "y(1) <0.415.0>", "0x00002aaaaef10860 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,mnesia_kernel_sup},{status,waiting},{initial_call,{proc_lib,ini", "y(1) <0.414.0>", "0x00002aaaaef10878 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,mnesia_event},{status,waiting},{initial_call,{proc_lib,init_p,5", "y(1) <0.413.0>", "0x00002aaaaef10890 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,mnesia_sup},{status,waiting},{initial_call,{proc_lib,init_p,5}}", "y(1) <0.412.0>", "0x00002aaaaef108a8 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,[]},{status,waiting},{initial_call,{application_master,start_it", "y(1) <0.411.0>", "0x00002aaaaef108c0 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,[]},{status,waiting},{initial_call,{proc_lib,init_p,5}},{backtr", "y(1) <0.410.0>", "0x00002aaaaef108d8 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,ns_mnesia},{status,waiting},{initial_call,{proc_lib,init_p,5}},", "y(1) <0.408.0>", "0x00002aaaaef108f0 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,[]},{status,waiting},{initial_call,{net_kernel,ticker,2}},{back", "y(1) <0.393.0>", "0x00002aaaaef10908 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,[]},{status,waiting},{initial_call,{inet_tcp_dist,accept_loop,2", "y(1) <0.392.0>", "0x00002aaaaef10920 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,net_kernel},{status,waiting},{initial_call,{proc_lib,init_p,5}}", "y(1) <0.391.0>", "0x00002aaaaef10938 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,auth},{status,waiting},{initial_call,{proc_lib,init_p,5}},{back", "y(1) <0.390.0>", "0x00002aaaaef10950 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,erl_epmd},{status,waiting},{initial_call,{proc_lib,init_p,5}},{", "y(1) <0.389.0>", "0x00002aaaaef10968 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,net_sup},{status,waiting},{initial_call,{proc_lib,init_p,5}},{b", "y(1) <0.388.0>", "0x00002aaaaef10980 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,[]},{status,waiting},{initial_call,{proc_lib,init_p,5}},{backtr", "y(1) <0.360.0>", "0x00002aaaaef10998 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,hot_keys_keeper},{status,waiting},{initial_call,{proc_lib,init_", "y(1) <0.349.0>", "0x00002aaaaef109b0 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,inet_gethost_native},{status,waiting},{initial_call,{inet_getho", "y(1) <0.305.0>", "0x00002aaaaef109c8 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,inet_gethost_native_sup},{status,waiting},{initial_call,{proc_l", "y(1) <0.304.0>", "0x00002aaaaef109e0 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,'ns_memcached-default'},{status,waiting},{initial_call,{proc_li", "y(1) <0.303.0>", "0x00002aaaaef109f8 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,'ns_vbm_sup-default'},{status,waiting},{initial_call,{proc_lib,", "y(1) <0.302.0>", "0x00002aaaaef10a10 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,[]},{status,waiting},{initial_call,{proc_lib,init_p,5}},{backtr", "y(1) <0.160.0>", "0x00002aaaaef10a28 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,[]},{status,waiting},{initial_call,{proc_lib,init_p,5}},{backtr", "y(1) <0.159.0>", "0x00002aaaaef10a40 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,[]},{status,waiting},{initial_call,{proc_lib,init_p,5}},{backtr", "y(1) <0.158.0>", "0x00002aaaaef10a58 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,dets},{status,waiting},{initial_call,{proc_lib,init_p,5}},{back", "y(1) <0.136.0>", "0x00002aaaaef10a70 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,dets_sup},{status,waiting},{initial_call,{proc_lib,init_p,5}},{", "y(1) <0.135.0>", "0x00002aaaaef10a88 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,disk_log_server},{status,waiting},{initial_call,{proc_lib,init_", "y(1) <0.132.0>", "0x00002aaaaef10aa0 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,disk_log_sup},{status,waiting},{initial_call,{proc_lib,init_p,5", "y(1) <0.131.0>", "0x00002aaaaef10ab8 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,[]},{status,waiting},{initial_call,{proc_lib,init_p,5}},{backtr", "y(1) <0.109.0>", "0x00002aaaaef10ad0 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,ns_good_bucket_sup},{status,waiting},{initial_call,{proc_lib,in", "y(1) <0.108.0>", "0x00002aaaaef10ae8 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,ns_good_bucket_worker},{status,waiting},{initial_call,{proc_lib", "y(1) <0.107.0>", "0x00002aaaaef10b00 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,ns_stats_event},{status,waiting},{initial_call,{proc_lib,init_p", "y(1) <0.106.0>", "0x00002aaaaef10b18 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,ns_tick_event},{status,waiting},{initial_call,{proc_lib,init_p,", "y(1) <0.105.0>", "0x00002aaaaef10b30 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,[]},{status,waiting},{initial_call,{proc_lib,init_p,5}},{backtr", "y(1) <0.104.0>", "0x00002aaaaef10b48 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,[]},{status,waiting},{initial_call,{proc_lib,init_p,5}},{backtr", "y(1) <0.103.0>", "0x00002aaaaef10b60 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,ns_port_sup},{status,waiting},{initial_call,{proc_lib,init_p,5}", "y(1) <0.99.0>", "0x00002aaaaef10b78 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,menelaus_web},{status,waiting},{initial_call,{proc_lib,init_p,5", "y(1) <0.95.0>", "0x00002aaaaef10b90 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,menelaus_sup},{status,waiting},{initial_call,{proc_lib,init_p,5", "y(1) <0.94.0>", "0x00002aaaaef10ba8 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,ns_heart},{status,waiting},{initial_call,{proc_lib,init_p,5}},{", "y(1) <0.90.0>", "0x00002aaaaef10bc0 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,ns_config_rep},{status,waiting},{initial_call,{proc_lib,init_p,", "y(1) <0.85.0>", "0x00002aaaaef10bd8 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,ns_node_disco},{status,waiting},{initial_call,{proc_lib,init_p,", "y(1) <0.78.0>", "0x00002aaaaef10bf0 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,ns_node_disco_events},{status,waiting},{initial_call,{proc_lib,", "y(1) <0.77.0>", "0x00002aaaaef10c08 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,ns_node_disco_sup},{status,waiting},{initial_call,{proc_lib,ini", "y(1) <0.76.0>", "0x00002aaaaef10c20 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,ns_mail},{status,waiting},{initial_call,{proc_lib,init_p,5}},{b", "y(1) <0.74.0>", "0x00002aaaaef10c38 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,ns_mail_sup},{status,waiting},{initial_call,{proc_lib,init_p,5}", "y(1) <0.73.0>", "0x00002aaaaef10c50 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,ns_log_events},{status,waiting},{initial_call,{proc_lib,init_p,", "y(1) <0.72.0>", "0x00002aaaaef10c68 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,ns_log},{status,waiting},{initial_call,{proc_lib,init_p,5}},{ba", "y(1) <0.71.0>", "0x00002aaaaef10c80 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,ns_server_sup},{status,waiting},{initial_call,{proc_lib,init_p,", "y(1) <0.70.0>", "0x00002aaaaef10c98 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,ns_config},{status,waiting},{initial_call,{proc_lib,init_p,5}},", "y(1) <0.65.0>", "0x00002aaaaef10cb0 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,ns_config_events},{status,waiting},{initial_call,{proc_lib,init", "y(1) <0.64.0>", "0x00002aaaaef10cc8 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,ns_config_sup},{status,waiting},{initial_call,{proc_lib,init_p,", "y(1) <0.63.0>", "0x00002aaaaef10ce0 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,ns_cluster},{status,waiting},{initial_call,{proc_lib,init_p,5}}", "y(1) <0.62.0>", "0x00002aaaaef10cf8 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,dist_manager},{status,waiting},{initial_call,{proc_lib,init_p,5", "y(1) <0.55.0>", "0x00002aaaaef10d10 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,ns_server_cluster_sup},{status,waiting},{initial_call,{proc_lib", "y(1) <0.54.0>", "0x00002aaaaef10d28 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,timer_server},{status,waiting},{initial_call,{proc_lib,init_p,5", "y(1) <0.53.0>", "0x00002aaaaef10d40 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,[]},{status,waiting},{initial_call,{application_master,start_it", "y(1) <0.52.0>", "0x00002aaaaef10d58 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,[]},{status,waiting},{initial_call,{proc_lib,init_p,5}},{backtr", "y(1) <0.51.0>", "0x00002aaaaef10d70 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,[]},{status,waiting},{initial_call,{erlang,apply,2}},{backtrace", "y(1) <0.49.0>", "0x00002aaaaef10d88 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,[]},{status,waiting},{initial_call,{erlang,apply,2}},{backtrace", "y(1) <0.48.0>", "0x00002aaaaef10da0 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,cpu_sup},{status,waiting},{initial_call,{proc_lib,init_p,5}},{b", "y(1) <0.47.0>", "0x00002aaaaef10db8 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,[]},{status,waiting},{initial_call,{erlang,apply,2}},{backtrace", "y(1) <0.46.0>", "0x00002aaaaef10dd0 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,memsup},{status,waiting},{initial_call,{proc_lib,init_p,5}},{ba", "y(1) <0.45.0>", "0x00002aaaaef10de8 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,disksup},{status,waiting},{initial_call,{proc_lib,init_p,5}},{b", "y(1) <0.44.0>", "0x00002aaaaef10e00 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,os_mon_sup},{status,waiting},{initial_call,{proc_lib,init_p,5}}", "y(1) <0.43.0>", "0x00002aaaaef10e18 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,[]},{status,waiting},{initial_call,{application_master,start_it", "y(1) <0.42.0>", "0x00002aaaaef10e30 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,[]},{status,waiting},{initial_call,{proc_lib,init_p,5}},{backtr", "y(1) <0.41.0>", "0x00002aaaaef10e48 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,release_handler},{status,waiting},{initial_call,{proc_lib,init_", "y(1) <0.38.0>", "0x00002aaaaef10e60 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,overload},{status,waiting},{initial_call,{proc_lib,init_p,5}},{", "y(1) <0.37.0>", "0x00002aaaaef10e78 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,alarm_handler},{status,waiting},{initial_call,{proc_lib,init_p,", "y(1) <0.36.0>", "0x00002aaaaef10e90 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,sasl_safe_sup},{status,waiting},{initial_call,{proc_lib,init_p,", "y(1) <0.35.0>", "0x00002aaaaef10ea8 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,sasl_sup},{status,waiting},{initial_call,{proc_lib,init_p,5}},{", "y(1) <0.34.0>", "0x00002aaaaef10ec0 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,[]},{status,waiting},{initial_call,{application_master,start_it", "y(1) <0.33.0>", "0x00002aaaaef10ed8 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,[]},{status,waiting},{initial_call,{proc_lib,init_p,5}},{backtr", "y(1) <0.32.0>", "0x00002aaaaef10ef0 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,kernel_safe_sup},{status,waiting},{initial_call,{proc_lib,init_", "y(1) <0.27.0>", "0x00002aaaaef10f08 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,[]},{status,waiting},{initial_call,{proc_lib,init_p,5}},{backtr", "y(1) <0.26.0>", "0x00002aaaaef10f20 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,user},{status,waiting},{initial_call,{user,server,2}},{backtrac", "y(1) <0.25.0>", "0x00002aaaaef10f38 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,[]},{status,waiting},{initial_call,{proc_lib,init_p,5}},{backtr", "y(1) <0.24.0>", "0x00002aaaaef10f50 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,standard_error},{status,waiting},{initial_call,{standard_error,", "y(1) <0.23.0>", "0x00002aaaaef10f68 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,standard_error_sup},{status,waiting},{initial_call,{proc_lib,in", "y(1) <0.22.0>", "0x00002aaaaef10f80 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,code_server},{status,waiting},{initial_call,{erlang,apply,2}},{", "y(1) <0.21.0>", "0x00002aaaaef10f98 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,file_server_2},{status,waiting},{initial_call,{proc_lib,init_p,", "y(1) <0.20.0>", "0x00002aaaaef10fb0 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,global_group},{status,waiting},{initial_call,{proc_lib,init_p,5", "y(1) <0.19.0>", "0x00002aaaaef10fc8 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,inet_db},{status,waiting},{initial_call,{proc_lib,init_p,5}},{b", "y(1) <0.17.0>", "0x00002aaaaef10fe0 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,[]},{status,waiting},{initial_call,{erlang,apply,2}},{backtrace", "y(1) <0.16.0>", "0x00002aaaaef10ff8 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,[]},{status,waiting},{initial_call,{erlang,apply,2}},{backtrace", "y(1) <0.15.0>", "0x00002aaaaef11010 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,[]},{status,waiting},{initial_call,{erlang,apply,2}},{backtrace", "y(1) <0.14.0>", "0x00002aaaaef11028 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,global_name_server},{status,waiting},{initial_call,{proc_lib,in", "y(1) <0.13.0>", "0x00002aaaaef11040 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,rex},{status,waiting},{initial_call,{proc_lib,init_p,5}},{backt", "y(1) <0.12.0>", "0x00002aaaaef11058 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,kernel_sup},{status,waiting},{initial_call,{proc_lib,init_p,5}}", "y(1) <0.11.0>", "0x00002aaaaef11070 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,[]},{status,waiting},{initial_call,{application_master,start_it", "y(1) <0.10.0>", "0x00002aaaaef11088 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,[]},{status,waiting},{initial_call,{proc_lib,init_p,5}},{backtr", "y(1) <0.9.0>", "0x00002aaaaef110a0 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,application_controller},{status,waiting},{initial_call,{erlang,", "y(1) <0.7.0>", "0x00002aaaaef110b8 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,error_logger},{status,waiting},{initial_call,{proc_lib,init_p,5", "y(1) <0.6.0>", "0x00002aaaaef110d0 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,erl_prim_loader},{status,waiting},{initial_call,{erlang,apply,2", "y(1) <0.3.0>", "0x00002aaaaef110e8 Return addr 0x00002aaaab1d4fb8 (diag_handler:do_diag_per_node/0 + 224)", "y(0) [{registered_name,init},{status,waiting},{initial_call,{otp_ring0,start,2}},{back", "y(1) <0.0.0>", "0x00002aaaaef11100 Return addr 0x00002aaaaab60208 (rpc:'-handle_call_call/6-fun-0-'/5 + 20", "y(0) []","y(1) []", "y(2) [{version,[{kernel,\"2.13.4\"},{mnesia,\"4.4.12\"},{stdlib,\"1.16.4\"},{os_mon,\"2.2.4\"}", "y(3) [{{node,'ns_1@127.0.0.1',ns_log},[{filename,\"/var/opt/membase/1.6.5.4r/data/ns_1/", "y(4) [\"bucket_engine_1.6.5.3_1_g5ba5371-Linux.x86_64.tar.gz\",\"curl-7.21.1-w64_patched.", "y(5) [{kernel,\"2.13.4\"},{mnesia,\"4.4.12\"},{stdlib,\"1.16.4\"},{os_mon,\"2.2.4\"},{inets,\"5", "0x00002aaaaef11138 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002aaaaab60208 (rpc:'-handle_call_call/6-fun-0-'/5 + 208)", "y(1) []","y(2) []","y(3) []", "y(4) <0.12.0>"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,2}]}, {heap_size,832040}, {total_heap_size,1346269}, {links,[]}, {memory,10770928}, {message_queue_len,0}, {reductions,1364538}, {trap_exit,false}]}]}, {memory,{2058731520,680501248,{<0.12.0>,6657168}}}, {disk, [{"/",11204616,49}, {"/boot",101086,14}, {"/dev/shm",1005240,0}]}]}, {'ns_1@10.1.5.229', [{version, [{kernel,"2.13.4"}, {mnesia,"4.4.12"}, {stdlib,"1.16.4"}, {os_mon,"2.2.4"}, {ns_server,"1.6.5.4r"}, {menelaus,"1.6.5.4r"}, {sasl,"2.1.8"}]}, {manifest, ["bucket_engine_1.6.5.3_1_g5ba5371-Linux.x86_64.tar.gz", "curl-7.21.1-w64_patched.tar.gz", "ep-engine_1.6.5.3.1_13_g81a6cf1-Linux.x86_64.tar.gz", "google-perftools-1.6","google-perftools-1.6.tar.gz", "libconflate_1.6.5-Linux.x86_64.tar.gz", "libevent-2.0.7-rc.tar.gz", "libmemcached-0.41_trond-norbye_mingw32-revno895.tar.gz", "libvbucket_1.6.4-Linux.x86_64.tar.gz", "membase-cli_1.6.5-Linux.x86_64.tar.gz", "memcached_1.4.4_382_g9df3289-Linux.x86_64.tar.gz", "moxi_1.6.5.2-Linux.x86_64.tar.gz", "ns_server_1.6.5.4r.tar.gz", "vbucketmigrator_1.6.5_2_gd2c8834-Linux.x86_64.tar.gz", "wallace_1.6.5.4r-8-g00eec2b-Linux.x86_64"]}, {config, [{{node,'ns_1@127.0.0.1',ns_log}, [{filename, "/var/opt/membase/1.6.5.4r/data/ns_1/ns_log"}]}, {otp, [{'_vclock',[{'ns_1@10.1.5.229',{1,63472280023}}]}, {cookie,bujipmggocppuqmx}]}, {memory_quota,1570}, {buckets, [{'_vclock', [{'ns_1@10.1.5.227',{2,63472280025}}, {'ns_1@127.0.0.1',{3,63472280007}}]}, {configs, [{"default", [{num_replicas,1}, {ram_quota,1096810496}, {auth_type,sasl}, {sasl_password,[]}, {type,membase}, {num_vbuckets,1024}, {db_shards,4}, {ht_size,3079}, {tap_keepalive,0}, {tap_noop_interval,20}, {max_txn_size,1000}, {ht_locks,5}, {servers,['ns_1@10.1.5.227','ns_1@10.1.5.229']}, {map, [['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined]]}]}]}]}, {{node,'ns_1@127.0.0.1',membership},active}, {rest_creds, [{'_vclock',[{'ns_1@127.0.0.1',{1,63472279813}}]}, {creds, [{"Administrator",[{password,'filtered-out'}]}]}]}, {port_servers, [{moxi,"./bin/moxi/moxi", ["-Z", {"port_listen=~B,default_bucket_name=default,downstream_max=1024,downstream_conn_max=4,connect_max_errors=5,connect_retry_interval=30000,connect_timeout=400,auth_timeout=100,cycle=200,downstream_conn_queue_timeout=200,downstream_timeout=5000,wait_queue_timeout=200", [port]}, "-z", {"url=http://127.0.0.1:~B/pools/default/saslBucketsStreaming", [{rest,port}]}, "-p","0","-Y","y","-O","stderr", {"~s",[verbosity]}], [{env, [{"EVENT_NOSELECT","1"}, {"MOXI_SASL_PLAIN_USR", {"~s",[{ns_moxi_sup,rest_user,[]}]}}, {"MOXI_SASL_PLAIN_PWD", {"~s",[{ns_moxi_sup,rest_pass,[]}]}}]}, use_stdio,stderr_to_stdout,stream]}, {memcached,"./bin/memcached/memcached", ["-X","./bin/memcached/stdin_term_handler.so","-p", {"~B",[port]}, "-E","./bin/bucket_engine/bucket_engine.so","-B", "binary","-r","-c","10000","-e", {"admin=~s;default_bucket_name=default;auto_create=false", [admin_user]}, {"~s",[verbosity]}], [{env, [{"EVENT_NOSELECT","1"}, {"MEMCACHED_TOP_KEYS","100"}, {"ISASL_PWFILE",{"~s",[{isasl,path}]}}, {"ISASL_DB_CHECK_TIME","1"}]}, use_stdio,stderr_to_stdout,stream]}]}, {alerts, [{email,[]}, {email_alerts,false}, {email_server, [{user,undefined}, {pass,'filtered-out'}, {addr,undefined}, {port,undefined}, {encrypt,false}]}, {alerts, [server_down,server_unresponsive,server_up, server_joined,server_left,bucket_created, bucket_deleted,bucket_auth_failed]}]}, {rebalance_status,running}, {nodes_wanted, [{'_vclock',[{'ns_1@10.1.5.227',{2,63472280023}}]}, 'ns_1@10.1.5.227','ns_1@10.1.5.229']}, {rest, [{'_vclock',[{'ns_1@127.0.0.1',{1,63472279813}}]}, {port,8091}]}, {{node,'ns_1@10.1.5.227',memcached}, [{'_vclock',[{'ns_1@10.1.5.227',{1,63472280022}}]}, {port,11210}, {dbdir,"/var/opt/membase/1.6.5.4r/data/ns_1"}, {admin_user,"_admin"}, {admin_pass,"_admin"}, {bucket_engine,"./bin/bucket_engine/bucket_engine.so"}, {engines, [{membase, [{engine,"bin/ep_engine/ep.so"}, {initfile,"/etc/opt/membase/1.6.5.4r/init.sql"}]}, {memcached, [{engine,"bin/memcached/default_engine.so"}]}]}, {verbosity,[]}]}, {{node,'ns_1@10.1.5.227',ns_log}, [{'_vclock',[{'ns_1@10.1.5.227',{1,63472280022}}]}, {filename, "/var/opt/membase/1.6.5.4r/data/ns_1/ns_log"}]}, {{node,'ns_1@127.0.0.1',isasl}, [{path,"/var/opt/membase/1.6.5.4r/data/ns_1/isasl.pw"}]}, {{node,'ns_1@10.1.5.229',memcached}, [{'_vclock',[{'ns_1@10.1.5.229',{1,63472280023}}]}, {port,11210}, {dbdir,"/var/opt/membase/1.6.5.4r/data/ns_1"}, {admin_user,"_admin"}, {admin_pass,"_admin"}, {bucket_engine,"./bin/bucket_engine/bucket_engine.so"}, {engines, [{membase, [{engine,"bin/ep_engine/ep.so"}, {initfile,"/etc/opt/membase/1.6.5.4r/init.sql"}]}, {memcached, [{engine,"bin/memcached/default_engine.so"}]}]}, {verbosity,[]}]}, {{node,'ns_1@10.1.5.229',ns_log}, [{'_vclock',[{'ns_1@10.1.5.229',{1,63472280023}}]}, {filename, "/var/opt/membase/1.6.5.4r/data/ns_1/ns_log"}]}, {moxi,[{port,11211},{verbosity,[]}]}, {{node,'ns_1@10.1.5.227',membership},active}, {{node,'ns_1@10.1.5.229',membership},active}, {{node,'ns_1@10.1.5.227',isasl}, [{'_vclock',[{'ns_1@10.1.5.227',{1,63472280022}}]}, {path,"/var/opt/membase/1.6.5.4r/data/ns_1/isasl.pw"}]}, {replication,[{enabled,true}]}, {{node,'ns_1@127.0.0.1',memcached}, [{port,11210}, {dbdir,"/var/opt/membase/1.6.5.4r/data/ns_1"}, {admin_user,"_admin"}, {admin_pass,"_admin"}, {bucket_engine,"./bin/bucket_engine/bucket_engine.so"}, {engines, [{membase, [{engine,"bin/ep_engine/ep.so"}, {initfile,"/etc/opt/membase/1.6.5.4r/init.sql"}]}, {memcached, [{engine,"bin/memcached/default_engine.so"}]}]}, {verbosity,[]}]}, {{node,'ns_1@10.1.5.229',isasl}, [{'_vclock',[{'ns_1@10.1.5.229',{1,63472280023}}]}, {path, "/var/opt/membase/1.6.5.4r/data/ns_1/isasl.pw"}]}]}, {basic_info, [{version, [{kernel,"2.13.4"}, {mnesia,"4.4.12"}, {stdlib,"1.16.4"}, {os_mon,"2.2.4"}, {ns_server,"1.6.5.4r"}, {menelaus,"1.6.5.4r"}, {sasl,"2.1.8"}]}, {system_arch,"x86_64-unknown-linux-gnu"}, {wall_clock,1444}, {memory_data, {2058731520,727810048,{<9352.12.0>,4114680}}}, {disk_data, [{"/",11204616,44}, {"/boot",101086,14}, {"/dev/shm",1005240,0}]}]}, {processes, [{<9352.0.0>, [{registered_name,init}, {status,waiting}, {initial_call,{otp_ring0,start,2}}, {backtrace, ["Program counter: 0x00002ba7c3064468 (init:loop/1 + 40)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002aaaaac56e30 Return addr 0x0000000000846c78 ()", "y(0) {state,[{'-root',[<<45 bytes>>]},{'-progname',[<<3 bytes>>]},{'-home',[<<12 bytes"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,87}]}, {heap_size,4181}, {total_heap_size,5778}, {links,[<9352.6.0>,<9352.7.0>,<9352.3.0>]}, {memory,47048}, {message_queue_len,0}, {reductions,34524}, {trap_exit,true}]}, {<9352.3.0>, [{registered_name,erl_prim_loader}, {status,waiting}, {initial_call,{erlang,apply,2}}, {backtrace, ["Program counter: 0x00002ba7c30c3ef8 (erl_prim_loader:loop/3 + 184)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002aaaaac9ac48 Return addr 0x0000000000846c78 ()", "y(0) []", "y(1) [\"/opt/membase/1.6.5.4r/bin/ns_server/ebin\",\"/opt/membase/1.6.5.4r/bin/ns_server/", "y(2) <0.2.0>", "y(3) {state,efile,[],none,#Port<0.2>,infinity,undefined,true,{prim_state,false,undefin", "y(4) infinity"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,324}]}, {heap_size,2584}, {total_heap_size,20295}, {links,[#Port<9352.2>,<9352.0.0>]}, {memory,163144}, {message_queue_len,0}, {reductions,1066890}, {trap_exit,true}]}, {<9352.6.0>, [{registered_name,error_logger}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x00002ba7c30b4c48 (gen_event:fetch_msg/5 + 88)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002aaaaabce510 Return addr 0x00002ba7c30df110 (proc_lib:init_p_do_apply/3 + 56)", "y(0) false","y(1) []", "y(2) [{handler,ns_log_mf_h,false,{state,\"/var/opt/membase/1.6.5.4r/logs\",10485760,10,3", "y(3) error_logger","y(4) <0.2.0>", "0x00002aaaaabce540 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002ba7c30df130 (proc_lib:init_p_do_apply/3 + 88)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,38}]}, {heap_size,4181}, {total_heap_size,8362}, {links,[<9352.0.0>,<9352.25.0>,#Port<9352.1102>]}, {memory,67824}, {message_queue_len,0}, {reductions,91295}, {trap_exit,true}]}, {<9352.7.0>, [{registered_name,application_controller}, {status,waiting}, {initial_call,{erlang,apply,2}}, {backtrace, ["Program counter: 0x00002aaaaaafedf0 (gen_server:loop/6 + 288)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002ba7c31388f8 Return addr 0x0000000000846c78 ()", "y(0) []","y(1) infinity", "y(2) application_controller", "y(3) {state,[],[],[],[{mnesia,<0.389.0>},{ns_server,<0.52.0>},{os_mon,<0.41.0>},{sasl,", "y(4) application_controller", "y(5) <0.2.0>"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,3}]}, {heap_size,4181}, {total_heap_size,21892}, {links, [<9352.41.0>,<9352.52.0>,<9352.389.0>,<9352.9.0>, <9352.32.0>,<9352.0.0>]}, {memory,176184}, {message_queue_len,0}, {reductions,27003}, {trap_exit,true}]}, {<9352.9.0>, [{registered_name,[]}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x00002aaaaaaf6d08 (application_master:main_loop/2 + 64)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002ba7c3110dd0 Return addr 0x00002ba7c30df110 (proc_lib:init_p_do_apply/3 + 56)", "y(0) []", "y(1) {state,<0.10.0>,{appl_data,kernel,[application_controller,erl_reply,auth,boot_ser", "y(2) <0.7.0>", "0x00002ba7c3110df0 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002ba7c30df130 (proc_lib:init_p_do_apply/3 + 88)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,1}]}, {heap_size,377}, {total_heap_size,754}, {links,[<9352.7.0>,<9352.10.0>]}, {memory,6920}, {message_queue_len,0}, {reductions,44}, {trap_exit,true}]}, {<9352.10.0>, [{registered_name,[]}, {status,waiting}, {initial_call,{application_master,start_it,4}}, {backtrace, ["Program counter: 0x00002aaaaaaf9390 (application_master:loop_it/4 + 80)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002ba7c3113e10 Return addr 0x0000000000846c78 ()", "y(0) []","y(1) kernel","y(2) <0.11.0>", "y(3) <0.9.0>"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,0}]}, {heap_size,233}, {total_heap_size,233}, {links,[<9352.9.0>,<9352.11.0>]}, {memory,2648}, {message_queue_len,0}, {reductions,72}, {trap_exit,true}]}, {<9352.11.0>, [{registered_name,kernel_sup}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x00002aaaaaafedf0 (gen_server:loop/6 + 288)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002aaaab080af0 Return addr 0x00002ba7c30df110 (proc_lib:init_p_do_apply/3 + 56)", "y(0) []","y(1) infinity", "y(2) supervisor", "y(3) {state,{local,kernel_sup},one_for_all,[{child,<0.265.0>,net_sup_dynamic,{erl_dist", "y(4) kernel_sup","y(5) <0.10.0>", "0x00002aaaab080b28 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002ba7c30df130 (proc_lib:init_p_do_apply/3 + 88)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,13}]}, {heap_size,1597}, {total_heap_size,1974}, {links, [<9352.22.0>,<9352.26.0>,<9352.27.0>,<9352.265.0>, <9352.24.0>,<9352.17.0>,<9352.20.0>,<9352.21.0>, <9352.19.0>,<9352.12.0>,<9352.13.0>,<9352.10.0>]}, {memory,17080}, {message_queue_len,0}, {reductions,1614}, {trap_exit,true}]}, {<9352.12.0>, [{registered_name,rex}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x00002aaaaaafedf0 (gen_server:loop/6 + 288)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002aaaacdd99f0 Return addr 0x00002ba7c30df110 (proc_lib:init_p_do_apply/3 + 56)", "y(0) []","y(1) infinity","y(2) rpc", "y(3) {1,{<0.1369.0>,{<10272.6876.0>,{#Ref<10272.0.0.150229>,'ns_1@10.1.5.229'}},nil,ni", "y(4) rex","y(5) <0.11.0>", "0x00002aaaacdd9a28 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002ba7c30df130 (proc_lib:init_p_do_apply/3 + 88)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,2}]}, {heap_size,196418}, {total_heap_size,514229}, {links,[<9352.11.0>]}, {memory,4114872}, {message_queue_len,0}, {reductions,9997}, {trap_exit,true}]}, {<9352.13.0>, [{registered_name,global_name_server}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x00002aaaaaafedf0 (gen_server:loop/6 + 288)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002aaaaabf6fa0 Return addr 0x00002ba7c30df110 (proc_lib:init_p_do_apply/3 + 56)", "y(0) []","y(1) infinity","y(2) global", "y(3) {state,true,['ns_1@10.1.5.227'],['ns_1@10.1.5.227'],[],[],'ns_1@10.1.5.229',<0.14", "y(4) global_name_server","y(5) <0.11.0>", "0x00002aaaaabf6fd8 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002ba7c30df130 (proc_lib:init_p_do_apply/3 + 88)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,14}]}, {heap_size,987}, {total_heap_size,1364}, {links, [<9352.14.0>,<9352.16.0>,<9352.15.0>,<9352.11.0>]}, {memory,12088}, {message_queue_len,0}, {reductions,2359}, {trap_exit,true}]}, {<9352.14.0>, [{registered_name,[]}, {status,waiting}, {initial_call,{erlang,apply,2}}, {backtrace, ["Program counter: 0x00002aaaaab85598 (global:loop_the_locker/1 + 1176)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002ba7c314b4c0 Return addr 0x00002aaaaab850c0 (global:init_the_locker/1 + 384)", "y(0) {multi,[],[],['ns_1@10.1.5.227'],'ns_1@10.1.5.229',false,false}", "y(1) infinity", "0x00002ba7c314b4d8 Return addr 0x0000000000846c78 ()"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,3}]}, {heap_size,377}, {total_heap_size,754}, {links,[<9352.13.0>]}, {memory,6776}, {message_queue_len,0}, {reductions,385}, {trap_exit,true}]}, {<9352.15.0>, [{registered_name,[]}, {status,waiting}, {initial_call,{erlang,apply,2}}, {backtrace, ["Program counter: 0x00002aaaaab8d6f0 (global:collect_deletions/2 + 152)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002aaaaabad308 Return addr 0x00002aaaaab8d5f0 (global:loop_the_deleter/1 + 72)", "y(0) infinity","y(1) []", "y(2) <0.13.0>", "0x00002aaaaabad328 Return addr 0x0000000000846c78 ()", "y(0) <0.13.0>"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,3}]}, {heap_size,610}, {total_heap_size,987}, {links,[<9352.13.0>]}, {memory,8640}, {message_queue_len,0}, {reductions,338}, {trap_exit,false}]}, {<9352.16.0>, [{registered_name,[]}, {status,waiting}, {initial_call,{erlang,apply,2}}, {backtrace, ["Program counter: 0x00002aaaaab8d930 (global:loop_the_registrar/0 + 24)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002aaaaac4eb80 Return addr 0x0000000000846c78 ()", "y(0) []"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,3}]}, {heap_size,610}, {total_heap_size,987}, {links,[<9352.13.0>]}, {memory,8640}, {message_queue_len,0}, {reductions,364}, {trap_exit,false}]}, {<9352.17.0>, [{registered_name,inet_db}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x00002aaaaaafedf0 (gen_server:loop/6 + 288)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002ba7c31213d8 Return addr 0x00002ba7c30df110 (proc_lib:init_p_do_apply/3 + 56)", "y(0) []","y(1) infinity","y(2) inet_db", "y(3) {state,inet_db,inet_cache,inet_hosts_byname,inet_hosts_byaddr,inet_hosts_file_byn", "y(4) inet_db","y(5) <0.11.0>", "0x00002ba7c3121410 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002ba7c30df130 (proc_lib:init_p_do_apply/3 + 88)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,1}]}, {heap_size,987}, {total_heap_size,2584}, {links,[<9352.11.0>]}, {memory,21520}, {message_queue_len,0}, {reductions,1467}, {trap_exit,true}]}, {<9352.19.0>, [{registered_name,global_group}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x00002aaaaaafedf0 (gen_server:loop/6 + 288)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002ba7c3189f48 Return addr 0x00002ba7c30df110 (proc_lib:init_p_do_apply/3 + 56)", "y(0) []","y(1) infinity", "y(2) global_group", "y(3) {state,no_conf,true,[],[],[],[],[],'nonode@nohost',[],normal,normal}", "y(4) global_group","y(5) <0.11.0>", "0x00002ba7c3189f80 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002ba7c30df130 (proc_lib:init_p_do_apply/3 + 88)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,0}]}, {heap_size,233}, {total_heap_size,233}, {links,[<9352.11.0>]}, {memory,2712}, {message_queue_len,0}, {reductions,92}, {trap_exit,true}]}, {<9352.20.0>, [{registered_name,file_server_2}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x00002aaaaaafedf0 (gen_server:loop/6 + 288)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002aaaab0c2058 Return addr 0x00002ba7c30df110 (proc_lib:init_p_do_apply/3 + 56)", "y(0) []","y(1) infinity", "y(2) file_server","y(3) #Port<0.262>", "y(4) file_server_2","y(5) <0.11.0>", "0x00002aaaab0c2090 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002ba7c30df130 (proc_lib:init_p_do_apply/3 + 88)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,111}]}, {heap_size,6765}, {total_heap_size,17711}, {links,[#Port<9352.262>,<9352.11.0>]}, {memory,142576}, {message_queue_len,0}, {reductions,636189}, {trap_exit,true}]}, {<9352.21.0>, [{registered_name,code_server}, {status,waiting}, {initial_call,{erlang,apply,2}}, {backtrace, ["Program counter: 0x00002aaaaad22fc8 (code_server:loop/1 + 128)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002aaaab043aa0 Return addr 0x0000000000846c78 ()", "y(0) {state,<0.11.0>,\"/opt/membase/1.6.5.4r/erlang-13b03/lib/erlang\",[\"/opt/membase/1.", "y(1) <0.11.0>"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,269}]}, {heap_size,6765}, {total_heap_size,24476}, {links,[<9352.11.0>]}, {memory,196552}, {message_queue_len,0}, {reductions,264378}, {trap_exit,true}]}, {<9352.22.0>, [{registered_name,standard_error_sup}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x00002aaaaaafedf0 (gen_server:loop/6 + 288)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002ba7c3112af8 Return addr 0x00002ba7c30df110 (proc_lib:init_p_do_apply/3 + 56)", "y(0) []","y(1) infinity", "y(2) supervisor_bridge", "y(3) {state,standard_error,<0.23.0>,<0.23.0>,{local,standard_error_sup}}", "y(4) standard_error_sup","y(5) <0.11.0>", "0x00002ba7c3112b30 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002ba7c30df130 (proc_lib:init_p_do_apply/3 + 88)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,0}]}, {heap_size,233}, {total_heap_size,233}, {links,[<9352.11.0>,<9352.23.0>]}, {memory,2752}, {message_queue_len,0}, {reductions,40}, {trap_exit,true}]}, {<9352.23.0>, [{registered_name,standard_error}, {status,waiting}, {initial_call,{standard_error,server,2}}, {backtrace, ["Program counter: 0x00002aaaaad60458 (standard_error:server_loop/1 + 40)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002ba7c3113280 Return addr 0x0000000000846c78 ()", "y(0) #Port<0.568>"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,0}]}, {heap_size,233}, {total_heap_size,233}, {links,[<9352.22.0>,#Port<9352.568>]}, {memory,2752}, {message_queue_len,0}, {reductions,7}, {trap_exit,true}]}, {<9352.24.0>, [{registered_name,[]}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x00002aaaaaafedf0 (gen_server:loop/6 + 288)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002ba7c31277b8 Return addr 0x00002ba7c30df110 (proc_lib:init_p_do_apply/3 + 56)", "y(0) []","y(1) infinity", "y(2) supervisor_bridge", "y(3) {state,user_sup,<0.25.0>,<0.25.0>,{<0.24.0>,user_sup}}", "y(4) <0.24.0>","y(5) <0.11.0>", "0x00002ba7c31277f0 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002ba7c30df130 (proc_lib:init_p_do_apply/3 + 88)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,0}]}, {heap_size,1597}, {total_heap_size,1597}, {links,[<9352.11.0>,<9352.25.0>]}, {memory,13664}, {message_queue_len,0}, {reductions,71}, {trap_exit,true}]}, {<9352.25.0>, [{registered_name,user}, {status,waiting}, {initial_call,{user,server,2}}, {backtrace, ["Program counter: 0x00002aaaaad77ea0 (user:server_loop/2 + 56)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002aaaab052148 Return addr 0x0000000000846c78 ()", "y(0) []","y(1) []","y(2) {[],[]}", "y(3) #Port<0.587>"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,23}]}, {heap_size,6765}, {total_heap_size,13530}, {links,[<9352.6.0>,<9352.24.0>,#Port<9352.587>]}, {memory,109168}, {message_queue_len,0}, {reductions,48728}, {trap_exit,true}]}, {<9352.26.0>, [{registered_name,[]}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x00002aaaaaafedf0 (gen_server:loop/6 + 288)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002ba7c310d738 Return addr 0x00002ba7c30df110 (proc_lib:init_p_do_apply/3 + 56)", "y(0) []","y(1) infinity", "y(2) kernel_config","y(3) []", "y(4) <0.26.0>","y(5) <0.11.0>", "0x00002ba7c310d770 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002ba7c30df130 (proc_lib:init_p_do_apply/3 + 88)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,0}]}, {heap_size,233}, {total_heap_size,233}, {links,[<9352.11.0>]}, {memory,2712}, {message_queue_len,0}, {reductions,176}, {trap_exit,true}]}, {<9352.27.0>, [{registered_name,kernel_safe_sup}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x00002aaaaaafedf0 (gen_server:loop/6 + 288)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002aaaaac42658 Return addr 0x00002ba7c30df110 (proc_lib:init_p_do_apply/3 + 56)", "y(0) []","y(1) infinity", "y(2) supervisor", "y(3) {state,{local,kernel_safe_sup},one_for_one,[{child,<0.256.0>,inet_gethost_native_", "y(4) kernel_safe_sup","y(5) <0.11.0>", "0x00002aaaaac42690 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002ba7c30df130 (proc_lib:init_p_do_apply/3 + 88)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,4}]}, {heap_size,377}, {total_heap_size,754}, {links, [<9352.130.0>,<9352.134.0>,<9352.256.0>,<9352.133.0>, <9352.50.0>,<9352.129.0>,<9352.11.0>]}, {memory,7120}, {message_queue_len,0}, {reductions,442}, {trap_exit,true}]}, {<9352.32.0>, [{registered_name,[]}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x00002aaaaaaf6d08 (application_master:main_loop/2 + 64)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002ba7c3117000 Return addr 0x00002ba7c30df110 (proc_lib:init_p_do_apply/3 + 56)", "y(0) []", "y(1) {state,<0.33.0>,{appl_data,sasl,[sasl_sup,alarm_handler,overload,release_handler]", "y(2) <0.7.0>", "0x00002ba7c3117020 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002ba7c30df130 (proc_lib:init_p_do_apply/3 + 88)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,0}]}, {heap_size,233}, {total_heap_size,233}, {links,[<9352.7.0>,<9352.33.0>]}, {memory,2752}, {message_queue_len,0}, {reductions,23}, {trap_exit,true}]}, {<9352.33.0>, [{registered_name,[]}, {status,waiting}, {initial_call,{application_master,start_it,4}}, {backtrace, ["Program counter: 0x00002aaaaaaf9390 (application_master:loop_it/4 + 80)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002ba7c3114560 Return addr 0x0000000000846c78 ()", "y(0) {state,tty,undefined}","y(1) sasl", "y(2) <0.34.0>","y(3) <0.32.0>"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,0}]}, {heap_size,233}, {total_heap_size,233}, {links,[<9352.32.0>,<9352.34.0>]}, {memory,2648}, {message_queue_len,0}, {reductions,70}, {trap_exit,true}]}, {<9352.34.0>, [{registered_name,sasl_sup}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x00002aaaaaafedf0 (gen_server:loop/6 + 288)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002ba7c311ca58 Return addr 0x00002ba7c30df110 (proc_lib:init_p_do_apply/3 + 56)", "y(0) []","y(1) infinity", "y(2) supervisor", "y(3) {state,{local,sasl_sup},one_for_one,[{child,<0.38.0>,release_handler,{release_han", "y(4) sasl_sup","y(5) <0.33.0>", "0x00002ba7c311ca90 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002ba7c30df130 (proc_lib:init_p_do_apply/3 + 88)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,1}]}, {heap_size,233}, {total_heap_size,610}, {links,[<9352.35.0>,<9352.38.0>,<9352.33.0>]}, {memory,5808}, {message_queue_len,0}, {reductions,158}, {trap_exit,true}]}, {<9352.35.0>, [{registered_name,sasl_safe_sup}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x00002aaaaaafedf0 (gen_server:loop/6 + 288)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002ba7c318c588 Return addr 0x00002ba7c30df110 (proc_lib:init_p_do_apply/3 + 56)", "y(0) []","y(1) infinity", "y(2) supervisor", "y(3) {state,{local,sasl_safe_sup},one_for_one,[{child,<0.37.0>,overload,{overload,star", "y(4) sasl_safe_sup","y(5) <0.34.0>", "0x00002ba7c318c5c0 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002ba7c30df130 (proc_lib:init_p_do_apply/3 + 88)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,1}]}, {heap_size,233}, {total_heap_size,610}, {links,[<9352.36.0>,<9352.37.0>,<9352.34.0>]}, {memory,5808}, {message_queue_len,0}, {reductions,174}, {trap_exit,true}]}, {<9352.36.0>, [{registered_name,alarm_handler}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x00002ba7c30b4c48 (gen_event:fetch_msg/5 + 88)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002ba7c3115b38 Return addr 0x00002ba7c30df110 (proc_lib:init_p_do_apply/3 + 56)", "y(0) false","y(1) []", "y(2) [{handler,alarm_handler,false,[],false}]", "y(3) alarm_handler","y(4) <0.35.0>", "0x00002ba7c3115b68 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002ba7c30df130 (proc_lib:init_p_do_apply/3 + 88)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,0}]}, {heap_size,233}, {total_heap_size,233}, {links,[<9352.35.0>]}, {memory,2712}, {message_queue_len,0}, {reductions,28}, {trap_exit,true}]}, {<9352.37.0>, [{registered_name,overload}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x00002aaaaaafedf0 (gen_server:loop/6 + 288)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002ba7c314d348 Return addr 0x00002ba7c30df110 (proc_lib:init_p_do_apply/3 + 56)", "y(0) []","y(1) infinity", "y(2) overload", "y(3) {state,0,0,8.000000e-01,151,1.000000e-01,{0,0},clear}", "y(4) overload","y(5) <0.35.0>", "0x00002ba7c314d380 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002ba7c30df130 (proc_lib:init_p_do_apply/3 + 88)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,0}]}, {heap_size,233}, {total_heap_size,233}, {links,[<9352.35.0>]}, {memory,2712}, {message_queue_len,0}, {reductions,39}, {trap_exit,false}]}, {<9352.38.0>, [{registered_name,release_handler}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x00002aaaaaafedf0 (gen_server:loop/6 + 288)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002ba7c3175948 Return addr 0x00002ba7c30df110 (proc_lib:init_p_do_apply/3 + 56)", "y(0) []","y(1) infinity", "y(2) release_handler", "y(3) {state,[],\"/opt/membase/1.6.5.4r/erlang-13b03/lib/erlang\",\"/opt/membase/1.6.5.4r/", "y(4) release_handler","y(5) <0.34.0>", "0x00002ba7c3175980 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002ba7c30df130 (proc_lib:init_p_do_apply/3 + 88)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,2}]}, {heap_size,2584}, {total_heap_size,5168}, {links,[<9352.34.0>]}, {memory,42192}, {message_queue_len,0}, {reductions,2414}, {trap_exit,false}]}, {<9352.41.0>, [{registered_name,[]}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x00002aaaaaaf6d08 (application_master:main_loop/2 + 64)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002ba7c3127f20 Return addr 0x00002ba7c30df110 (proc_lib:init_p_do_apply/3 + 56)", "y(0) []", "y(1) {state,<0.42.0>,{appl_data,os_mon,[os_mon_sup,os_mon_sysinfo,disksup,memsup,cpu_s", "y(2) <0.7.0>", "0x00002ba7c3127f40 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002ba7c30df130 (proc_lib:init_p_do_apply/3 + 88)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,0}]}, {heap_size,233}, {total_heap_size,233}, {links,[<9352.7.0>,<9352.42.0>]}, {memory,2752}, {message_queue_len,0}, {reductions,23}, {trap_exit,true}]}, {<9352.42.0>, [{registered_name,[]}, {status,waiting}, {initial_call,{application_master,start_it,4}}, {backtrace, ["Program counter: 0x00002aaaaaaf9390 (application_master:loop_it/4 + 80)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002ba7c318a6b8 Return addr 0x0000000000846c78 ()", "y(0) []","y(1) os_mon","y(2) <0.43.0>", "y(3) <0.41.0>"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,0}]}, {heap_size,233}, {total_heap_size,233}, {links,[<9352.41.0>,<9352.43.0>]}, {memory,2648}, {message_queue_len,0}, {reductions,40}, {trap_exit,true}]}, {<9352.43.0>, [{registered_name,os_mon_sup}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x00002aaaaaafedf0 (gen_server:loop/6 + 288)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002ba7c311e4c8 Return addr 0x00002ba7c30df110 (proc_lib:init_p_do_apply/3 + 56)", "y(0) []","y(1) infinity", "y(2) supervisor", "y(3) {state,{local,os_mon_sup},one_for_one,[{child,<0.47.0>,cpu_sup,{cpu_sup,start_lin", "y(4) os_mon_sup","y(5) <0.42.0>", "0x00002ba7c311e500 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002ba7c30df130 (proc_lib:init_p_do_apply/3 + 88)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,1}]}, {heap_size,377}, {total_heap_size,754}, {links, [<9352.44.0>,<9352.45.0>,<9352.47.0>,<9352.42.0>]}, {memory,7000}, {message_queue_len,0}, {reductions,273}, {trap_exit,true}]}, {<9352.44.0>, [{registered_name,disksup}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x00002aaaaaafedf0 (gen_server:loop/6 + 288)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002aaaabbd9238 Return addr 0x00002ba7c30df110 (proc_lib:init_p_do_apply/3 + 56)", "y(0) []","y(1) infinity","y(2) disksup", "y(3) {state,80,60000,{unix,linux},[{\"/\",11204616,44},{\"/boot\",101086,14},{\"/dev/shm\",1", "y(4) disksup","y(5) <0.43.0>", "0x00002aaaabbd9270 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002ba7c30df130 (proc_lib:init_p_do_apply/3 + 88)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,41}]}, {heap_size,2584}, {total_heap_size,5168}, {links,[<9352.43.0>,#Port<9352.988>]}, {memory,42232}, {message_queue_len,0}, {reductions,35339}, {trap_exit,true}]}, {<9352.45.0>, [{registered_name,memsup}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x00002aaaaaafedf0 (gen_server:loop/6 + 288)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002aaaaabf1ed8 Return addr 0x00002ba7c30df110 (proc_lib:init_p_do_apply/3 + 56)", "y(0) []","y(1) infinity","y(2) memsup", "y(3) {state,{unix,linux},true,{727810048,2058731520},{<0.12.0>,4114680},false,60000,30", "y(4) memsup","y(5) <0.43.0>", "0x00002aaaaabf1f10 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002ba7c30df130 (proc_lib:init_p_do_apply/3 + 88)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,29}]}, {heap_size,2584}, {total_heap_size,3194}, {links,[<9352.43.0>,<9352.46.0>]}, {memory,26440}, {message_queue_len,0}, {reductions,56760}, {trap_exit,true}]}, {<9352.46.0>, [{registered_name,[]}, {status,waiting}, {initial_call,{erlang,apply,2}}, {backtrace, ["Program counter: 0x00002aaaaaf49880 (memsup:port_idle/1 + 40)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002ba7c312d000 Return addr 0x0000000000846c78 ()", "y(0) []","y(1) #Port<0.1020>"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,40}]}, {heap_size,2584}, {total_heap_size,2961}, {links,[<9352.45.0>,#Port<9352.1020>]}, {memory,24472}, {message_queue_len,0}, {reductions,15741}, {trap_exit,true}]}, {<9352.47.0>, [{registered_name,cpu_sup}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x00002aaaaaafedf0 (gen_server:loop/6 + 288)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002ba7c311d8f8 Return addr 0x00002ba7c30df110 (proc_lib:init_p_do_apply/3 + 56)", "y(0) []","y(1) infinity","y(2) cpu_sup", "y(3) {state,<0.48.0>,{unix,linux}}", "y(4) cpu_sup","y(5) <0.43.0>", "0x00002ba7c311d930 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002ba7c30df130 (proc_lib:init_p_do_apply/3 + 88)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,0}]}, {heap_size,233}, {total_heap_size,233}, {links,[<9352.43.0>]}, {memory,2712}, {message_queue_len,0}, {reductions,33}, {trap_exit,true}]}, {<9352.48.0>, [{registered_name,[]}, {status,waiting}, {initial_call,{erlang,apply,2}}, {backtrace, ["Program counter: 0x00002aaaaaf3a558 (cpu_sup:measurement_server_loop/1 + 40)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002ba7c31865f0 Return addr 0x0000000000846c78 ()", "y(0) []","y(1) []","y(2) []", "y(3) []","y(4) []","y(5) []", "y(6) []","y(7) []", "y(8) {internal,<0.49.0>,[],{unix,linux}}"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,0}]}, {heap_size,233}, {total_heap_size,233}, {links,[<9352.49.0>]}, {memory,2608}, {message_queue_len,0}, {reductions,11}, {trap_exit,true}]}, {<9352.49.0>, [{registered_name,[]}, {status,waiting}, {initial_call,{erlang,apply,2}}, {backtrace, ["Program counter: 0x00002aaaaaf3b730 (cpu_sup:port_server_loop/2 + 64)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002ba7c317fb10 Return addr 0x0000000000846c78 ()", "y(0) []","y(1) 6000", "y(2) #Port<0.1062>"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,3}]}, {heap_size,610}, {total_heap_size,987}, {links,[<9352.48.0>,#Port<9352.1062>]}, {memory,8680}, {message_queue_len,0}, {reductions,335}, {trap_exit,false}]}, {<9352.50.0>, [{registered_name,timer_server}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x00002aaaaaafedf0 (gen_server:loop/6 + 288)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002aaaab07b5d0 Return addr 0x00002ba7c30df110 (proc_lib:init_p_do_apply/3 + 56)", "y(0) []","y(1) 773","y(2) timer", "y(3) []","y(4) timer_server", "y(5) <0.27.0>", "0x00002aaaab07b608 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002ba7c30df130 (proc_lib:init_p_do_apply/3 + 88)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,867}]}, {heap_size,610}, {total_heap_size,987}, {links, [<9352.351.0>,<9352.401.0>,<9352.442.0>,<9352.485.0>, <9352.439.0>,<9352.352.0>,<9352.397.0>,<9352.333.0>, <9352.340.0>,<9352.27.0>]}, {memory,9104}, {message_queue_len,0}, {reductions,487277}, {trap_exit,true}]}, {<9352.52.0>, [{registered_name,[]}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x00002aaaaaaf6d08 (application_master:main_loop/2 + 64)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002ba7c314dab0 Return addr 0x00002ba7c30df110 (proc_lib:init_p_do_apply/3 + 56)", "y(0) []", "y(1) {state,<0.53.0>,{appl_data,ns_server,[ns_server_sup,ns_config,ns_config_sup,ns_co", "y(2) <0.7.0>", "0x00002ba7c314dad0 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002ba7c30df130 (proc_lib:init_p_do_apply/3 + 88)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,1}]}, {heap_size,233}, {total_heap_size,610}, {links,[<9352.7.0>,<9352.53.0>]}, {memory,5768}, {message_queue_len,0}, {reductions,29}, {trap_exit,true}]}, {<9352.53.0>, [{registered_name,[]}, {status,waiting}, {initial_call,{application_master,start_it,4}}, {backtrace, ["Program counter: 0x00002aaaaaaf9390 (application_master:loop_it/4 + 80)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002ba7c314a8e8 Return addr 0x0000000000846c78 ()", "y(0) []","y(1) ns_server", "y(2) <0.54.0>","y(3) <0.52.0>"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,0}]}, {heap_size,233}, {total_heap_size,233}, {links,[<9352.52.0>,<9352.54.0>]}, {memory,2648}, {message_queue_len,0}, {reductions,50}, {trap_exit,true}]}, {<9352.54.0>, [{registered_name,ns_server_cluster_sup}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x00002aaaaaafedf0 (gen_server:loop/6 + 288)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002ba7c317aa10 Return addr 0x00002ba7c30df110 (proc_lib:init_p_do_apply/3 + 56)", "y(0) []","y(1) infinity", "y(2) supervisor", "y(3) {state,{local,ns_server_cluster_sup},one_for_one,[{child,<0.332.0>,ns_server_sup,", "y(4) ns_server_cluster_sup","y(5) <0.53.0>", "0x00002ba7c317aa48 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002ba7c30df130 (proc_lib:init_p_do_apply/3 + 88)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,0}]}, {heap_size,2584}, {total_heap_size,2584}, {links, [<9352.55.0>,<9352.63.0>,<9352.332.0>,<9352.62.0>, <9352.53.0>]}, {memory,21680}, {message_queue_len,0}, {reductions,1906}, {trap_exit,true}]}, {<9352.55.0>, [{registered_name,dist_manager}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x00002aaaaaafedf0 (gen_server:loop/6 + 288)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002aaaaabb65f8 Return addr 0x00002ba7c30df110 (proc_lib:init_p_do_apply/3 + 56)", "y(0) []","y(1) infinity", "y(2) dist_manager", "y(3) {state,true,\"10.1.5.229\"}", "y(4) dist_manager","y(5) <0.54.0>", "0x00002aaaaabb6630 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002ba7c30df130 (proc_lib:init_p_do_apply/3 + 88)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,6}]}, {heap_size,377}, {total_heap_size,754}, {links,[<9352.54.0>]}, {memory,6880}, {message_queue_len,0}, {reductions,601}, {trap_exit,false}]}, {<9352.62.0>, [{registered_name,ns_cluster}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x00002aaaaaafedf0 (gen_server:loop/6 + 288)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002aaaab0b07e0 Return addr 0x00002ba7c30df110 (proc_lib:init_p_do_apply/3 + 56)", "y(0) []","y(1) infinity", "y(2) ns_cluster","y(3) {state}", "y(4) ns_cluster","y(5) <0.54.0>", "0x00002aaaab0b0818 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002ba7c30df130 (proc_lib:init_p_do_apply/3 + 88)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,1}]}, {heap_size,4181}, {total_heap_size,10946}, {links,[<9352.54.0>]}, {memory,88416}, {message_queue_len,0}, {reductions,11570}, {trap_exit,false}]}, {<9352.63.0>, [{registered_name,ns_config_sup}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x00002aaaaaafedf0 (gen_server:loop/6 + 288)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002aaaaabec240 Return addr 0x00002ba7c30df110 (proc_lib:init_p_do_apply/3 + 56)", "y(0) []","y(1) infinity", "y(2) supervisor", "y(3) {state,{local,ns_config_sup},rest_for_one,[{child,undefined,ns_config_log,{ns_con", "y(4) ns_config_sup","y(5) <0.54.0>", "0x00002aaaaabec278 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002ba7c30df130 (proc_lib:init_p_do_apply/3 + 88)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,0}]}, {heap_size,6765}, {total_heap_size,6765}, {links,[<9352.64.0>,<9352.65.0>,<9352.54.0>]}, {memory,55048}, {message_queue_len,0}, {reductions,582}, {trap_exit,true}]}, {<9352.64.0>, [{registered_name,ns_config_events}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x0000000000846c70 (unknown function)", "CP: 0x0000000000846c78 ()", "arity = 3"," proc_lib"," wake_up", " [gen_event,wake_hib,[<0.63.0>,ns_config_events,[{handler,ns_pubsub,#Ref<0.0.0.4124>,{st"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,2}]}, {heap_size,6998}, {total_heap_size,6998}, {links, [<9352.383.0>,<9352.420.0>,<9352.421.0>,<9352.63.0>]}, {memory,57024}, {message_queue_len,0}, {reductions,111738}, {trap_exit,true}]}, {<9352.65.0>, [{registered_name,ns_config}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x00002aaaaaafedf0 (gen_server:loop/6 + 288)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002aaaac6428e0 Return addr 0x00002ba7c30df110 (proc_lib:init_p_do_apply/3 + 56)", "y(0) []","y(1) infinity", "y(2) ns_config", "y(3) {config,{full,\"/etc/opt/membase/1.6.5.4r/config\",undefined,ns_config_default},[[{", "y(4) ns_config","y(5) <0.63.0>", "0x00002aaaac642918 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002ba7c30df130 (proc_lib:init_p_do_apply/3 + 88)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,20}]}, {heap_size,46368}, {total_heap_size,121393}, {links,[<9352.63.0>]}, {memory,971992}, {message_queue_len,0}, {reductions,107735}, {trap_exit,true}]}, {<9352.129.0>, [{registered_name,disk_log_sup}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x00002aaaaaafedf0 (gen_server:loop/6 + 288)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002aaaaac8f7b8 Return addr 0x00002ba7c30df110 (proc_lib:init_p_do_apply/3 + 56)", "y(0) []","y(1) infinity", "y(2) supervisor", "y(3) {state,{local,disk_log_sup},simple_one_for_one,[{child,undefined,disk_log,{disk_l", "y(4) disk_log_sup","y(5) <0.27.0>", "0x00002aaaaac8f7f0 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002ba7c30df130 (proc_lib:init_p_do_apply/3 + 88)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,15}]}, {heap_size,1597}, {total_heap_size,2207}, {links,[<9352.405.0>,<9352.27.0>]}, {memory,18544}, {message_queue_len,0}, {reductions,3774}, {trap_exit,true}]}, {<9352.130.0>, [{registered_name,disk_log_server}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x00002aaaaaafedf0 (gen_server:loop/6 + 288)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002aaaab1935b0 Return addr 0x00002ba7c30df110 (proc_lib:init_p_do_apply/3 + 56)", "y(0) []","y(1) infinity", "y(2) disk_log_server","y(3) {state,[]}", "y(4) disk_log_server","y(5) <0.27.0>", "0x00002aaaab1935e8 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002ba7c30df130 (proc_lib:init_p_do_apply/3 + 88)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,60}]}, {heap_size,1597}, {total_heap_size,2207}, {links,[<9352.405.0>,<9352.27.0>]}, {memory,18544}, {message_queue_len,0}, {reductions,3862}, {trap_exit,true}]}, {<9352.133.0>, [{registered_name,dets_sup}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x00002aaaaaafedf0 (gen_server:loop/6 + 288)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002aaaab162cd8 Return addr 0x00002ba7c30df110 (proc_lib:init_p_do_apply/3 + 56)", "y(0) []","y(1) infinity", "y(2) supervisor", "y(3) {state,{local,dets_sup},simple_one_for_one,[{child,undefined,dets,{dets,istart_li", "y(4) dets_sup","y(5) <0.27.0>", "0x00002aaaab162d10 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002ba7c30df130 (proc_lib:init_p_do_apply/3 + 88)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,8}]}, {heap_size,377}, {total_heap_size,987}, {links,[<9352.27.0>]}, {memory,8744}, {message_queue_len,0}, {reductions,839}, {trap_exit,true}]}, {<9352.134.0>, [{registered_name,dets}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x00002aaaaaafedf0 (gen_server:loop/6 + 288)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002aaaaabfc078 Return addr 0x00002ba7c30df110 (proc_lib:init_p_do_apply/3 + 56)", "y(0) []","y(1) infinity", "y(2) dets_server", "y(3) {state,32802,[<0.27.0>],[]}", "y(4) dets","y(5) <0.27.0>", "0x00002aaaaabfc0b0 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002ba7c30df130 (proc_lib:init_p_do_apply/3 + 88)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,14}]}, {heap_size,377}, {total_heap_size,754}, {links,[<9352.27.0>]}, {memory,6880}, {message_queue_len,0}, {reductions,1187}, {trap_exit,true}]}, {<9352.256.0>, [{registered_name,inet_gethost_native_sup}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x00002aaaaaafedf0 (gen_server:loop/6 + 288)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002ba7c3146b00 Return addr 0x00002ba7c30df110 (proc_lib:init_p_do_apply/3 + 56)", "y(0) []","y(1) infinity", "y(2) supervisor_bridge", "y(3) {state,inet_gethost_native,<0.257.0>,<0.257.0>,{local,inet_gethost_native_sup}}", "y(4) inet_gethost_native_sup", "y(5) <0.27.0>", "0x00002ba7c3146b38 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002ba7c30df130 (proc_lib:init_p_do_apply/3 + 88)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,0}]}, {heap_size,233}, {total_heap_size,233}, {links,[<9352.27.0>,<9352.257.0>]}, {memory,2752}, {message_queue_len,0}, {reductions,41}, {trap_exit,true}]}, {<9352.257.0>, [{registered_name,inet_gethost_native}, {status,waiting}, {initial_call,{inet_gethost_native,server_init,2}}, {backtrace, ["Program counter: 0x00002aaaaacd6440 (inet_gethost_native:main_loop/1 + 40)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002aaaaabdd060 Return addr 0x0000000000846c78 ()", "y(0) {state,#Port<0.2508>,8000,40989,45091,<0.256.0>,4,{statistics,0,0,0,0,0,0,0,0}}"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,2}]}, {heap_size,610}, {total_heap_size,987}, {links,[<9352.256.0>,#Port<9352.2508>]}, {memory,8784}, {message_queue_len,0}, {reductions,252}, {trap_exit,true}]}, {<9352.265.0>, [{registered_name,net_sup}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x00002aaaaaafedf0 (gen_server:loop/6 + 288)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002ba7c31120d8 Return addr 0x00002ba7c30df110 (proc_lib:init_p_do_apply/3 + 56)", "y(0) []","y(1) infinity", "y(2) supervisor", "y(3) {state,{local,net_sup},one_for_all,[{child,<0.268.0>,net_kernel,{net_kernel,start", "y(4) net_sup","y(5) <0.11.0>", "0x00002ba7c3112110 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002ba7c30df130 (proc_lib:init_p_do_apply/3 + 88)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,1}]}, {heap_size,377}, {total_heap_size,987}, {links, [<9352.266.0>,<9352.267.0>,<9352.268.0>,<9352.11.0>]}, {memory,8864}, {message_queue_len,0}, {reductions,244}, {trap_exit,true}]}, {<9352.266.0>, [{registered_name,erl_epmd}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x00002aaaaaafedf0 (gen_server:loop/6 + 288)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002ba7c318ccd8 Return addr 0x00002ba7c30df110 (proc_lib:init_p_do_apply/3 + 56)", "y(0) []","y(1) infinity", "y(2) erl_epmd", "y(3) {state,#Port<0.2549>,21100,ns_1}", "y(4) erl_epmd","y(5) <0.265.0>", "0x00002ba7c318cd10 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002ba7c30df130 (proc_lib:init_p_do_apply/3 + 88)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,0}]}, {heap_size,233}, {total_heap_size,233}, {links,[<9352.265.0>,#Port<9352.2549>]}, {memory,2752}, {message_queue_len,0}, {reductions,127}, {trap_exit,false}]}, {<9352.267.0>, [{registered_name,auth}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x00002aaaaaafedf0 (gen_server:loop/6 + 288)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002ba7c310f1a8 Return addr 0x00002ba7c30df110 (proc_lib:init_p_do_apply/3 + 56)", "y(0) []","y(1) infinity","y(2) auth", "y(3) {state,bujipmggocppuqmx,61457}", "y(4) auth","y(5) <0.265.0>", "0x00002ba7c310f1e0 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002ba7c30df130 (proc_lib:init_p_do_apply/3 + 88)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,3}]}, {heap_size,233}, {total_heap_size,610}, {links,[<9352.265.0>]}, {memory,5728}, {message_queue_len,0}, {reductions,285}, {trap_exit,true}]}, {<9352.268.0>, [{registered_name,net_kernel}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x00002aaaaaafedf0 (gen_server:loop/6 + 288)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002aaaaac5d1d8 Return addr 0x00002ba7c30df110 (proc_lib:init_p_do_apply/3 + 56)", "y(0) []","y(1) infinity", "y(2) net_kernel", "y(3) {state,'ns_1@10.1.5.229','ns_1@10.1.5.229',longnames,{tick,<0.270.0>,15000},7000,", "y(4) net_kernel","y(5) <0.265.0>", "0x00002aaaaac5d210 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002ba7c30df130 (proc_lib:init_p_do_apply/3 + 88)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,13}]}, {heap_size,610}, {total_heap_size,987}, {links, [<9352.265.0>,<9352.270.0>,<9352.328.0>,<9352.269.0>, #Port<9352.2547>]}, {memory,8936}, {message_queue_len,0}, {reductions,1905}, {trap_exit,true}]}, {<9352.269.0>, [{registered_name,[]}, {status,waiting}, {initial_call,{inet_tcp_dist,accept_loop,2}}, {backtrace, ["Program counter: 0x00002ba7c3096d78 (prim_inet:accept0/2 + 184)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002ba7c3114c90 Return addr 0x00002aaaaaf7d900 (inet_tcp:accept/1 + 40)", "y(0) 80","y(1) #Port<0.2547>", "0x00002ba7c3114ca8 Return addr 0x00002aaaaaf81d48 (inet_tcp_dist:accept_loop/2 + 96)", "y(0) []", "0x00002ba7c3114cb8 Return addr 0x0000000000846c78 ()", "y(0) []","y(1) #Port<0.2547>", "y(2) <0.268.0>"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,2}]}, {heap_size,233}, {total_heap_size,610}, {links,[<9352.268.0>]}, {memory,5696}, {message_queue_len,0}, {reductions,525}, {trap_exit,false}]}, {<9352.270.0>, [{registered_name,[]}, {status,waiting}, {initial_call,{net_kernel,ticker,2}}, {backtrace, ["Program counter: 0x00002aaaaad42130 (net_kernel:ticker_loop/2 + 56)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002aaaaac73218 Return addr 0x0000000000846c78 ()", "y(0) 15000","y(1) <0.268.0>"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,0}]}, {heap_size,233}, {total_heap_size,233}, {links,[<9352.268.0>]}, {memory,2608}, {message_queue_len,0}, {reductions,171}, {trap_exit,false}]}, {<9352.328.0>, [{registered_name,[]}, {status,waiting}, {initial_call,{inet_tcp_dist,do_setup,6}}, {backtrace, ["Program counter: 0x00002aaaab90abb0 (dist_util:con_loop/9 + 144)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002aaaaabd16e8 Return addr 0x0000000000846c78 ()", "y(0) []", "y(1) #Fun", "y(2) #Fun", "y(3) {tick,6570,3516,0,4}","y(4) normal", "y(5) 'ns_1@10.1.5.229'", "y(6) {net_address,{{10,1,5,227},21100},\"10.1.5.227\",tcp,inet}", "y(7) #Port<0.2620>", "y(8) 'ns_1@10.1.5.227'","y(9) <0.268.0>"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,10}]}, {heap_size,1597}, {total_heap_size,1974}, {links,[<9352.268.0>,#Port<9352.2620>]}, {memory,16576}, {message_queue_len,0}, {reductions,2368}, {trap_exit,false}]}, {<9352.332.0>, [{registered_name,ns_server_sup}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x00002aaaaaafedf0 (gen_server:loop/6 + 288)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002aaaab416168 Return addr 0x00002ba7c30df110 (proc_lib:init_p_do_apply/3 + 56)", "y(0) []","y(1) infinity", "y(2) supervisor", "y(3) {state,{local,ns_server_sup},one_for_one,[{child,<0.422.0>,ns_tick,{ns_tick,start", "y(4) ns_server_sup","y(5) <0.54.0>", "0x00002aaaab4161a0 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002ba7c30df130 (proc_lib:init_p_do_apply/3 + 88)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,0}]}, {heap_size,121393}, {total_heap_size,121393}, {links, [<9352.358.0>,<9352.382.0>,<9352.419.0>,<9352.421.0>, <9352.422.0>,<9352.420.0>,<9352.385.0>,<9352.386.0>, <9352.383.0>,<9352.380.0>,<9352.381.0>,<9352.370.0>, <9352.335.0>,<9352.351.0>,<9352.352.0>,<9352.338.0>, <9352.333.0>,<9352.334.0>,<9352.54.0>]}, {memory,972712}, {message_queue_len,0}, {reductions,107091}, {trap_exit,true}]}, {<9352.333.0>, [{registered_name,ns_log}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x00002aaaaaafedf0 (gen_server:loop/6 + 288)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002aaaaac4b4d8 Return addr 0x00002ba7c30df110 (proc_lib:init_p_do_apply/3 + 56)", "y(0) []","y(1) infinity","y(2) ns_log", "y(3) {state,[{log_entry,{1305,60644,666628},'ns_1@127.0.0.1',ns_node_disco,3,\"Initial ", "y(4) ns_log","y(5) <0.332.0>", "0x00002aaaaac4b510 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002ba7c30df130 (proc_lib:init_p_do_apply/3 + 88)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,3}]}, {heap_size,4181}, {total_heap_size,10946}, {links,[<9352.50.0>,<9352.332.0>]}, {memory,88456}, {message_queue_len,0}, {reductions,2700}, {trap_exit,false}]}, {<9352.334.0>, [{registered_name,ns_log_events}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x0000000000846c70 (unknown function)", "CP: 0x0000000000846c78 ()", "arity = 3"," proc_lib"," wake_up", " [gen_event,wake_hib,[<0.332.0>,ns_log_events,[{handler,ns_mail_log,false,{state},false}"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,0}]}, {heap_size,34}, {total_heap_size,34}, {links,[<9352.332.0>]}, {memory,1120}, {message_queue_len,0}, {reductions,283}, {trap_exit,true}]}, {<9352.335.0>, [{registered_name,ns_mail_sup}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x00002aaaaaafedf0 (gen_server:loop/6 + 288)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002aaaaabb6d48 Return addr 0x00002ba7c30df110 (proc_lib:init_p_do_apply/3 + 56)", "y(0) []","y(1) infinity", "y(2) supervisor", "y(3) {state,{local,ns_mail_sup},one_for_all,[{child,undefined,ns_mail_log,{ns_mail_log", "y(4) ns_mail_sup","y(5) <0.332.0>", "0x00002aaaaabb6d80 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002ba7c30df130 (proc_lib:init_p_do_apply/3 + 88)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,1}]}, {heap_size,233}, {total_heap_size,610}, {links,[<9352.336.0>,<9352.332.0>]}, {memory,5768}, {message_queue_len,0}, {reductions,652}, {trap_exit,true}]}, {<9352.336.0>, [{registered_name,ns_mail}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x00002aaaaaafedf0 (gen_server:loop/6 + 288)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002ba7c3111508 Return addr 0x00002ba7c30df110 (proc_lib:init_p_do_apply/3 + 56)", "y(0) []","y(1) infinity","y(2) ns_mail", "y(3) empty_state","y(4) ns_mail", "y(5) <0.335.0>", "0x00002ba7c3111540 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002ba7c30df130 (proc_lib:init_p_do_apply/3 + 88)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,0}]}, {heap_size,233}, {total_heap_size,233}, {links,[<9352.335.0>]}, {memory,2712}, {message_queue_len,0}, {reductions,27}, {trap_exit,true}]}, {<9352.338.0>, [{registered_name,ns_node_disco_sup}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x00002aaaaaafedf0 (gen_server:loop/6 + 288)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002ba7c31153e0 Return addr 0x00002ba7c30df110 (proc_lib:init_p_do_apply/3 + 56)", "y(0) []","y(1) infinity", "y(2) supervisor", "y(3) {state,{local,ns_node_disco_sup},rest_for_one,[{child,<0.346.0>,ns_config_rep,{ns", "y(4) ns_node_disco_sup","y(5) <0.332.0>", "0x00002ba7c3115418 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002ba7c30df130 (proc_lib:init_p_do_apply/3 + 88)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,4}]}, {heap_size,233}, {total_heap_size,610}, {links, [<9352.340.0>,<9352.346.0>,<9352.339.0>,<9352.332.0>]}, {memory,5848}, {message_queue_len,0}, {reductions,802}, {trap_exit,true}]}, {<9352.339.0>, [{registered_name,ns_node_disco_events}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x00002ba7c30b4c48 (gen_event:fetch_msg/5 + 88)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002ba7c31463b8 Return addr 0x00002ba7c30df1f0 (proc_lib:wake_up/3 + 120)", "y(0) false","y(1) []", "y(2) [{handler,menelaus_event,ns_node_disco_events,{state,ns_node_disco_events,undefin", "y(3) ns_node_disco_events","y(4) <0.338.0>", "0x00002ba7c31463e8 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002ba7c30df210 (proc_lib:wake_up/3 + 152)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,1}]}, {heap_size,233}, {total_heap_size,322}, {links,[<9352.338.0>]}, {memory,3496}, {message_queue_len,0}, {reductions,60}, {trap_exit,true}]}, {<9352.340.0>, [{registered_name,ns_node_disco}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x00002aaaaaafedf0 (gen_server:loop/6 + 288)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002aaaab9a45d8 Return addr 0x00002ba7c30df110 (proc_lib:init_p_do_apply/3 + 56)", "y(0) []","y(1) infinity", "y(2) ns_node_disco", "y(3) {state,false,{interval,#Ref<0.0.0.3623>}}", "y(4) ns_node_disco","y(5) <0.338.0>", "0x00002aaaab9a4610 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002ba7c30df130 (proc_lib:init_p_do_apply/3 + 88)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,309}]}, {heap_size,17711}, {total_heap_size,46368}, {links,[<9352.338.0>,<9352.50.0>]}, {memory,371832}, {message_queue_len,0}, {reductions,10842}, {trap_exit,false}]}, {<9352.346.0>, [{registered_name,ns_config_rep}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x00002aaaaaafedf0 (gen_server:loop/6 + 288)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002aaaac5af8e0 Return addr 0x00002ba7c30df110 (proc_lib:init_p_do_apply/3 + 56)", "y(0) []","y(1) infinity", "y(2) ns_config_rep","y(3) {state}", "y(4) ns_config_rep","y(5) <0.338.0>", "0x00002aaaac5af918 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002ba7c30df130 (proc_lib:init_p_do_apply/3 + 88)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,20}]}, {heap_size,46368}, {total_heap_size,75025}, {links,[<9352.338.0>]}, {memory,601048}, {message_queue_len,0}, {reductions,23222}, {trap_exit,false}]}, {<9352.351.0>, [{registered_name,ns_heart}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x00002aaaaaafedf0 (gen_server:loop/6 + 288)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002aaaab0cf3c8 Return addr 0x00002ba7c30df110 (proc_lib:init_p_do_apply/3 + 56)", "y(0) []","y(1) infinity", "y(2) ns_heart", "y(3) [{meminfo,<<777 bytes>>},{replication,[{\"default\",1.000000e+00}]},{system_memory_", "y(4) ns_heart","y(5) <0.332.0>", "0x00002aaaab0cf400 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002ba7c30df130 (proc_lib:init_p_do_apply/3 + 88)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,5}]}, {heap_size,6765}, {total_heap_size,35422}, {links,[<9352.332.0>,<9352.50.0>]}, {memory,284264}, {message_queue_len,0}, {reductions,2164694}, {trap_exit,false}]}, {<9352.352.0>, [{registered_name,ns_doctor}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x00002aaaaaafedf0 (gen_server:loop/6 + 288)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002aaaaac80558 Return addr 0x00002ba7c30df110 (proc_lib:init_p_do_apply/3 + 56)", "y(0) []","y(1) infinity", "y(2) ns_doctor", "y(3) {state,{dict,2,16,16,8,80,48,{[],[],[],[],[],[],[],[],[],[],[],[],[],[],[],[]},{{", "y(4) ns_doctor","y(5) <0.332.0>", "0x00002aaaaac80590 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002ba7c30df130 (proc_lib:init_p_do_apply/3 + 88)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,627}]}, {heap_size,6765}, {total_heap_size,13530}, {links,[<9352.332.0>,<9352.50.0>]}, {memory,109128}, {message_queue_len,0}, {reductions,69434}, {trap_exit,false}]}, {<9352.358.0>, [{registered_name,menelaus_sup}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x00002aaaaaafedf0 (gen_server:loop/6 + 288)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002aaaab05c2a0 Return addr 0x00002ba7c30df110 (proc_lib:init_p_do_apply/3 + 56)", "y(0) []","y(1) infinity", "y(2) supervisor", "y(3) {state,{local,menelaus_sup},one_for_one,[{child,<0.485.0>,hot_keys_keeper,{hot_ke", "y(4) menelaus_sup","y(5) <0.332.0>", "0x00002aaaab05c2d8 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002ba7c30df130 (proc_lib:init_p_do_apply/3 + 88)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,1}]}, {heap_size,4181}, {total_heap_size,32838}, {links,[<9352.364.0>,<9352.485.0>,<9352.332.0>]}, {memory,263632}, {message_queue_len,0}, {reductions,1864}, {trap_exit,true}]}, {<9352.364.0>, [{registered_name,menelaus_web}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x00002aaaaaafedf0 (gen_server:loop/6 + 288)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002aaaaacabf20 Return addr 0x00002ba7c30df110 (proc_lib:init_p_do_apply/3 + 56)", "y(0) []","y(1) infinity", "y(2) mochiweb_socket_server", "y(3) {mochiweb_socket_server,8091,#Fun,{local,menelaus_web}", "y(4) menelaus_web","y(5) <0.358.0>", "0x00002aaaaacabf58 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002ba7c30df130 (proc_lib:init_p_do_apply/3 + 88)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,1}]}, {heap_size,377}, {total_heap_size,987}, {links, [<9352.358.0>,<9352.365.0>,<9352.407.0>, #Port<9352.2628>]}, {memory,8864}, {message_queue_len,0}, {reductions,331}, {trap_exit,true}]}, {<9352.365.0>, [{registered_name,[]}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x00002aaaab227430 (menelaus_web:handle_streaming/4 + 392)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002aaaac69d190 Return addr 0x00002aaaab223ae0 (menelaus_web:loop/3 + 24272)", "y(0) {struct,[{buckets,[{struct,[{name,<<7 bytes>>},{nodeLocator,vbucket},{saslPasswor", "y(1) {struct,[{buckets,[{struct,[{name,<<7 bytes>>},{nodeLocator,vbucket},{saslPasswor", "y(2) {mochiweb_response,{mochiweb_request,#Port<0.2640>,'GET',\"/pools/default/saslBuck", "y(3) {mochiweb_request,#Port<0.2640>,'GET',\"/pools/default/saslBucketsStreaming\",{1,1}", "y(4) #Fun", "0x00002aaaac69d1c0 Return addr 0x00002aaaab23cc08 (mochiweb_http:headers/5 + 1360)", "y(0) []","y(1) []","y(2) []", "y(3) []", "y(4) {mochiweb_request,#Port<0.2640>,'GET',\"/pools/default/saslBucketsStreaming\",{1,1}", "y(5) Catch 0x00002aaaab223b00 (menelaus_web:loop/3 + 24304)", "0x00002aaaac69d1f8 Return addr 0x00002ba7c30df110 (proc_lib:init_p_do_apply/3 + 56)", "y(0) #Fun", "y(1) []","y(2) []", "y(3) {mochiweb_request,#Port<0.2640>,'GET',\"/pools/default/saslBucketsStreaming\",{1,1}", "0x00002aaaac69d220 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002ba7c30df130 (proc_lib:init_p_do_apply/3 + 88)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,1}]}, {heap_size,46368}, {total_heap_size,121393}, {links,[<9352.364.0>,#Port<9352.2640>]}, {memory,972176}, {message_queue_len,0}, {reductions,3784016}, {trap_exit,false}]}, {<9352.370.0>, [{registered_name,ns_port_sup}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x00002aaaaaafedf0 (gen_server:loop/6 + 288)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002aaaabb1ef48 Return addr 0x00002ba7c30df110 (proc_lib:init_p_do_apply/3 + 56)", "y(0) []","y(1) infinity", "y(2) supervisor", "y(3) {state,{local,ns_port_sup},one_for_one,[{child,<0.486.0>,{memcached,\"./bin/memcac", "y(4) ns_port_sup","y(5) <0.332.0>", "0x00002aaaabb1ef80 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002ba7c30df130 (proc_lib:init_p_do_apply/3 + 88)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,1}]}, {heap_size,28657}, {total_heap_size,75025}, {links,[<9352.374.0>,<9352.486.0>,<9352.332.0>]}, {memory,601128}, {message_queue_len,0}, {reductions,3027}, {trap_exit,true}]}, {<9352.374.0>, [{registered_name,[]}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x00002aaaaaafedf0 (gen_server:loop/6 + 288)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002aaaaabd7808 Return addr 0x00002ba7c30df110 (proc_lib:init_p_do_apply/3 + 56)", "y(0) []","y(1) infinity", "y(2) supervisor_cushion", "y(3) {state,moxi,5000,{1305,60824,469605},<0.375.0>}", "y(4) <0.374.0>","y(5) <0.370.0>", "0x00002aaaaabd7840 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002ba7c30df130 (proc_lib:init_p_do_apply/3 + 88)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,1}]}, {heap_size,987}, {total_heap_size,2584}, {links,[<9352.370.0>,<9352.375.0>]}, {memory,21560}, {message_queue_len,0}, {reductions,150}, {trap_exit,true}]}, {<9352.375.0>, [{registered_name,[]}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x00002aaaaaafedf0 (gen_server:loop/6 + 288)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002aaaaabb10b0 Return addr 0x00002ba7c30df110 (proc_lib:init_p_do_apply/3 + 56)", "y(0) []","y(1) infinity", "y(2) ns_port_server", "y(3) {state,#Port<0.2629>,moxi,{[\"2011-05-10 13:53:44: (cproxy_config.c.325) env: MOXI", "y(4) <0.375.0>","y(5) <0.374.0>", "0x00002aaaaabb10e8 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002ba7c30df130 (proc_lib:init_p_do_apply/3 + 88)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,1}]}, {heap_size,1597}, {total_heap_size,3194}, {links,[<9352.374.0>,#Port<9352.2629>]}, {memory,26440}, {message_queue_len,0}, {reductions,273}, {trap_exit,true}]}, {<9352.380.0>, [{registered_name,ns_tick_event}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x00002ba7c30b4c48 (gen_event:fetch_msg/5 + 88)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002aaaaac93ca8 Return addr 0x00002ba7c30df110 (proc_lib:init_p_do_apply/3 + 56)", "y(0) false","y(1) []", "y(2) [{handler,ns_pubsub,#Ref<0.0.0.4248>,{state,#Fun,ignored},<", "y(3) ns_tick_event","y(4) <0.332.0>", "0x00002aaaaac93cd8 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002ba7c30df130 (proc_lib:init_p_do_apply/3 + 88)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,25}]}, {heap_size,1597}, {total_heap_size,1974}, {links,[<9352.332.0>,<9352.437.0>]}, {memory,16680}, {message_queue_len,0}, {reductions,13954}, {trap_exit,true}]}, {<9352.381.0>, [{registered_name,ns_stats_event}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x00002ba7c30b4c48 (gen_event:fetch_msg/5 + 88)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002aaaab085760 Return addr 0x00002ba7c30df110 (proc_lib:init_p_do_apply/3 + 56)", "y(0) false","y(1) []", "y(2) [{handler,ns_pubsub,#Ref<0.0.0.4472>,{state,#Fun,ignored},<", "y(3) ns_stats_event","y(4) <0.332.0>", "0x00002aaaab085790 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002ba7c30df130 (proc_lib:init_p_do_apply/3 + 88)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,2511}]}, {heap_size,377}, {total_heap_size,987}, {links,[<9352.332.0>,<9352.442.0>]}, {memory,8784}, {message_queue_len,0}, {reductions,27661}, {trap_exit,true}]}, {<9352.382.0>, [{registered_name,ns_good_bucket_worker}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x00002aaaaaafedf0 (gen_server:loop/6 + 288)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002aaaab083128 Return addr 0x00002ba7c30df110 (proc_lib:init_p_do_apply/3 + 56)", "y(0) []","y(1) infinity", "y(2) work_queue","y(3) []", "y(4) ns_good_bucket_worker", "y(5) <0.332.0>", "0x00002aaaab083160 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002ba7c30df130 (proc_lib:init_p_do_apply/3 + 88)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,0}]}, {heap_size,233}, {total_heap_size,233}, {links,[<9352.332.0>]}, {memory,2712}, {message_queue_len,0}, {reductions,95}, {trap_exit,false}]}, {<9352.383.0>, [{registered_name,ns_good_bucket_sup}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x00002aaaaaafedf0 (gen_server:loop/6 + 288)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002aaaabad19a8 Return addr 0x00002ba7c30df110 (proc_lib:init_p_do_apply/3 + 56)", "y(0) []","y(1) infinity", "y(2) supervisor", "y(3) {state,{local,ns_good_bucket_sup},one_for_one,[{child,<0.439.0>,{ns_memcached,\"de", "y(4) ns_good_bucket_sup","y(5) <0.332.0>", "0x00002aaaabad19e0 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002ba7c30df130 (proc_lib:init_p_do_apply/3 + 88)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,0}]}, {heap_size,17711}, {total_heap_size,17711}, {links, [<9352.332.0>,<9352.438.0>,<9352.439.0>,<9352.64.0>]}, {memory,142656}, {message_queue_len,0}, {reductions,214}, {trap_exit,true}]}, {<9352.385.0>, [{registered_name,[]}, {status,waiting}, {initial_call,{erlang,apply,2}}, {backtrace, ["Program counter: 0x00002aaaaaf6a300 (misc:wait_for_process/2 + 208)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002ba7c310de98 Return addr 0x00002aaaaaf6fe08 (misc:'-start_singleton/4-fun-0-'/2 + 11", "y(0) []","y(1) #Ref<0.0.0.3902>", "y(2) infinity", "0x00002ba7c310deb8 Return addr 0x0000000000846c78 ()", "y(0) <10272.109.0>","y(1) ns_orchestrator"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,0}]}, {heap_size,233}, {total_heap_size,233}, {links,[<9352.332.0>]}, {memory,2680}, {message_queue_len,0}, {reductions,6}, {trap_exit,false}]}, {<9352.386.0>, [{registered_name,ns_mnesia}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x00002aaaaaafedf0 (gen_server:loop/6 + 288)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002aaaab05f490 Return addr 0x00002ba7c30df110 (proc_lib:init_p_do_apply/3 + 56)", "y(0) []","y(1) infinity", "y(2) ns_mnesia","y(3) {state}", "y(4) ns_mnesia","y(5) <0.332.0>", "0x00002aaaab05f4c8 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002ba7c30df130 (proc_lib:init_p_do_apply/3 + 88)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,6}]}, {heap_size,1597}, {total_heap_size,2207}, {links,[<9352.395.0>,<9352.332.0>]}, {memory,18544}, {message_queue_len,0}, {reductions,1401}, {trap_exit,true}]}, {<9352.387.0>, [{registered_name,[]}, {status,waiting}, {initial_call,{erlang,apply,2}}, {backtrace, ["Program counter: 0x00002aaaaaf70698 (misc:'-wait_for_process/2-fun-0-'/3 + 224)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002aaaaac72670 Return addr 0x0000000000846c78 ()", "y(0) []","y(1) <0.385.0>", "y(2) #Ref<0.0.0.3902>"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,0}]}, {heap_size,233}, {total_heap_size,233}, {links,[<0.109.0>]}, {memory,2712}, {message_queue_len,0}, {reductions,13}, {trap_exit,true}]}, {<9352.389.0>, [{registered_name,[]}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x00002aaaaaaf6d08 (application_master:main_loop/2 + 64)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002aaaab07d4b8 Return addr 0x00002ba7c30df110 (proc_lib:init_p_do_apply/3 + 56)", "y(0) []", "y(1) {state,<0.390.0>,{appl_data,mnesia,[mnesia_dumper_load_regulator,mnesia_event,mne", "y(2) <0.7.0>", "0x00002aaaab07d4d8 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002ba7c30df130 (proc_lib:init_p_do_apply/3 + 88)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,5}]}, {heap_size,233}, {total_heap_size,610}, {links,[<9352.7.0>,<9352.390.0>]}, {memory,5768}, {message_queue_len,0}, {reductions,45}, {trap_exit,true}]}, {<9352.390.0>, [{registered_name,[]}, {status,waiting}, {initial_call,{application_master,start_it,4}}, {backtrace, ["Program counter: 0x00002aaaaaaf9390 (application_master:loop_it/4 + 80)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002ba7c314bc08 Return addr 0x0000000000846c78 ()", "y(0) {normal,[]}","y(1) mnesia_sup", "y(2) <0.391.0>","y(3) <0.389.0>"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,0}]}, {heap_size,233}, {total_heap_size,233}, {links,[<9352.389.0>,<9352.391.0>]}, {memory,2648}, {message_queue_len,0}, {reductions,32}, {trap_exit,true}]}, {<9352.391.0>, [{registered_name,mnesia_sup}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x00002aaaaaafedf0 (gen_server:loop/6 + 288)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002aaaab161538 Return addr 0x00002ba7c30df110 (proc_lib:init_p_do_apply/3 + 56)", "y(0) []","y(1) infinity", "y(2) supervisor", "y(3) {state,{local,mnesia_sup},one_for_all,[{child,<0.393.0>,mnesia_kernel_sup,{mnesia", "y(4) mnesia_sup","y(5) <0.390.0>", "0x00002aaaab161570 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002ba7c30df130 (proc_lib:init_p_do_apply/3 + 88)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,1}]}, {heap_size,377}, {total_heap_size,754}, {links,[<9352.392.0>,<9352.393.0>,<9352.390.0>]}, {memory,6960}, {message_queue_len,0}, {reductions,191}, {trap_exit,true}]}, {<9352.392.0>, [{registered_name,mnesia_event}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x00002ba7c30b4c48 (gen_event:fetch_msg/5 + 88)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002aaaab0b4cf0 Return addr 0x00002ba7c30df110 (proc_lib:init_p_do_apply/3 + 56)", "y(0) false","y(1) []", "y(2) [{handler,mnesia_event,false,{state,[],false,[]},false}]", "y(3) mnesia_event","y(4) <0.391.0>", "0x00002aaaab0b4d20 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002ba7c30df130 (proc_lib:init_p_do_apply/3 + 88)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,7}]}, {heap_size,610}, {total_heap_size,987}, {links,[<9352.391.0>,<9352.395.0>]}, {memory,8784}, {message_queue_len,0}, {reductions,402}, {trap_exit,true}]}, {<9352.393.0>, [{registered_name,mnesia_kernel_sup}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x00002aaaaaafedf0 (gen_server:loop/6 + 288)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002aaaab07cd50 Return addr 0x00002ba7c30df110 (proc_lib:init_p_do_apply/3 + 56)", "y(0) []","y(1) infinity", "y(2) supervisor", "y(3) {state,{local,mnesia_kernel_sup},one_for_all,[{child,<0.402.0>,mnesia_late_loader", "y(4) mnesia_kernel_sup","y(5) <0.391.0>", "0x00002aaaab07cd88 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002ba7c30df130 (proc_lib:init_p_do_apply/3 + 88)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,4}]}, {heap_size,610}, {total_heap_size,987}, {links, [<9352.396.0>,<9352.400.0>,<9352.401.0>,<9352.402.0>, <9352.398.0>,<9352.399.0>,<9352.397.0>,<9352.394.0>, <9352.395.0>,<9352.391.0>]}, {memory,9104}, {message_queue_len,0}, {reductions,545}, {trap_exit,true}]}, {<9352.394.0>, [{registered_name,mnesia_monitor}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x00002aaaaaafedf0 (gen_server:loop/6 + 288)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002aaaaac85620 Return addr 0x00002ba7c30df110 (proc_lib:init_p_do_apply/3 + 56)", "y(0) []","y(1) infinity", "y(2) mnesia_monitor", "y(3) {state,<0.393.0>,[],[],true,[],undefined,[]}", "y(4) mnesia_monitor","y(5) <0.393.0>", "0x00002aaaaac85658 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002ba7c30df130 (proc_lib:init_p_do_apply/3 + 88)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,25}]}, {heap_size,2584}, {total_heap_size,3194}, {links,[<9352.405.0>,<9352.393.0>]}, {memory,26440}, {message_queue_len,0}, {reductions,5324}, {trap_exit,true}]}, {<9352.395.0>, [{registered_name,mnesia_subscr}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x00002aaaaaafedf0 (gen_server:loop/6 + 288)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002aaaab0441e8 Return addr 0x00002ba7c30df110 (proc_lib:init_p_do_apply/3 + 56)", "y(0) []","y(1) infinity", "y(2) mnesia_subscr", "y(3) {state,<0.393.0>,86042}", "y(4) mnesia_subscr","y(5) <0.393.0>", "0x00002aaaab044220 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002ba7c30df130 (proc_lib:init_p_do_apply/3 + 88)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,0}]}, {heap_size,233}, {total_heap_size,233}, {links,[<9352.392.0>,<9352.393.0>,<9352.386.0>]}, {memory,2792}, {message_queue_len,0}, {reductions,111}, {trap_exit,true}]}, {<9352.396.0>, [{registered_name,mnesia_locker}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x00002aaaab504b60 (mnesia_locker:loop/1 + 40)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002aaaaabf3d88 Return addr 0x00002aaaaaeb3e80 (mnesia_sp:init_proc/4 + 264)", "y(0) []","y(1) []","y(2) []", "y(3) []","y(4) []", "y(5) {state,<0.393.0>}", "0x00002aaaaabf3dc0 Return addr 0x00002ba7c30df110 (proc_lib:init_p_do_apply/3 + 56)", "y(0) Catch 0x00002aaaaaeb3e80 (mnesia_sp:init_proc/4 + 264)", "y(1) mnesia_locker","y(2) []","y(3) []", "y(4) [<0.393.0>]", "0x00002aaaaabf3df0 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002ba7c30df130 (proc_lib:init_p_do_apply/3 + 88)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,452}]}, {heap_size,987}, {total_heap_size,1364}, {links,[<9352.393.0>]}, {memory,11760}, {message_queue_len,0}, {reductions,189967}, {trap_exit,true}]}, {<9352.397.0>, [{registered_name,mnesia_recover}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x00002aaaaaafedf0 (gen_server:loop/6 + 288)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002ba7c317e7d0 Return addr 0x00002ba7c30df110 (proc_lib:init_p_do_apply/3 + 56)", "y(0) []","y(1) infinity", "y(2) mnesia_recover", "y(3) {state,<0.393.0>,undefined,undefined,undefined,0,true,[]}", "y(4) mnesia_recover","y(5) <0.393.0>", "0x00002ba7c317e808 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002ba7c30df130 (proc_lib:init_p_do_apply/3 + 88)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,7}]}, {heap_size,1597}, {total_heap_size,1974}, {links,[<9352.393.0>,<9352.50.0>]}, {memory,16680}, {message_queue_len,0}, {reductions,2612}, {trap_exit,true}]}, {<9352.398.0>, [{registered_name,mnesia_tm}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x00002aaaab619a80 (mnesia_tm:doit_loop/1 + 216)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002ba7c3184b70 Return addr 0x00002aaaaaeb3e80 (mnesia_sp:init_proc/4 + 264)", "y(0) []","y(1) []", "y(2) {state,{0,nil},{0,nil},<0.393.0>,[],[],[]}", "y(3) []","y(4) []","y(5) <0.393.0>", "y(6) {0,nil}","y(7) {0,nil}", "0x00002ba7c3184bb8 Return addr 0x00002ba7c30df110 (proc_lib:init_p_do_apply/3 + 56)", "y(0) Catch 0x00002aaaaaeb3e80 (mnesia_sp:init_proc/4 + 264)", "y(1) mnesia_tm","y(2) []","y(3) []", "y(4) [<0.393.0>]", "0x00002ba7c3184be8 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002ba7c30df130 (proc_lib:init_p_do_apply/3 + 88)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,94}]}, {heap_size,2584}, {total_heap_size,3194}, {links,[<9352.393.0>]}, {memory,26400}, {message_queue_len,0}, {reductions,432534}, {trap_exit,true}]}, {<9352.399.0>, [{registered_name,mnesia_checkpoint_sup}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x00002aaaaaafedf0 (gen_server:loop/6 + 288)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002ba7c3117e88 Return addr 0x00002ba7c30df110 (proc_lib:init_p_do_apply/3 + 56)", "y(0) []","y(1) infinity", "y(2) supervisor", "y(3) {state,{local,mnesia_checkpoint_sup},simple_one_for_one,[{child,undefined,mnesia_", "y(4) mnesia_checkpoint_sup", "y(5) <0.393.0>", "0x00002ba7c3117ec0 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002ba7c30df130 (proc_lib:init_p_do_apply/3 + 88)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,0}]}, {heap_size,233}, {total_heap_size,233}, {links,[<9352.393.0>]}, {memory,2712}, {message_queue_len,0}, {reductions,61}, {trap_exit,true}]}, {<9352.400.0>, [{registered_name,mnesia_snmp_sup}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x00002aaaaaafedf0 (gen_server:loop/6 + 288)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002ba7c314c338 Return addr 0x00002ba7c30df110 (proc_lib:init_p_do_apply/3 + 56)", "y(0) []","y(1) infinity", "y(2) supervisor", "y(3) {state,{local,mnesia_snmp_sup},simple_one_for_one,[{child,undefined,mnesia_snmp_s", "y(4) mnesia_snmp_sup","y(5) <0.393.0>", "0x00002ba7c314c370 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002ba7c30df130 (proc_lib:init_p_do_apply/3 + 88)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,0}]}, {heap_size,233}, {total_heap_size,233}, {links,[<9352.393.0>]}, {memory,2712}, {message_queue_len,0}, {reductions,61}, {trap_exit,true}]}, {<9352.401.0>, [{registered_name,mnesia_controller}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x00002aaaaaafedf0 (gen_server:loop/6 + 288)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002aaaaabb5a28 Return addr 0x00002ba7c30df110 (proc_lib:init_p_do_apply/3 + 56)", "y(0) []","y(1) infinity", "y(2) mnesia_controller", "y(3) {state,<0.393.0>,true,[],[],{0,nil},[],[],{0,nil},undefined,[],[],{interval,#Ref<", "y(4) mnesia_controller","y(5) <0.393.0>", "0x00002aaaaabb5a60 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002ba7c30df130 (proc_lib:init_p_do_apply/3 + 88)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,4}]}, {heap_size,377}, {total_heap_size,754}, {links,[<9352.393.0>,<9352.50.0>]}, {memory,6920}, {message_queue_len,0}, {reductions,679}, {trap_exit,true}]}, {<9352.402.0>, [{registered_name,mnesia_late_loader}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x00002aaaab671de8 (mnesia_late_loader:loop/1 + 40)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002aaaab084b70 Return addr 0x00002aaaaaeb3e80 (mnesia_sp:init_proc/4 + 264)", "y(0) []","y(1) []", "y(2) {state,<0.393.0>}", "0x00002aaaab084b90 Return addr 0x00002ba7c30df110 (proc_lib:init_p_do_apply/3 + 56)", "y(0) Catch 0x00002aaaaaeb3e80 (mnesia_sp:init_proc/4 + 264)", "y(1) mnesia_late_loader","y(2) []", "y(3) []","y(4) [<0.393.0>]", "0x00002aaaab084bc0 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002ba7c30df130 (proc_lib:init_p_do_apply/3 + 88)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,1}]}, {heap_size,233}, {total_heap_size,610}, {links,[<9352.393.0>]}, {memory,5728}, {message_queue_len,0}, {reductions,174}, {trap_exit,false}]}, {<9352.405.0>, [{registered_name,[]}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x00002aaaab6e93b8 (disk_log:loop/1 + 168)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002aaaaabfb4d0 Return addr 0x00002ba7c30df110 (proc_lib:init_p_do_apply/3 + 56)", "y(0) {state,[],[],<0.129.0>,<0.130.0>,12,{arg,latest_log,undefined,\"/var/opt/membase/1", "0x00002aaaaabfb4e0 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002ba7c30df130 (proc_lib:init_p_do_apply/3 + 88)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,3}]}, {heap_size,1597}, {total_heap_size,3194}, {links, [<9352.130.0>,<9352.394.0>,<9352.129.0>, #Port<9352.3038>]}, {memory,26520}, {message_queue_len,0}, {reductions,138645}, {trap_exit,true}]}, {<9352.407.0>, [{registered_name,[]}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x00002ba7c3096d78 (prim_inet:accept0/2 + 184)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002aaaab0b3280 Return addr 0x00002aaaaaf7d900 (inet_tcp:accept/1 + 40)", "y(0) 90","y(1) #Port<0.2628>", "0x00002aaaab0b3298 Return addr 0x00002aaaab24b018 (mochiweb_socket_server:acceptor_loop/1 ", "y(0) []", "0x00002aaaab0b32a8 Return addr 0x00002ba7c30df110 (proc_lib:init_p_do_apply/3 + 56)", "y(0) #Fun", "y(1) <0.364.0>", "y(2) Catch 0x00002aaaab24b018 (mochiweb_socket_server:acceptor_loop/1 + 160)", "0x00002aaaab0b32c8 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002ba7c30df130 (proc_lib:init_p_do_apply/3 + 88)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,1}]}, {heap_size,377}, {total_heap_size,754}, {links,[<9352.364.0>]}, {memory,6952}, {message_queue_len,0}, {reductions,39}, {trap_exit,false}]}, {<9352.419.0>, [{registered_name,ns_bad_bucket_worker}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x00002aaaaaafedf0 (gen_server:loop/6 + 288)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002ba7c3117738 Return addr 0x00002ba7c30df110 (proc_lib:init_p_do_apply/3 + 56)", "y(0) []","y(1) infinity", "y(2) work_queue","y(3) []", "y(4) ns_bad_bucket_worker","y(5) <0.332.0>", "0x00002ba7c3117770 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002ba7c30df130 (proc_lib:init_p_do_apply/3 + 88)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,1}]}, {heap_size,233}, {total_heap_size,610}, {links,[<9352.332.0>]}, {memory,5728}, {message_queue_len,0}, {reductions,129}, {trap_exit,false}]}, {<9352.420.0>, [{registered_name,ns_bad_bucket_sup}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x00002aaaaaafedf0 (gen_server:loop/6 + 288)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002aaaab93c958 Return addr 0x00002ba7c30df110 (proc_lib:init_p_do_apply/3 + 56)", "y(0) []","y(1) infinity", "y(2) supervisor", "y(3) {state,{local,ns_bad_bucket_sup},one_for_one,[{child,<0.480.0>,{stats_reader,\"def", "y(4) ns_bad_bucket_sup","y(5) <0.332.0>", "0x00002aaaab93c990 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002ba7c30df130 (proc_lib:init_p_do_apply/3 + 88)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,0}]}, {heap_size,17711}, {total_heap_size,17711}, {links, [<9352.332.0>,<9352.442.0>,<9352.480.0>,<9352.437.0>, <9352.64.0>]}, {memory,142696}, {message_queue_len,0}, {reductions,301}, {trap_exit,true}]}, {<9352.421.0>, [{registered_name,ns_moxi_sup}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x00002aaaaaafedf0 (gen_server:loop/6 + 288)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002aaaab95f2d8 Return addr 0x00002ba7c30df110 (proc_lib:init_p_do_apply/3 + 56)", "y(0) []","y(1) infinity", "y(2) supervisor", "y(3) {state,{local,ns_moxi_sup},one_for_one,[],{dict,0,16,16,8,80,48,{[],[],[],[],[],[", "y(4) ns_moxi_sup","y(5) <0.332.0>", "0x00002aaaab95f310 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002ba7c30df130 (proc_lib:init_p_do_apply/3 + 88)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,0}]}, {heap_size,17711}, {total_heap_size,17711}, {links,[<9352.332.0>,<9352.64.0>]}, {memory,142576}, {message_queue_len,0}, {reductions,649}, {trap_exit,true}]}, {<9352.422.0>, [{registered_name,[]}, {status,waiting}, {initial_call,{erlang,apply,2}}, {backtrace, ["Program counter: 0x00002aaaaaf6a300 (misc:wait_for_process/2 + 208)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002aaaaac4d3d8 Return addr 0x00002aaaaaf6fe08 (misc:'-start_singleton/4-fun-0-'/2 + 11", "y(0) []","y(1) #Ref<0.0.0.4128>", "y(2) infinity", "0x00002aaaaac4d3f8 Return addr 0x0000000000846c78 ()", "y(0) <10272.473.0>","y(1) ns_tick"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,0}]}, {heap_size,233}, {total_heap_size,233}, {links,[<9352.332.0>]}, {memory,2680}, {message_queue_len,0}, {reductions,6}, {trap_exit,false}]}, {<9352.423.0>, [{registered_name,[]}, {status,waiting}, {initial_call,{erlang,apply,2}}, {backtrace, ["Program counter: 0x00002aaaaaf70698 (misc:'-wait_for_process/2-fun-0-'/3 + 224)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002ba7c31301f8 Return addr 0x0000000000846c78 ()", "y(0) []","y(1) <0.422.0>", "y(2) #Ref<0.0.0.4128>"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,0}]}, {heap_size,233}, {total_heap_size,233}, {links,[<0.473.0>]}, {memory,2712}, {message_queue_len,0}, {reductions,13}, {trap_exit,true}]}, {<9352.437.0>, [{registered_name,[]}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x00002aaaaaafedf0 (gen_server:loop/6 + 288)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002aaaaac3b6a8 Return addr 0x00002ba7c30df110 (proc_lib:init_p_do_apply/3 + 56)", "y(0) []","y(1) infinity", "y(2) stats_collector", "y(3) {state,\"default\",[434575,6873845,0,0,0,0,0,0,0,0,0,0,0,0,0,0,1718,0,0,0,0,0,0,0,0", "y(4) <0.437.0>","y(5) <0.420.0>", "0x00002aaaaac3b6e0 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002ba7c30df130 (proc_lib:init_p_do_apply/3 + 88)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,14}]}, {heap_size,6765}, {total_heap_size,17711}, {links,[<9352.420.0>,<9352.380.0>]}, {memory,142576}, {message_queue_len,0}, {reductions,8849545}, {trap_exit,false}]}, {<9352.438.0>, [{registered_name,'ns_vbm_sup-default'}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x00002aaaaaafedf0 (gen_server:loop/6 + 288)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002ba7c314ca88 Return addr 0x00002ba7c30df110 (proc_lib:init_p_do_apply/3 + 56)", "y(0) []","y(1) infinity", "y(2) supervisor", "y(3) {state,{local,'ns_vbm_sup-default'},one_for_one,[],{dict,0,16,16,8,80,48,{[],[],[", "y(4) 'ns_vbm_sup-default'","y(5) <0.383.0>", "0x00002ba7c314cac0 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002ba7c30df130 (proc_lib:init_p_do_apply/3 + 88)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,0}]}, {heap_size,233}, {total_heap_size,233}, {links,[<9352.383.0>]}, {memory,2712}, {message_queue_len,0}, {reductions,583}, {trap_exit,true}]}, {<9352.439.0>, [{registered_name,'ns_memcached-default'}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x00002aaaaaafedf0 (gen_server:loop/6 + 288)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002aaaab9ac888 Return addr 0x00002ba7c30df110 (proc_lib:init_p_do_apply/3 + 56)", "y(0) []","y(1) infinity", "y(2) ns_memcached", "y(3) {state,\"default\",#Port<0.2772>}", "y(4) <0.439.0>","y(5) <0.383.0>", "0x00002aaaab9ac8c0 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002ba7c30df130 (proc_lib:init_p_do_apply/3 + 88)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,38}]}, {heap_size,4181}, {total_heap_size,32838}, {links,[<9352.50.0>,<9352.383.0>,#Port<9352.2772>]}, {memory,263632}, {message_queue_len,0}, {reductions,7410406}, {trap_exit,true}]}, {<9352.442.0>, [{registered_name,'stats_archiver-default'}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x00002aaaaaafedf0 (gen_server:loop/6 + 288)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002aaaabbe65a8 Return addr 0x00002ba7c30df110 (proc_lib:init_p_do_apply/3 + 56)", "y(0) []","y(1) infinity", "y(2) stats_archiver", "y(3) {state,\"default\"}", "y(4) 'stats_archiver-default'", "y(5) <0.420.0>", "0x00002aaaabbe65e0 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002ba7c30df130 (proc_lib:init_p_do_apply/3 + 88)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,183}]}, {heap_size,6765}, {total_heap_size,24476}, {links,[<9352.381.0>,<9352.420.0>,<9352.50.0>]}, {memory,196736}, {message_queue_len,0}, {reductions,3879725}, {trap_exit,false}]}, {<9352.480.0>, [{registered_name,'stats_reader-default'}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x00002aaaaaafedf0 (gen_server:loop/6 + 288)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002aaaaac63120 Return addr 0x00002ba7c30df110 (proc_lib:init_p_do_apply/3 + 56)", "y(0) []","y(1) infinity", "y(2) stats_reader", "y(3) {state,\"default\"}", "y(4) 'stats_reader-default'", "y(5) <0.420.0>", "0x00002aaaaac63158 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002ba7c30df130 (proc_lib:init_p_do_apply/3 + 88)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,87}]}, {heap_size,2584}, {total_heap_size,3571}, {links,[<9352.420.0>]}, {memory,29416}, {message_queue_len,0}, {reductions,107218}, {trap_exit,false}]}, {<9352.485.0>, [{registered_name,hot_keys_keeper}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x00002aaaaaafedf0 (gen_server:loop/6 + 288)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002aaaaac5bec0 Return addr 0x00002ba7c30df110 (proc_lib:init_p_do_apply/3 + 56)", "y(0) []","y(1) infinity", "y(2) hot_keys_keeper", "y(3) {state,[{\"default\",[{\"b7aedb5d-898e-4666-9727-1265f77c4d24-10-0\",[{ops,7.955449e-", "y(4) hot_keys_keeper","y(5) <0.358.0>", "0x00002aaaaac5bef8 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002ba7c30df130 (proc_lib:init_p_do_apply/3 + 88)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,249}]}, {heap_size,2584}, {total_heap_size,6765}, {links,[<9352.358.0>,<9352.50.0>]}, {memory,55008}, {message_queue_len,0}, {reductions,7836}, {trap_exit,false}]}, {<9352.486.0>, [{registered_name,[]}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x00002aaaaaafedf0 (gen_server:loop/6 + 288)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002ba7c311b080 Return addr 0x00002ba7c30df110 (proc_lib:init_p_do_apply/3 + 56)", "y(0) []","y(1) infinity", "y(2) supervisor_cushion", "y(3) {state,memcached,5000,{1305,60829,522207},<0.487.0>}", "y(4) <0.486.0>","y(5) <0.370.0>", "0x00002ba7c311b0b8 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002ba7c30df130 (proc_lib:init_p_do_apply/3 + 88)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,0}]}, {heap_size,987}, {total_heap_size,987}, {links,[<9352.370.0>,<9352.487.0>]}, {memory,8784}, {message_queue_len,0}, {reductions,55}, {trap_exit,true}]}, {<9352.487.0>, [{registered_name,[]}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x00002aaaaaafedf0 (gen_server:loop/6 + 288)", "CP: 0x0000000000000000 (invalid)","arity = 0", "0x00002aaaaac64438 Return addr 0x00002ba7c30df110 (proc_lib:init_p_do_apply/3 + 56)", "y(0) []","y(1) infinity", "y(2) ns_port_server", "y(3) {state,#Port<0.2771>,memcached,{[\"WARNING: Found duplicate entry for \\\"tap_keepal", "y(4) <0.487.0>","y(5) <0.486.0>", "0x00002aaaaac64470 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002ba7c30df130 (proc_lib:init_p_do_apply/3 + 88)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,0}]}, {heap_size,610}, {total_heap_size,610}, {links,[<9352.486.0>,#Port<9352.2771>]}, {memory,5768}, {message_queue_len,0}, {reductions,185}, {trap_exit,true}]}, {<9352.1369.0>, [{registered_name,[]}, {status,running}, {initial_call,{erlang,apply,2}}, {backtrace, ["Program counter: 0x00002ba7c301cc00 (unknown function)", "CP: 0x00002aaaab1d4da8 (diag_handler:grab_process_info/1 + 64)", "0x00002aaaab13c940 Return addr 0x00002aaaab1d7308 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) []", "0x00002aaaab13c950 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) []","y(1) <0.1369.0>", "0x00002aaaab13c968 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,[]},{status,waiting},{initial_call,{proc_lib,init_p,5}},{backtr", "y(1) <0.487.0>", "0x00002aaaab13c980 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,[]},{status,waiting},{initial_call,{proc_lib,init_p,5}},{backtr", "y(1) <0.486.0>", "0x00002aaaab13c998 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,hot_keys_keeper},{status,waiting},{initial_call,{proc_lib,init_", "y(1) <0.485.0>", "0x00002aaaab13c9b0 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,'stats_reader-default'},{status,waiting},{initial_call,{proc_li", "y(1) <0.480.0>", "0x00002aaaab13c9c8 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,'stats_archiver-default'},{status,waiting},{initial_call,{proc_", "y(1) <0.442.0>", "0x00002aaaab13c9e0 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,'ns_memcached-default'},{status,waiting},{initial_call,{proc_li", "y(1) <0.439.0>", "0x00002aaaab13c9f8 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,'ns_vbm_sup-default'},{status,waiting},{initial_call,{proc_lib,", "y(1) <0.438.0>", "0x00002aaaab13ca10 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,[]},{status,waiting},{initial_call,{proc_lib,init_p,5}},{backtr", "y(1) <0.437.0>", "0x00002aaaab13ca28 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,[]},{status,waiting},{initial_call,{erlang,apply,2}},{backtrace", "y(1) <0.423.0>", "0x00002aaaab13ca40 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,[]},{status,waiting},{initial_call,{erlang,apply,2}},{backtrace", "y(1) <0.422.0>", "0x00002aaaab13ca58 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,ns_moxi_sup},{status,waiting},{initial_call,{proc_lib,init_p,5}", "y(1) <0.421.0>", "0x00002aaaab13ca70 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,ns_bad_bucket_sup},{status,waiting},{initial_call,{proc_lib,ini", "y(1) <0.420.0>", "0x00002aaaab13ca88 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,ns_bad_bucket_worker},{status,waiting},{initial_call,{proc_lib,", "y(1) <0.419.0>", "0x00002aaaab13caa0 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,[]},{status,waiting},{initial_call,{proc_lib,init_p,5}},{backtr", "y(1) <0.407.0>", "0x00002aaaab13cab8 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,[]},{status,waiting},{initial_call,{proc_lib,init_p,5}},{backtr", "y(1) <0.405.0>", "0x00002aaaab13cad0 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,mnesia_late_loader},{status,waiting},{initial_call,{proc_lib,in", "y(1) <0.402.0>", "0x00002aaaab13cae8 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,mnesia_controller},{status,waiting},{initial_call,{proc_lib,ini", "y(1) <0.401.0>", "0x00002aaaab13cb00 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,mnesia_snmp_sup},{status,waiting},{initial_call,{proc_lib,init_", "y(1) <0.400.0>", "0x00002aaaab13cb18 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,mnesia_checkpoint_sup},{status,waiting},{initial_call,{proc_lib", "y(1) <0.399.0>", "0x00002aaaab13cb30 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,mnesia_tm},{status,waiting},{initial_call,{proc_lib,init_p,5}},", "y(1) <0.398.0>", "0x00002aaaab13cb48 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,mnesia_recover},{status,waiting},{initial_call,{proc_lib,init_p", "y(1) <0.397.0>", "0x00002aaaab13cb60 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,mnesia_locker},{status,waiting},{initial_call,{proc_lib,init_p,", "y(1) <0.396.0>", "0x00002aaaab13cb78 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,mnesia_subscr},{status,waiting},{initial_call,{proc_lib,init_p,", "y(1) <0.395.0>", "0x00002aaaab13cb90 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,mnesia_monitor},{status,waiting},{initial_call,{proc_lib,init_p", "y(1) <0.394.0>", "0x00002aaaab13cba8 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,mnesia_kernel_sup},{status,waiting},{initial_call,{proc_lib,ini", "y(1) <0.393.0>", "0x00002aaaab13cbc0 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,mnesia_event},{status,waiting},{initial_call,{proc_lib,init_p,5", "y(1) <0.392.0>", "0x00002aaaab13cbd8 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,mnesia_sup},{status,waiting},{initial_call,{proc_lib,init_p,5}}", "y(1) <0.391.0>", "0x00002aaaab13cbf0 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,[]},{status,waiting},{initial_call,{application_master,start_it", "y(1) <0.390.0>", "0x00002aaaab13cc08 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,[]},{status,waiting},{initial_call,{proc_lib,init_p,5}},{backtr", "y(1) <0.389.0>", "0x00002aaaab13cc20 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,[]},{status,waiting},{initial_call,{erlang,apply,2}},{backtrace", "y(1) <0.387.0>", "0x00002aaaab13cc38 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,ns_mnesia},{status,waiting},{initial_call,{proc_lib,init_p,5}},", "y(1) <0.386.0>", "0x00002aaaab13cc50 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,[]},{status,waiting},{initial_call,{erlang,apply,2}},{backtrace", "y(1) <0.385.0>", "0x00002aaaab13cc68 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,ns_good_bucket_sup},{status,waiting},{initial_call,{proc_lib,in", "y(1) <0.383.0>", "0x00002aaaab13cc80 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,ns_good_bucket_worker},{status,waiting},{initial_call,{proc_lib", "y(1) <0.382.0>", "0x00002aaaab13cc98 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,ns_stats_event},{status,waiting},{initial_call,{proc_lib,init_p", "y(1) <0.381.0>", "0x00002aaaab13ccb0 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,ns_tick_event},{status,waiting},{initial_call,{proc_lib,init_p,", "y(1) <0.380.0>", "0x00002aaaab13ccc8 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,[]},{status,waiting},{initial_call,{proc_lib,init_p,5}},{backtr", "y(1) <0.375.0>", "0x00002aaaab13cce0 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,[]},{status,waiting},{initial_call,{proc_lib,init_p,5}},{backtr", "y(1) <0.374.0>", "0x00002aaaab13ccf8 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,ns_port_sup},{status,waiting},{initial_call,{proc_lib,init_p,5}", "y(1) <0.370.0>", "0x00002aaaab13cd10 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,[]},{status,waiting},{initial_call,{proc_lib,init_p,5}},{backtr", "y(1) <0.365.0>", "0x00002aaaab13cd28 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,menelaus_web},{status,waiting},{initial_call,{proc_lib,init_p,5", "y(1) <0.364.0>", "0x00002aaaab13cd40 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,menelaus_sup},{status,waiting},{initial_call,{proc_lib,init_p,5", "y(1) <0.358.0>", "0x00002aaaab13cd58 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,ns_doctor},{status,waiting},{initial_call,{proc_lib,init_p,5}},", "y(1) <0.352.0>", "0x00002aaaab13cd70 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,ns_heart},{status,waiting},{initial_call,{proc_lib,init_p,5}},{", "y(1) <0.351.0>", "0x00002aaaab13cd88 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,ns_config_rep},{status,waiting},{initial_call,{proc_lib,init_p,", "y(1) <0.346.0>", "0x00002aaaab13cda0 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,ns_node_disco},{status,waiting},{initial_call,{proc_lib,init_p,", "y(1) <0.340.0>", "0x00002aaaab13cdb8 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,ns_node_disco_events},{status,waiting},{initial_call,{proc_lib,", "y(1) <0.339.0>", "0x00002aaaab13cdd0 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,ns_node_disco_sup},{status,waiting},{initial_call,{proc_lib,ini", "y(1) <0.338.0>", "0x00002aaaab13cde8 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,ns_mail},{status,waiting},{initial_call,{proc_lib,init_p,5}},{b", "y(1) <0.336.0>", "0x00002aaaab13ce00 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,ns_mail_sup},{status,waiting},{initial_call,{proc_lib,init_p,5}", "y(1) <0.335.0>", "0x00002aaaab13ce18 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,ns_log_events},{status,waiting},{initial_call,{proc_lib,init_p,", "y(1) <0.334.0>", "0x00002aaaab13ce30 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,ns_log},{status,waiting},{initial_call,{proc_lib,init_p,5}},{ba", "y(1) <0.333.0>", "0x00002aaaab13ce48 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,ns_server_sup},{status,waiting},{initial_call,{proc_lib,init_p,", "y(1) <0.332.0>", "0x00002aaaab13ce60 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,[]},{status,waiting},{initial_call,{inet_tcp_dist,do_setup,6}},", "y(1) <0.328.0>", "0x00002aaaab13ce78 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,[]},{status,waiting},{initial_call,{net_kernel,ticker,2}},{back", "y(1) <0.270.0>", "0x00002aaaab13ce90 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,[]},{status,waiting},{initial_call,{inet_tcp_dist,accept_loop,2", "y(1) <0.269.0>", "0x00002aaaab13cea8 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,net_kernel},{status,waiting},{initial_call,{proc_lib,init_p,5}}", "y(1) <0.268.0>", "0x00002aaaab13cec0 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,auth},{status,waiting},{initial_call,{proc_lib,init_p,5}},{back", "y(1) <0.267.0>", "0x00002aaaab13ced8 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,erl_epmd},{status,waiting},{initial_call,{proc_lib,init_p,5}},{", "y(1) <0.266.0>", "0x00002aaaab13cef0 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,net_sup},{status,waiting},{initial_call,{proc_lib,init_p,5}},{b", "y(1) <0.265.0>", "0x00002aaaab13cf08 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,inet_gethost_native},{status,waiting},{initial_call,{inet_getho", "y(1) <0.257.0>", "0x00002aaaab13cf20 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,inet_gethost_native_sup},{status,waiting},{initial_call,{proc_l", "y(1) <0.256.0>", "0x00002aaaab13cf38 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,dets},{status,waiting},{initial_call,{proc_lib,init_p,5}},{back", "y(1) <0.134.0>", "0x00002aaaab13cf50 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,dets_sup},{status,waiting},{initial_call,{proc_lib,init_p,5}},{", "y(1) <0.133.0>", "0x00002aaaab13cf68 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,disk_log_server},{status,waiting},{initial_call,{proc_lib,init_", "y(1) <0.130.0>", "0x00002aaaab13cf80 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,disk_log_sup},{status,waiting},{initial_call,{proc_lib,init_p,5", "y(1) <0.129.0>", "0x00002aaaab13cf98 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,ns_config},{status,waiting},{initial_call,{proc_lib,init_p,5}},", "y(1) <0.65.0>", "0x00002aaaab13cfb0 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,ns_config_events},{status,waiting},{initial_call,{proc_lib,init", "y(1) <0.64.0>", "0x00002aaaab13cfc8 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,ns_config_sup},{status,waiting},{initial_call,{proc_lib,init_p,", "y(1) <0.63.0>", "0x00002aaaab13cfe0 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,ns_cluster},{status,waiting},{initial_call,{proc_lib,init_p,5}}", "y(1) <0.62.0>", "0x00002aaaab13cff8 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,dist_manager},{status,waiting},{initial_call,{proc_lib,init_p,5", "y(1) <0.55.0>", "0x00002aaaab13d010 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,ns_server_cluster_sup},{status,waiting},{initial_call,{proc_lib", "y(1) <0.54.0>", "0x00002aaaab13d028 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,[]},{status,waiting},{initial_call,{application_master,start_it", "y(1) <0.53.0>", "0x00002aaaab13d040 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,[]},{status,waiting},{initial_call,{proc_lib,init_p,5}},{backtr", "y(1) <0.52.0>", "0x00002aaaab13d058 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,timer_server},{status,waiting},{initial_call,{proc_lib,init_p,5", "y(1) <0.50.0>", "0x00002aaaab13d070 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,[]},{status,waiting},{initial_call,{erlang,apply,2}},{backtrace", "y(1) <0.49.0>", "0x00002aaaab13d088 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,[]},{status,waiting},{initial_call,{erlang,apply,2}},{backtrace", "y(1) <0.48.0>", "0x00002aaaab13d0a0 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,cpu_sup},{status,waiting},{initial_call,{proc_lib,init_p,5}},{b", "y(1) <0.47.0>", "0x00002aaaab13d0b8 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,[]},{status,waiting},{initial_call,{erlang,apply,2}},{backtrace", "y(1) <0.46.0>", "0x00002aaaab13d0d0 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,memsup},{status,waiting},{initial_call,{proc_lib,init_p,5}},{ba", "y(1) <0.45.0>", "0x00002aaaab13d0e8 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,disksup},{status,waiting},{initial_call,{proc_lib,init_p,5}},{b", "y(1) <0.44.0>", "0x00002aaaab13d100 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,os_mon_sup},{status,waiting},{initial_call,{proc_lib,init_p,5}}", "y(1) <0.43.0>", "0x00002aaaab13d118 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,[]},{status,waiting},{initial_call,{application_master,start_it", "y(1) <0.42.0>", "0x00002aaaab13d130 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,[]},{status,waiting},{initial_call,{proc_lib,init_p,5}},{backtr", "y(1) <0.41.0>", "0x00002aaaab13d148 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,release_handler},{status,waiting},{initial_call,{proc_lib,init_", "y(1) <0.38.0>", "0x00002aaaab13d160 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,overload},{status,waiting},{initial_call,{proc_lib,init_p,5}},{", "y(1) <0.37.0>", "0x00002aaaab13d178 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,alarm_handler},{status,waiting},{initial_call,{proc_lib,init_p,", "y(1) <0.36.0>", "0x00002aaaab13d190 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,sasl_safe_sup},{status,waiting},{initial_call,{proc_lib,init_p,", "y(1) <0.35.0>", "0x00002aaaab13d1a8 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,sasl_sup},{status,waiting},{initial_call,{proc_lib,init_p,5}},{", "y(1) <0.34.0>", "0x00002aaaab13d1c0 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,[]},{status,waiting},{initial_call,{application_master,start_it", "y(1) <0.33.0>", "0x00002aaaab13d1d8 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,[]},{status,waiting},{initial_call,{proc_lib,init_p,5}},{backtr", "y(1) <0.32.0>", "0x00002aaaab13d1f0 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,kernel_safe_sup},{status,waiting},{initial_call,{proc_lib,init_", "y(1) <0.27.0>", "0x00002aaaab13d208 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,[]},{status,waiting},{initial_call,{proc_lib,init_p,5}},{backtr", "y(1) <0.26.0>", "0x00002aaaab13d220 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,user},{status,waiting},{initial_call,{user,server,2}},{backtrac", "y(1) <0.25.0>", "0x00002aaaab13d238 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,[]},{status,waiting},{initial_call,{proc_lib,init_p,5}},{backtr", "y(1) <0.24.0>", "0x00002aaaab13d250 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,standard_error},{status,waiting},{initial_call,{standard_error,", "y(1) <0.23.0>", "0x00002aaaab13d268 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,standard_error_sup},{status,waiting},{initial_call,{proc_lib,in", "y(1) <0.22.0>", "0x00002aaaab13d280 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,code_server},{status,waiting},{initial_call,{erlang,apply,2}},{", "y(1) <0.21.0>", "0x00002aaaab13d298 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,file_server_2},{status,waiting},{initial_call,{proc_lib,init_p,", "y(1) <0.20.0>", "0x00002aaaab13d2b0 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,global_group},{status,waiting},{initial_call,{proc_lib,init_p,5", "y(1) <0.19.0>", "0x00002aaaab13d2c8 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,inet_db},{status,waiting},{initial_call,{proc_lib,init_p,5}},{b", "y(1) <0.17.0>", "0x00002aaaab13d2e0 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,[]},{status,waiting},{initial_call,{erlang,apply,2}},{backtrace", "y(1) <0.16.0>", "0x00002aaaab13d2f8 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,[]},{status,waiting},{initial_call,{erlang,apply,2}},{backtrace", "y(1) <0.15.0>", "0x00002aaaab13d310 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,[]},{status,waiting},{initial_call,{erlang,apply,2}},{backtrace", "y(1) <0.14.0>", "0x00002aaaab13d328 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,global_name_server},{status,waiting},{initial_call,{proc_lib,in", "y(1) <0.13.0>", "0x00002aaaab13d340 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,rex},{status,waiting},{initial_call,{proc_lib,init_p,5}},{backt", "y(1) <0.12.0>", "0x00002aaaab13d358 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,kernel_sup},{status,waiting},{initial_call,{proc_lib,init_p,5}}", "y(1) <0.11.0>", "0x00002aaaab13d370 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,[]},{status,waiting},{initial_call,{application_master,start_it", "y(1) <0.10.0>", "0x00002aaaab13d388 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,[]},{status,waiting},{initial_call,{proc_lib,init_p,5}},{backtr", "y(1) <0.9.0>", "0x00002aaaab13d3a0 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,application_controller},{status,waiting},{initial_call,{erlang,", "y(1) <0.7.0>", "0x00002aaaab13d3b8 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,error_logger},{status,waiting},{initial_call,{proc_lib,init_p,5", "y(1) <0.6.0>", "0x00002aaaab13d3d0 Return addr 0x00002aaaab1d7348 (diag_handler:'-do_diag_per_node/0-lc$^0", "y(0) [{registered_name,erl_prim_loader},{status,waiting},{initial_call,{erlang,apply,2", "y(1) <0.3.0>", "0x00002aaaab13d3e8 Return addr 0x00002aaaab1d4fb8 (diag_handler:do_diag_per_node/0 + 224)", "y(0) [{registered_name,init},{status,waiting},{initial_call,{otp_ring0,start,2}},{back", "y(1) <0.0.0>", "0x00002aaaab13d400 Return addr 0x00002aaaaab60208 (rpc:'-handle_call_call/6-fun-0-'/5 + 20", "y(0) []","y(1) []", "y(2) [{version,[{kernel,\"2.13.4\"},{mnesia,\"4.4.12\"},{stdlib,\"1.16.4\"},{os_mon,\"2.2.4\"}", "y(3) [{{node,'ns_1@127.0.0.1',ns_log},[{filename,\"/var/opt/membase/1.6.5.4r/data/ns_1/", "y(4) [\"bucket_engine_1.6.5.3_1_g5ba5371-Linux.x86_64.tar.gz\",\"curl-7.21.1-w64_patched.", "y(5) [{kernel,\"2.13.4\"},{mnesia,\"4.4.12\"},{stdlib,\"1.16.4\"},{os_mon,\"2.2.4\"},{ns_serve", "0x00002aaaab13d438 Return addr 0x0000000000846c78 ()", "y(0) Catch 0x00002aaaaab60208 (rpc:'-handle_call_call/6-fun-0-'/5 + 208)", "y(1) []","y(2) []","y(3) []", "y(4) <0.12.0>"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,19}]}, {heap_size,46368}, {total_heap_size,364179}, {links,[]}, {memory,2914208}, {message_queue_len,0}, {reductions,610790}, {trap_exit,false}]}]}, {memory,{2058731520,727810048,{<9352.12.0>,4114680}}}, {disk, [{"/",11204616,44}, {"/boot",101086,14}, {"/dev/shm",1005240,0}]}]}] nodes_info = [{struct,[{uptime,<<"1480">>}, {memoryTotal,2058731520}, {memoryFree,1378230272}, {mcdMemoryReserved,1570}, {mcdMemoryAllocated,1570}, {otpNode,<<"ns_1@10.1.5.227">>}, {otpCookie,<<"bujipmggocppuqmx">>}, {clusterMembership,<<"active">>}, {status,<<"healthy">>}, {hostname,<<"10.1.5.227:8091">>}, {clusterCompatibility,1}, {version,<<"1.6.5.4r">>}, {os,<<"x86_64-unknown-linux-gnu">>}, {ports,{struct,[{proxy,11211},{direct,11210}]}}]}, {struct,[{uptime,<<"1443">>}, {memoryTotal,2058731520}, {memoryFree,1330921472}, {mcdMemoryReserved,1570}, {mcdMemoryAllocated,1570}, {otpNode,<<"ns_1@10.1.5.229">>}, {otpCookie,<<"bujipmggocppuqmx">>}, {clusterMembership,<<"active">>}, {status,<<"healthy">>}, {hostname,<<"10.1.5.229:8091">>}, {clusterCompatibility,1}, {version,<<"1.6.5.4r">>}, {os,<<"x86_64-unknown-linux-gnu">>}, {ports,{struct,[{proxy,11211},{direct,11210}]}}]}] buckets = [{"default", [{num_replicas,1}, {ram_quota,1096810496}, {auth_type,sasl}, {sasl_password,[]}, {type,membase}, {num_vbuckets,1024}, {db_shards,4}, {ht_size,3079}, {tap_keepalive,0}, {tap_noop_interval,20}, {max_txn_size,1000}, {ht_locks,5}, {servers,['ns_1@10.1.5.227','ns_1@10.1.5.229']}, {map,[['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined]]}]}] logs: ------------------------------- 2011-05-10 13:50:08.183 ns_node_disco:3:info:cookie update - Initial otp cookie generated: bujipmggocppuqmx 2011-05-10 13:50:08.500 menelaus_app:1:info:web start ok - Membase Server has started on web port 8091 on node 'ns_1@127.0.0.1'. 2011-05-10 13:53:27.153 menelaus_web:12:info:message - Created bucket "default" of type: membase 2011-05-10 13:53:30.285 ns_memcached:1:info:message - Bucket "default" loaded on node 'ns_1@127.0.0.1' in 2 seconds. 2011-05-10 13:53:42.656 ns_node_disco:5:warning:node down - Node nonode@nohost saw that node 'ns_1@127.0.0.1' went down. 2011-05-10 13:53:42.658 ns_node_disco:4:info:node up - Node 'ns_1@10.1.5.227' saw that node 'ns_1@10.1.5.227' came up. 2011-05-10 13:53:44.030 ns_node_disco:4:info:node up - Node 'ns_1@10.1.5.227' saw that node 'ns_1@10.1.5.229' came up. 2011-05-10 13:53:44.453 menelaus_app:1:info:web start ok - Membase Server has started on web port 8091 on node 'ns_1@10.1.5.229'. 2011-05-10 13:53:44.520 ns_port_server:0:info:message - Port server memcached on node 'ns_1@10.1.5.229' exited with status 71. Restarting. Messages: failed to listen on TCP port 11210: Address already in use 2011-05-10 13:53:44.520 supervisor_cushion:1:warning:port exited too soon after restart - Service memcached exited on node 'ns_1@10.1.5.229' in 0.03s 2011-05-10 13:53:44.628 ns_cluster:3:info:message - Node ns_1@10.1.5.229 joined cluster 2011-05-10 13:53:45.787 ns_orchestrator:4:info:message - Starting rebalance, KeepNodes = ['ns_1@10.1.5.227','ns_1@10.1.5.229'], EjectNodes = [] 2011-05-10 13:53:52.054 ns_memcached:1:info:message - Bucket "default" loaded on node 'ns_1@10.1.5.229' in 2 seconds. logs_node: ------------------------------- INFO REPORT <3.54.0> 2011-05-10 13:50:08 =============================================================================== nonode@nohost:log_os_info:25: OS type: {unix,linux} Version: {2,6,18} Runtime info: [{otp_release,"R13B03"}, {erl_version,"5.7.4"}, {erl_version_long,"Erlang R13B03 (erts-5.7.4) [source] [64-bit] [rq:1] [async-threads:16] [hipe] [kernel-poll:false]\n"}, {system_arch_raw,"x86_64-unknown-linux-gnu"}, {system_arch,"x86_64-unknown-linux-gnu"}, {localtime,{{2011,5,10},{13,50,8}}}, {memory,[{total,14709800}, {processes,1336728}, {processes_used,1322968}, {system,13373072}, {atom,396577}, {atom_used,393692}, {binary,41752}, {code,3478759}, {ets,324512}]}, {loaded,[ns_info,log_os_info,misc,ns_log_mf_h, ns_server_cluster_sup,ns_server,timer,io_lib_fread, cpu_sup,memsup,disksup,os_mon,io_lib_pretty,unicode, io_lib_format,io_lib,io,sasl_report,release_handler, calendar,overload,alarm_handler,sasl_report_tty_h, sasl,ns_bootstrap,file_io_server,orddict,erl_eval, file,c,error_logger_tty_h,queue,kernel_config,user, user_sup,supervisor_bridge,standard_error,gb_sets, hipe_unified_loader,packages,code_server,code, file_server,net_kernel,global_group,erl_distribution, string,ets,filename,inet_gethost_native,inet_parse, inet,inet_udp,os,inet_config,inet_db,global,gb_trees, rpc,dict,supervisor,kernel,application_master,sys, application,gen_server,erl_parse,proplists,erl_scan, lists,application_controller,proc_lib,gen,gen_event, error_logger,heart,error_handler,erlang, erl_prim_loader,prim_zip,zlib,prim_file,prim_inet, init,otp_ring0]}, {applications,[{kernel,"ERTS CXC 138 10","2.13.4"}, {stdlib,"ERTS CXC 138 10","1.16.4"}, {os_mon,"CPO CXC 138 46","2.2.4"}, {ns_server,"Membase server","1.6.5.4r"}, {menelaus,"Membase menelaus","1.6.5.4r"}, {sasl,"SASL CXC 138 11","2.1.8"}]}, {pre_loaded,[erlang,erl_prim_loader,prim_zip,zlib,prim_file, prim_inet,init,otp_ring0]}, {process_count,44}, {node,nonode@nohost}, {nodes,[]}, {registered,[rex,user,inet_db,kernel_sup,init, global_name_server,code_server,release_handler, file_server_2,application_controller,error_logger, kernel_safe_sup,overload,ns_server_cluster_sup, standard_error,alarm_handler,global_group, timer_server,sasl_sup,os_mon_sup,erl_prim_loader, cpu_sup,memsup,sasl_safe_sup,disksup, standard_error_sup]}, {cookie,nocookie}, {wordsize,8}, {wall_clock,0}] INFO REPORT <3.55.0> 2011-05-10 13:50:08 =============================================================================== reading ip config from "/etc/opt/membase/1.6.5.4r/ip" PROGRESS REPORT <3.56.0> 2011-05-10 13:50:08 =============================================================================== supervisor {local,net_sup} started [{pid,<3.57.0>}, {name,erl_epmd}, {mfa,{erl_epmd,start_link,[]}}, {restart_type,permanent}, {shutdown,2000}, {child_type,worker}] PROGRESS REPORT <3.56.0> 2011-05-10 13:50:08 =============================================================================== supervisor {local,net_sup} started [{pid,<3.58.0>}, {name,auth}, {mfa,{auth,start_link,[]}}, {restart_type,permanent}, {shutdown,2000}, {child_type,worker}] PROGRESS REPORT <6040.56.0> 2011-05-10 13:50:08 =============================================================================== supervisor {local,net_sup} started [{pid,<6040.59.0>}, {name,net_kernel}, {mfa,{net_kernel,start_link,[['ns_1@127.0.0.1',longnames]]}}, {restart_type,permanent}, {shutdown,2000}, {child_type,worker}] PROGRESS REPORT <6040.11.0> 2011-05-10 13:50:08 =============================================================================== supervisor {local,kernel_sup} started [{pid,<6040.56.0>}, {name,net_sup_dynamic}, {mfa,{erl_distribution,start_link,[['ns_1@127.0.0.1',longnames]]}}, {restart_type,permanent}, {shutdown,1000}, {child_type,supervisor}] PROGRESS REPORT <6040.54.0> 2011-05-10 13:50:08 =============================================================================== supervisor {local,ns_server_cluster_sup} started [{pid,<6040.55.0>}, {name,dist_manager}, {mfa,{dist_manager,start_link,[]}}, {restart_type,permanent}, {shutdown,10}, {child_type,worker}] PROGRESS REPORT <6040.54.0> 2011-05-10 13:50:08 =============================================================================== supervisor {local,ns_server_cluster_sup} started [{pid,<6040.62.0>}, {name,ns_cluster}, {mfa,{ns_cluster,start_link,[]}}, {restart_type,permanent}, {shutdown,5000}, {child_type,worker}] INFO REPORT <6040.63.0> 2011-05-10 13:50:08 =============================================================================== loading config from "/etc/opt/membase/1.6.5.4r/config" PROGRESS REPORT <6040.63.0> 2011-05-10 13:50:08 =============================================================================== supervisor {local,ns_config_sup} started [{pid,<6040.64.0>}, {name,ns_config_events}, {mfa,{gen_event,start_link,[{local,ns_config_events}]}}, {restart_type,permanent}, {shutdown,10}, {child_type,worker}] PROGRESS REPORT <6040.63.0> 2011-05-10 13:50:08 =============================================================================== supervisor {local,ns_config_sup} started [{pid,<6040.65.0>}, {name,ns_config}, {mfa,{ns_config,start_link, ["/etc/opt/membase/1.6.5.4r/config", ns_config_default]}}, {restart_type,permanent}, {shutdown,10}, {child_type,worker}] PROGRESS REPORT <6040.63.0> 2011-05-10 13:50:08 =============================================================================== supervisor {local,ns_config_sup} started [{pid,<6040.67.0>}, {name,ns_config_isasl_sync}, {mfa,{ns_config_isasl_sync,start_link,[]}}, {restart_type,transient}, {shutdown,10}, {child_type,worker}] INFO REPORT <6040.64.0> 2011-05-10 13:50:08 =============================================================================== isasl_sync init: ["/var/opt/membase/1.6.5.4r/data/ns_1/isasl.pw","_admin", "_admin"] INFO REPORT <6040.64.0> 2011-05-10 13:50:08 =============================================================================== isasl_sync init buckets: [] INFO REPORT <6040.64.0> 2011-05-10 13:50:08 =============================================================================== Writing isasl passwd file: "/var/opt/membase/1.6.5.4r/data/ns_1/isasl.pw" PROGRESS REPORT <6040.63.0> 2011-05-10 13:50:08 =============================================================================== supervisor {local,ns_config_sup} started [{pid,<6040.69.0>}, {name,ns_config_log}, {mfa,{ns_config_log,start_link,[]}}, {restart_type,transient}, {shutdown,10}, {child_type,worker}] PROGRESS REPORT <6040.54.0> 2011-05-10 13:50:08 =============================================================================== supervisor {local,ns_server_cluster_sup} started [{pid,<6040.63.0>}, {name,ns_config_sup}, {mfa,{ns_config_sup,start_link,[]}}, {restart_type,permanent}, {shutdown,infinity}, {child_type,supervisor}] INFO REPORT <6040.71.0> 2011-05-10 13:50:08 =============================================================================== ns_log:init(): Couldn't load logs from "/var/opt/membase/1.6.5.4r/data/ns_1/ns_log": {error, enoent} PROGRESS REPORT <6040.70.0> 2011-05-10 13:50:08 =============================================================================== supervisor {local,ns_server_sup} started [{pid,<6040.71.0>}, {name,ns_log}, {mfa,{ns_log,start_link,[]}}, {restart_type,permanent}, {shutdown,10}, {child_type,worker}] PROGRESS REPORT <6040.70.0> 2011-05-10 13:50:08 =============================================================================== supervisor {local,ns_server_sup} started [{pid,<6040.72.0>}, {name,ns_log_events}, {mfa,{gen_event,start_link,[{local,ns_log_events}]}}, {restart_type,permanent}, {shutdown,10}, {child_type,worker}] PROGRESS REPORT <6040.73.0> 2011-05-10 13:50:08 =============================================================================== supervisor {local,ns_mail_sup} started [{pid,<6040.74.0>}, {name,ns_mail}, {mfa,{ns_mail,start_link,[]}}, {restart_type,permanent}, {shutdown,10}, {child_type,worker}] PROGRESS REPORT <6040.73.0> 2011-05-10 13:50:08 =============================================================================== supervisor {local,ns_mail_sup} started [{pid,<6040.75.0>}, {name,ns_mail_log}, {mfa,{ns_mail_log,start_link,[]}}, {restart_type,transient}, {shutdown,10}, {child_type,worker}] PROGRESS REPORT <6040.70.0> 2011-05-10 13:50:08 =============================================================================== supervisor {local,ns_server_sup} started [{pid,<6040.73.0>}, {name,ns_mail_sup}, {mfa,{ns_mail_sup,start_link,[]}}, {restart_type,permanent}, {shutdown,infinity}, {child_type,supervisor}] PROGRESS REPORT <6040.76.0> 2011-05-10 13:50:08 =============================================================================== supervisor {local,ns_node_disco_sup} started [{pid,<6040.77.0>}, {name,ns_node_disco_events}, {mfa,{gen_event,start_link,[{local,ns_node_disco_events}]}}, {restart_type,permanent}, {shutdown,10}, {child_type,worker}] INFO REPORT <6040.78.0> 2011-05-10 13:50:08 =============================================================================== Initting ns_node_disco with [] INFO REPORT <6040.79.0> 2011-05-10 13:50:08 =============================================================================== ns_node_disco cookie_sync INFO REPORT <6040.79.0> 2011-05-10 13:50:08 =============================================================================== ns_log: logging ns_node_disco:3:Initial otp cookie generated: bujipmggocppuqmx INFO REPORT <6040.64.0> 2011-05-10 13:50:08 =============================================================================== config change: otp -> [{cookie,bujipmggocppuqmx}] INFO REPORT <6040.79.0> 2011-05-10 13:50:08 =============================================================================== ns_node_disco: nodes_wanted updated: ['ns_1@127.0.0.1'], with cookie: bujipmggocppuqmx INFO REPORT <6040.79.0> 2011-05-10 13:50:08 =============================================================================== ns_node_disco: nodes_wanted pong: ['ns_1@127.0.0.1'], with cookie: bujipmggocppuqmx PROGRESS REPORT <6040.76.0> 2011-05-10 13:50:08 =============================================================================== supervisor {local,ns_node_disco_sup} started [{pid,<6040.78.0>}, {name,ns_node_disco}, {mfa,{ns_node_disco,start_link,[]}}, {restart_type,permanent}, {shutdown,10}, {child_type,worker}] PROGRESS REPORT <6040.76.0> 2011-05-10 13:50:08 =============================================================================== supervisor {local,ns_node_disco_sup} started [{pid,<6040.83.0>}, {name,ns_node_disco_log}, {mfa,{ns_node_disco_log,start_link,[]}}, {restart_type,transient}, {shutdown,10}, {child_type,worker}] PROGRESS REPORT <6040.76.0> 2011-05-10 13:50:08 =============================================================================== supervisor {local,ns_node_disco_sup} started [{pid,<6040.84.0>}, {name,ns_node_disco_conf_events}, {mfa,{ns_node_disco_conf_events,start_link,[]}}, {restart_type,transient}, {shutdown,10}, {child_type,worker}] INFO REPORT <6040.85.0> 2011-05-10 13:50:08 =============================================================================== ns_config_rep init pulling INFO REPORT <6040.85.0> 2011-05-10 13:50:08 =============================================================================== ns_config_rep init pushing INFO REPORT <6040.85.0> 2011-05-10 13:50:08 =============================================================================== ns_config_rep init reannouncing INFO REPORT <6040.64.0> 2011-05-10 13:50:08 =============================================================================== ns_node_disco_conf_events config on otp INFO REPORT <6040.64.0> 2011-05-10 13:50:08 =============================================================================== config change: otp -> [{cookie,bujipmggocppuqmx}] INFO REPORT <6040.64.0> 2011-05-10 13:50:08 =============================================================================== config change: alerts -> [{email,[]}, {email_alerts,false}, [{user,undefined}, {pass,"********"}, {addr,undefined}, {port,undefined}, {encrypt,false}], {alerts,[server_down,server_unresponsive,server_up,server_joined,server_left, bucket_created,bucket_deleted,bucket_auth_failed]}] INFO REPORT <6040.64.0> 2011-05-10 13:50:08 =============================================================================== config change: buckets -> [{configs,[]}] INFO REPORT <6040.64.0> 2011-05-10 13:50:08 =============================================================================== config change: memory_quota -> 1570 INFO REPORT <6040.87.0> 2011-05-10 13:50:08 =============================================================================== ns_node_disco cookie_sync INFO REPORT <6040.64.0> 2011-05-10 13:50:08 =============================================================================== config change: moxi -> [{port,11211},{verbosity,[]}] INFO REPORT <6040.64.0> 2011-05-10 13:50:08 =============================================================================== ns_node_disco_conf_events config on nodes_wanted INFO REPORT <6040.64.0> 2011-05-10 13:50:08 =============================================================================== config change: nodes_wanted -> ['ns_1@127.0.0.1'] INFO REPORT <6040.64.0> 2011-05-10 13:50:08 =============================================================================== config change: port_servers -> [{moxi,"./bin/moxi/moxi", ["-Z", {"port_listen=~B,default_bucket_name=default,downstream_max=1024,downstream_conn_max=4,connect_max_errors=5,connect_retry_interval=30000,connect_timeout=400,auth_timeout=100,cycle=200,downstream_conn_queue_timeout=200,downstream_timeout=5000,wait_queue_timeout=200", [port]}, "-z", {"url=http://127.0.0.1:~B/pools/default/saslBucketsStreaming", [{rest,port}]}, "-p","0","-Y","y","-O","stderr", {"~s",[verbosity]}], [{env,[{"EVENT_NOSELECT","1"}, {"MOXI_SASL_PLAIN_USR",{"~s",[{ns_moxi_sup,rest_user,[]}]}}, {"MOXI_SASL_PLAIN_PWD",{"~s",[{ns_moxi_sup,rest_pass,[]}]}}]}, use_stdio,stderr_to_stdout,stream]}, {memcached,"./bin/memcached/memcached", ["-X","./bin/memcached/stdin_term_handler.so","-p", {"~B",[port]}, "-E","./bin/bucket_engine/bucket_engine.so","-B","binary","-r", "-c","10000","-e", {"admin=~s;default_bucket_name=default;auto_create=false", [admin_user]}, {"~s",[verbosity]}], [{env,[{"EVENT_NOSELECT","1"}, {"MEMCACHED_TOP_KEYS","100"}, {"ISASL_PWFILE",{"~s",[{isasl,path}]}}, {"ISASL_DB_CHECK_TIME","1"}]}, use_stdio,stderr_to_stdout,stream]}] INFO REPORT <6040.88.0> 2011-05-10 13:50:08 =============================================================================== ns_node_disco cookie_sync INFO REPORT <6040.64.0> 2011-05-10 13:50:08 =============================================================================== config change: replication -> [{enabled,true}] INFO REPORT <6040.87.0> 2011-05-10 13:50:08 =============================================================================== ns_node_disco: nodes_wanted updated: ['ns_1@127.0.0.1'], with cookie: bujipmggocppuqmx INFO REPORT <6040.64.0> 2011-05-10 13:50:08 =============================================================================== config change: rest -> [{port,8091}] INFO REPORT <6040.64.0> 2011-05-10 13:50:08 =============================================================================== config change: rest_creds -> ******** INFO REPORT <6040.64.0> 2011-05-10 13:50:08 =============================================================================== config change: {node,'ns_1@127.0.0.1',isasl} -> [{path,"/var/opt/membase/1.6.5.4r/data/ns_1/isasl.pw"}] INFO REPORT <6040.64.0> 2011-05-10 13:50:08 =============================================================================== config change: {node,'ns_1@127.0.0.1',membership} -> active INFO REPORT <6040.64.0> 2011-05-10 13:50:08 =============================================================================== config change: {node,'ns_1@127.0.0.1',memcached} -> [{port,11210}, {dbdir,"/var/opt/membase/1.6.5.4r/data/ns_1"}, {admin_user,"_admin"}, {admin_pass,"_admin"}, {bucket_engine,"./bin/bucket_engine/bucket_engine.so"}, {engines,[{membase,[{engine,"bin/ep_engine/ep.so"}, {initfile,"/etc/opt/membase/1.6.5.4r/init.sql"}]}, {memcached,[{engine,"bin/memcached/default_engine.so"}]}]}, {verbosity,[]}] INFO REPORT <6040.87.0> 2011-05-10 13:50:08 =============================================================================== ns_node_disco: nodes_wanted pong: ['ns_1@127.0.0.1'], with cookie: bujipmggocppuqmx INFO REPORT <6040.88.0> 2011-05-10 13:50:08 =============================================================================== ns_node_disco: nodes_wanted updated: ['ns_1@127.0.0.1'], with cookie: bujipmggocppuqmx INFO REPORT <6040.64.0> 2011-05-10 13:50:08 =============================================================================== config change: {node,'ns_1@127.0.0.1',ns_log} -> [{filename,"/var/opt/membase/1.6.5.4r/data/ns_1/ns_log"}] INFO REPORT <6040.64.0> 2011-05-10 13:50:08 =============================================================================== ns_node_disco_conf_events config all INFO REPORT <6040.88.0> 2011-05-10 13:50:08 =============================================================================== ns_node_disco: nodes_wanted pong: ['ns_1@127.0.0.1'], with cookie: bujipmggocppuqmx INFO REPORT <6040.85.0> 2011-05-10 13:50:08 =============================================================================== Pushing config PROGRESS REPORT <6040.76.0> 2011-05-10 13:50:08 =============================================================================== supervisor {local,ns_node_disco_sup} started [{pid,<6040.85.0>}, {name,ns_config_rep}, {mfa,{ns_config_rep,start_link,[]}}, {restart_type,permanent}, {shutdown,10}, {child_type,worker}] PROGRESS REPORT <6040.70.0> 2011-05-10 13:50:08 =============================================================================== supervisor {local,ns_server_sup} started [{pid,<6040.76.0>}, {name,ns_node_disco_sup}, {mfa,{ns_node_disco_sup,start_link,[]}}, {restart_type,permanent}, {shutdown,infinity}, {child_type,supervisor}] INFO REPORT <6040.85.0> 2011-05-10 13:50:08 =============================================================================== Pushing config done PROGRESS REPORT <6040.70.0> 2011-05-10 13:50:08 =============================================================================== supervisor {local,ns_server_sup} started [{pid,<6040.90.0>}, {name,ns_heart}, {mfa,{ns_heart,start_link,[]}}, {restart_type,permanent}, {shutdown,10}, {child_type,worker}] PROGRESS REPORT <6040.70.0> 2011-05-10 13:50:08 =============================================================================== supervisor {local,ns_server_sup} started [{pid,<6040.92.0>}, {name,ns_doctor}, {mfa,{ns_doctor,start_link,[]}}, {restart_type,permanent}, {shutdown,10}, {child_type,worker}] INFO REPORT <6040.92.0> 2011-05-10 13:50:08 =============================================================================== ns_1@127.0.0.1:ns_doctor:78: Got initial status [{'ns_1@127.0.0.1', [{last_heard, {1305,60608,420008}}, {active_buckets,[]}, {memory, [{total,16440504}, {processes,2245328}, {processes_used,2234096}, {system,14195176}, {atom,473097}, {atom_used,450887}, {binary,33072}, {code,4161949}, {ets,373176}]}, {cluster_compatibility_version, 1}, {version, [{kernel,"2.13.4"}, {stdlib,"1.16.4"}, {os_mon,"2.2.4"}, {ns_server,"1.6.5.4r"}, {menelaus,"1.6.5.4r"}, {sasl,"2.1.8"}]}, {system_arch, "x86_64-unknown-linux-gnu"}, {wall_clock,0}, {memory_data, {2058731520,557842432, {<6040.3.0>,196664}}}, {disk_data, [{"/",11204616,49}, {"/boot",101086,14}, {"/dev/shm",1005240,0}]}, {meminfo, <<"MemTotal: 2010480 kB\nMemFree: 1464464 kB\nBuffers: 58692 kB\nCached: 333308 kB\nSwapCached: 20084 kB\nActive: 234744 kB\nInactive: 198460 kB\nHighTotal: 0 kB\nHighFree: 0 kB\nLowTotal: 2010480 kB\nLowFree: 1464464 kB\nSwapTotal: 4030456 kB\nSwapFree: 3952288 kB\nDirty: 193020 kB\nWriteback: 0 kB\nAnonPages: 39060 kB\nMapped: 12728 kB\nSlab: 88236 kB\nPageTables: 5764 kB\nNFS_Unstable: 0 kB\nBounce: 0 kB\nCommitLimit: 5035696 kB\nCommitted_AS: 234840 kB\nVmallocTotal: 34359738367 kB\nVmallocUsed: 265484 kB\nVmallocChunk: 34359471011 kB\nHugePages_Total: 0\nHugePages_Free: 0\nHugePages_Rsvd: 0\nHugepagesize: 2048 kB\n">>}, {replication,[]}, {system_memory_data, [{system_total_memory, 2058731520}, {free_swap,4047142912}, {total_swap,4127186944}, {cached_memory,341307392}, {buffered_memory, 60100608}, {free_memory,1499611136}, {total_memory, 2058731520}]}, {statistics, [{wall_clock,{580,264}}, {context_switches, {4706,0}}, {garbage_collection, {963,2013630,0}}, {io, {{input,2189033}, {output,112859}}}, {reductions, {903234,903234}}, {run_queue,0}, {runtime,{230,230}}]}]}] PROGRESS REPORT <6040.94.0> 2011-05-10 13:50:08 =============================================================================== supervisor {local,menelaus_sup} started [{pid,<6040.95.0>}, {name,menelaus_web}, {mfa,{menelaus_web,start_link,[]}}, {restart_type,permanent}, {shutdown,5000}, {child_type,worker}] PROGRESS REPORT <6040.94.0> 2011-05-10 13:50:08 =============================================================================== supervisor {local,menelaus_sup} started [{pid,<6040.97.0>}, {name,menelaus_event}, {mfa,{menelaus_event,start_link,[]}}, {restart_type,transient}, {shutdown,5000}, {child_type,worker}] PROGRESS REPORT <6040.94.0> 2011-05-10 13:50:08 =============================================================================== supervisor {local,menelaus_sup} started [{pid,<6040.98.0>}, {name,hot_keys_keeper}, {mfa,{hot_keys_keeper,start_link,[]}}, {restart_type,permanent}, {shutdown,5000}, {child_type,worker}] INFO REPORT <6040.70.0> 2011-05-10 13:50:08 =============================================================================== ns_log: logging menelaus_app:1:Membase Server has started on web port 8091 on node 'ns_1@127.0.0.1'. PROGRESS REPORT <6040.70.0> 2011-05-10 13:50:08 =============================================================================== supervisor {local,ns_server_sup} started [{pid,<6040.94.0>}, {name,menelaus}, {mfa,{menelaus_app,start_subapp,[]}}, {restart_type,permanent}, {shutdown,infinity}, {child_type,supervisor}] PROGRESS REPORT <6040.99.0> 2011-05-10 13:50:08 =============================================================================== supervisor {local,ns_port_sup} started [{pid,<6040.100.0>}, {name,ns_port_init}, {mfa,{ns_port_init,start_link,[]}}, {restart_type,transient}, {shutdown,10}, {child_type,worker}] INFO REPORT <6040.101.0> 2011-05-10 13:50:08 =============================================================================== starting ns_port_server with delay of 5000 PROGRESS REPORT <6040.99.0> 2011-05-10 13:50:08 =============================================================================== supervisor {local,ns_port_sup} started [{pid,<6040.101.0>}, {name, {moxi,"./bin/moxi/moxi", ["-Z", "port_listen=11211,default_bucket_name=default,downstream_max=1024,downstream_conn_max=4,connect_max_errors=5,connect_retry_interval=30000,connect_timeout=400,auth_timeout=100,cycle=200,downstream_conn_queue_timeout=200,downstream_timeout=5000,wait_queue_timeout=200", "-z", "url=http://127.0.0.1:8091/pools/default/saslBucketsStreaming", "-p","0","-Y","y","-O","stderr",[]], [{env, [{"EVENT_NOSELECT","1"}, {"MOXI_SASL_PLAIN_USR",[]}, {"MOXI_SASL_PLAIN_PWD",[]}]}, use_stdio,stderr_to_stdout,stream]}}, {mfa, {supervisor_cushion,start_link, [moxi,5000,ns_port_server,start_link, [moxi,"./bin/moxi/moxi", ["-Z", "port_listen=11211,default_bucket_name=default,downstream_max=1024,downstream_conn_max=4,connect_max_errors=5,connect_retry_interval=30000,connect_timeout=400,auth_timeout=100,cycle=200,downstream_conn_queue_timeout=200,downstream_timeout=5000,wait_queue_timeout=200", "-z", "url=http://127.0.0.1:8091/pools/default/saslBucketsStreaming", "-p","0","-Y","y","-O","stderr",[]], [{env, [{"EVENT_NOSELECT","1"}, {"MOXI_SASL_PLAIN_USR",[]}, {"MOXI_SASL_PLAIN_PWD",[]}]}, use_stdio,stderr_to_stdout,stream]]]}}, {restart_type,permanent}, {shutdown,10}, {child_type,worker}] INFO REPORT <6040.103.0> 2011-05-10 13:50:08 =============================================================================== starting ns_port_server with delay of 5000 PROGRESS REPORT <6040.99.0> 2011-05-10 13:50:08 =============================================================================== supervisor {local,ns_port_sup} started [{pid,<6040.103.0>}, {name, {memcached,"./bin/memcached/memcached", ["-X","./bin/memcached/stdin_term_handler.so","-p","11210", "-E","./bin/bucket_engine/bucket_engine.so","-B","binary", "-r","-c","10000","-e", "admin=_admin;default_bucket_name=default;auto_create=false", []], [{env, [{"EVENT_NOSELECT","1"}, {"MEMCACHED_TOP_KEYS","100"}, {"ISASL_PWFILE", "/var/opt/membase/1.6.5.4r/data/ns_1/isasl.pw"}, {"ISASL_DB_CHECK_TIME","1"}]}, use_stdio,stderr_to_stdout,stream]}}, {mfa, {supervisor_cushion,start_link, [memcached,5000,ns_port_server,start_link, [memcached,"./bin/memcached/memcached", ["-X","./bin/memcached/stdin_term_handler.so","-p", "11210","-E","./bin/bucket_engine/bucket_engine.so","-B", "binary","-r","-c","10000","-e", "admin=_admin;default_bucket_name=default;auto_create=false", []], [{env, [{"EVENT_NOSELECT","1"}, {"MEMCACHED_TOP_KEYS","100"}, {"ISASL_PWFILE", "/var/opt/membase/1.6.5.4r/data/ns_1/isasl.pw"}, {"ISASL_DB_CHECK_TIME","1"}]}, use_stdio,stderr_to_stdout,stream]]]}}, {restart_type,permanent}, {shutdown,10}, {child_type,worker}] PROGRESS REPORT <6040.70.0> 2011-05-10 13:50:08 =============================================================================== supervisor {local,ns_server_sup} started [{pid,<6040.99.0>}, {name,ns_port_sup}, {mfa,{ns_port_sup,start_link,[]}}, {restart_type,permanent}, {shutdown,10}, {child_type,worker}] PROGRESS REPORT <6040.70.0> 2011-05-10 13:50:08 =============================================================================== supervisor {local,ns_server_sup} started [{pid,<6040.105.0>}, {name,ns_tick_event}, {mfa,{gen_event,start_link,[{local,ns_tick_event}]}}, {restart_type,permanent}, {shutdown,10}, {child_type,worker}] PROGRESS REPORT <6040.70.0> 2011-05-10 13:50:08 =============================================================================== supervisor {local,ns_server_sup} started [{pid,<6040.106.0>}, {name,ns_stats_event}, {mfa,{gen_event,start_link,[{local,ns_stats_event}]}}, {restart_type,permanent}, {shutdown,10}, {child_type,worker}] PROGRESS REPORT <6040.70.0> 2011-05-10 13:50:08 =============================================================================== supervisor {local,ns_server_sup} started [{pid,<6040.107.0>}, {name,ns_good_bucket_worker}, {mfa,{work_queue,start_link,[ns_good_bucket_worker]}}, {restart_type,permanent}, {shutdown,10}, {child_type,worker}] PROGRESS REPORT <6040.70.0> 2011-05-10 13:50:08 =============================================================================== supervisor {local,ns_server_sup} started [{pid,<6040.108.0>}, {name,ns_good_bucket_sup}, {mfa,{ns_bucket_sup,start_link, [ns_good_bucket_sup, #Fun, ns_good_bucket_worker]}}, {restart_type,permanent}, {shutdown,infinity}, {child_type,supervisor}] INFO REPORT <6040.70.0> 2011-05-10 13:50:08 =============================================================================== ns_1@127.0.0.1:misc:725: start_singleton(gen_fsm, ns_orchestrator, [], []): started as <6040.109.0> on 'ns_1@127.0.0.1' PROGRESS REPORT <6040.70.0> 2011-05-10 13:50:08 =============================================================================== supervisor {local,ns_server_sup} started [{pid,<6040.109.0>}, {name,ns_orchestrator}, {mfa,{ns_orchestrator,start_link,[]}}, {restart_type,permanent}, {shutdown,20}, {child_type,worker}] PROGRESS REPORT <6040.114.0> 2011-05-10 13:50:08 =============================================================================== supervisor {local,mnesia_sup} started [{pid,<6040.115.0>}, {name,mnesia_event}, {mfa,{mnesia_sup,start_event,[]}}, {restart_type,permanent}, {shutdown,30000}, {child_type,worker}] PROGRESS REPORT <6040.116.0> 2011-05-10 13:50:08 =============================================================================== supervisor {local,mnesia_kernel_sup} started [{pid,<6040.117.0>}, {name,mnesia_monitor}, {mfa,{mnesia_monitor,start,[]}}, {restart_type,permanent}, {shutdown,3000}, {child_type,worker}] PROGRESS REPORT <6040.116.0> 2011-05-10 13:50:08 =============================================================================== supervisor {local,mnesia_kernel_sup} started [{pid,<6040.118.0>}, {name,mnesia_subscr}, {mfa,{mnesia_subscr,start,[]}}, {restart_type,permanent}, {shutdown,3000}, {child_type,worker}] PROGRESS REPORT <6040.116.0> 2011-05-10 13:50:08 =============================================================================== supervisor {local,mnesia_kernel_sup} started [{pid,<6040.119.0>}, {name,mnesia_locker}, {mfa,{mnesia_locker,start,[]}}, {restart_type,permanent}, {shutdown,3000}, {child_type,worker}] PROGRESS REPORT <6040.116.0> 2011-05-10 13:50:08 =============================================================================== supervisor {local,mnesia_kernel_sup} started [{pid,<6040.120.0>}, {name,mnesia_recover}, {mfa,{mnesia_recover,start,[]}}, {restart_type,permanent}, {shutdown,180000}, {child_type,worker}] PROGRESS REPORT <6040.116.0> 2011-05-10 13:50:08 =============================================================================== supervisor {local,mnesia_kernel_sup} started [{pid,<6040.121.0>}, {name,mnesia_tm}, {mfa,{mnesia_tm,start,[]}}, {restart_type,permanent}, {shutdown,30000}, {child_type,worker}] PROGRESS REPORT <6040.116.0> 2011-05-10 13:50:08 =============================================================================== supervisor {local,mnesia_kernel_sup} started [{pid,<6040.122.0>}, {name,mnesia_checkpoint_sup}, {mfa,{mnesia_checkpoint_sup,start,[]}}, {restart_type,permanent}, {shutdown,infinity}, {child_type,supervisor}] PROGRESS REPORT <6040.116.0> 2011-05-10 13:50:08 =============================================================================== supervisor {local,mnesia_kernel_sup} started [{pid,<6040.123.0>}, {name,mnesia_snmp_sup}, {mfa,{mnesia_snmp_sup,start,[]}}, {restart_type,permanent}, {shutdown,infinity}, {child_type,supervisor}] PROGRESS REPORT <6040.116.0> 2011-05-10 13:50:08 =============================================================================== supervisor {local,mnesia_kernel_sup} started [{pid,<6040.124.0>}, {name,mnesia_controller}, {mfa,{mnesia_controller,start,[]}}, {restart_type,permanent}, {shutdown,3000}, {child_type,worker}] PROGRESS REPORT <6040.116.0> 2011-05-10 13:50:08 =============================================================================== supervisor {local,mnesia_kernel_sup} started [{pid,<6040.125.0>}, {name,mnesia_late_loader}, {mfa,{mnesia_late_loader,start,[]}}, {restart_type,permanent}, {shutdown,3000}, {child_type,worker}] PROGRESS REPORT <6040.114.0> 2011-05-10 13:50:08 =============================================================================== supervisor {local,mnesia_sup} started [{pid,<6040.116.0>}, {name,mnesia_kernel_sup}, {mfa,{mnesia_kernel_sup,start,[]}}, {restart_type,permanent}, {shutdown,infinity}, {child_type,supervisor}] PROGRESS REPORT <6040.7.0> 2011-05-10 13:50:08 =============================================================================== application mnesia started_at 'ns_1@127.0.0.1' PROGRESS REPORT <6040.27.0> 2011-05-10 13:50:08 =============================================================================== supervisor {local,kernel_safe_sup} started [{pid,<6040.131.0>}, {name,disk_log_sup}, {mfa,{disk_log_sup,start_link,[]}}, {restart_type,permanent}, {shutdown,1000}, {child_type,supervisor}] PROGRESS REPORT <6040.27.0> 2011-05-10 13:50:08 =============================================================================== supervisor {local,kernel_safe_sup} started [{pid,<6040.132.0>}, {name,disk_log_server}, {mfa,{disk_log_server,start_link,[]}}, {restart_type,permanent}, {shutdown,2000}, {child_type,worker}] PROGRESS REPORT <6040.27.0> 2011-05-10 13:50:08 =============================================================================== supervisor {local,kernel_safe_sup} started [{pid,<6040.135.0>}, {name,dets_sup}, {mfa,{dets_sup,start_link,[]}}, {restart_type,permanent}, {shutdown,1000}, {child_type,supervisor}] PROGRESS REPORT <6040.27.0> 2011-05-10 13:50:08 =============================================================================== supervisor {local,kernel_safe_sup} started [{pid,<6040.136.0>}, {name,dets}, {mfa,{dets_server,start_link,[]}}, {restart_type,permanent}, {shutdown,2000}, {child_type,worker}] INFO REPORT <6040.110.0> 2011-05-10 13:50:08 =============================================================================== ns_1@127.0.0.1:ns_mnesia:250: Committed schema to disk. INFO REPORT <6040.110.0> 2011-05-10 13:50:08 =============================================================================== ns_1@127.0.0.1:ns_mnesia:196: Current config: [{access_module,mnesia}, {auto_repair,true}, {backup_module,mnesia_backup}, {checkpoints,[]}, {db_nodes,['ns_1@127.0.0.1']}, {debug,verbose}, {directory, "/var/opt/membase/1.6.5.4r/mnesia"}, {dump_log_load_regulation, false}, {dump_log_time_threshold, 180000}, {dump_log_update_in_place,true}, {dump_log_write_threshold,1000}, {embedded_mnemosyne,false}, {event_module,mnesia_event}, {extra_db_nodes,[]}, {fallback_activated,false}, {held_locks,[]}, {ignore_fallback_at_startup, false}, {fallback_error_function, {mnesia,lkill}}, {is_running,yes}, {local_tables,[schema]}, {lock_queue,[]}, {log_version,"4.3"}, {master_node_tables,[]}, {max_wait_for_decision,10000}, {protocol_version,{7,6}}, {running_db_nodes, ['ns_1@127.0.0.1']}, {schema_location,opt_disc}, {schema_version,{3,0}}, {subscribers, [<6040.115.0>,<6040.110.0>]}, {tables,[schema]}, {transaction_commits,3}, {transaction_failures,0}, {transaction_log_writes,1}, {transaction_restarts,0}, {transactions,[]}, {use_dir,true}, {core_dir,false}, {no_table_loaders,2}, {dc_dump_limit,4}, {version,"4.4.12"}] INFO REPORT <6040.110.0> 2011-05-10 13:50:08 =============================================================================== ns_1@127.0.0.1:ns_mnesia:170: Mnesia table event: {write,schema, {schema,schema, [{name,schema}, {type,set}, {ram_copies,[]}, {disc_copies,['ns_1@127.0.0.1']}, {disc_only_copies,[]}, {load_order,0}, {access_mode,read_write}, {index,[]}, {snmp,[]}, {local_content,false}, {record_name,schema}, {attributes,[table,cstruct]}, {user_properties,[]}, {frag_properties,[]}, {cookie,{{1305,60608,606210},'ns_1@127.0.0.1'}}, {version,{{3,0},{'ns_1@127.0.0.1',{1305,60608,638527}}}}]}, [{schema,schema, [{name,schema}, {type,set}, {ram_copies,[]}, {disc_copies,[]}, {disc_only_copies,[]}, {load_order,0}, {access_mode,read_write}, {index,[]}, {snmp,[]}, {local_content,false}, {record_name,schema}, {attributes,[table,cstruct]}, {user_properties,[]}, {frag_properties,[]}, {cookie,{{1305,60608,606210},'ns_1@127.0.0.1'}}, {version,{{2,0},[]}}]}], {tid,3,<6040.127.0>}} PROGRESS REPORT <6040.70.0> 2011-05-10 13:50:08 =============================================================================== supervisor {local,ns_server_sup} started [{pid,<6040.110.0>}, {name,ns_mnesia}, {mfa,{ns_mnesia,start_link,[]}}, {restart_type,permanent}, {shutdown,10000}, {child_type,worker}] PROGRESS REPORT <6040.70.0> 2011-05-10 13:50:08 =============================================================================== supervisor {local,ns_server_sup} started [{pid,<6040.148.0>}, {name,ns_bad_bucket_worker}, {mfa,{work_queue,start_link,[ns_bad_bucket_worker]}}, {restart_type,permanent}, {shutdown,10}, {child_type,worker}] PROGRESS REPORT <6040.70.0> 2011-05-10 13:50:08 =============================================================================== supervisor {local,ns_server_sup} started [{pid,<6040.149.0>}, {name,ns_bad_bucket_sup}, {mfa,{ns_bucket_sup,start_link, [ns_bad_bucket_sup, #Fun, ns_bad_bucket_worker]}}, {restart_type,permanent}, {shutdown,infinity}, {child_type,supervisor}] PROGRESS REPORT <6040.70.0> 2011-05-10 13:50:08 =============================================================================== supervisor {local,ns_server_sup} started [{pid,<6040.150.0>}, {name,ns_moxi_sup}, {mfa,{ns_moxi_sup,start_link,[]}}, {restart_type,permanent}, {shutdown,infinity}, {child_type,supervisor}] INFO REPORT <6040.70.0> 2011-05-10 13:50:08 =============================================================================== ns_1@127.0.0.1:misc:725: start_singleton(gen_server, ns_tick, [], []): started as <6040.151.0> on 'ns_1@127.0.0.1' PROGRESS REPORT <6040.70.0> 2011-05-10 13:50:08 =============================================================================== supervisor {local,ns_server_sup} started [{pid,<6040.151.0>}, {name,ns_tick}, {mfa,{ns_tick,start_link,[]}}, {restart_type,permanent}, {shutdown,10}, {child_type,worker}] PROGRESS REPORT <6040.54.0> 2011-05-10 13:50:08 =============================================================================== supervisor {local,ns_server_cluster_sup} started [{pid,<6040.70.0>}, {name,ns_server_sup}, {mfa,{ns_server_sup,start_link,[]}}, {restart_type,permanent}, {shutdown,infinity}, {child_type,supervisor}] PROGRESS REPORT <6040.7.0> 2011-05-10 13:50:08 =============================================================================== application ns_server started_at 'ns_1@127.0.0.1' INFO REPORT <6040.64.0> 2011-05-10 13:50:13 =============================================================================== unsupervising port: {moxi,"./bin/moxi/moxi", ["-Z", "port_listen=11211,default_bucket_name=default,downstream_max=1024,downstream_conn_max=4,connect_max_errors=5,connect_retry_interval=30000,connect_timeout=400,auth_timeout=100,cycle=200,downstream_conn_queue_timeout=200,downstream_timeout=5000,wait_queue_timeout=200", "-z", "url=http://127.0.0.1:8091/pools/default/saslBucketsStreaming", "-p","0","-Y","y","-O","stderr",[]], [{env,[{"EVENT_NOSELECT","1"}, {"MOXI_SASL_PLAIN_USR",[]}, {"MOXI_SASL_PLAIN_PWD",[]}]}, use_stdio,stderr_to_stdout,stream]} INFO REPORT <6040.64.0> 2011-05-10 13:50:13 =============================================================================== supervising port: {moxi,"./bin/moxi/moxi", ["-Z", "port_listen=11211,default_bucket_name=default,downstream_max=1024,downstream_conn_max=4,connect_max_errors=5,connect_retry_interval=30000,connect_timeout=400,auth_timeout=100,cycle=200,downstream_conn_queue_timeout=200,downstream_timeout=5000,wait_queue_timeout=200", "-z", "url=http://127.0.0.1:8091/pools/default/saslBucketsStreaming", "-p","0","-Y","y","-O","stderr",[]], [{env,[{"EVENT_NOSELECT","1"}, {"MOXI_SASL_PLAIN_USR","Administrator"}, {"MOXI_SASL_PLAIN_PWD","password"}]}, use_stdio,stderr_to_stdout,stream]} INFO REPORT <6040.158.0> 2011-05-10 13:50:13 =============================================================================== starting ns_port_server with delay of 5000 PROGRESS REPORT <6040.99.0> 2011-05-10 13:50:13 =============================================================================== supervisor {local,ns_port_sup} started [{pid,<6040.158.0>}, {name, {moxi,"./bin/moxi/moxi", ["-Z", "port_listen=11211,default_bucket_name=default,downstream_max=1024,downstream_conn_max=4,connect_max_errors=5,connect_retry_interval=30000,connect_timeout=400,auth_timeout=100,cycle=200,downstream_conn_queue_timeout=200,downstream_timeout=5000,wait_queue_timeout=200", "-z", "url=http://127.0.0.1:8091/pools/default/saslBucketsStreaming", "-p","0","-Y","y","-O","stderr",[]], [{env, [{"EVENT_NOSELECT","1"}, {"MOXI_SASL_PLAIN_USR","Administrator"}, {"MOXI_SASL_PLAIN_PWD","password"}]}, use_stdio,stderr_to_stdout,stream]}}, {mfa, {supervisor_cushion,start_link, [moxi,5000,ns_port_server,start_link, [moxi,"./bin/moxi/moxi", ["-Z", "port_listen=11211,default_bucket_name=default,downstream_max=1024,downstream_conn_max=4,connect_max_errors=5,connect_retry_interval=30000,connect_timeout=400,auth_timeout=100,cycle=200,downstream_conn_queue_timeout=200,downstream_timeout=5000,wait_queue_timeout=200", "-z", "url=http://127.0.0.1:8091/pools/default/saslBucketsStreaming", "-p","0","-Y","y","-O","stderr",[]], [{env, [{"EVENT_NOSELECT","1"}, {"MOXI_SASL_PLAIN_USR","Administrator"}, {"MOXI_SASL_PLAIN_PWD","password"}]}, use_stdio,stderr_to_stdout,stream]]]}}, {restart_type,permanent}, {shutdown,10}, {child_type,worker}] INFO REPORT <6040.64.0> 2011-05-10 13:50:13 =============================================================================== config change: rest -> [{port,8091}] INFO REPORT <6040.64.0> 2011-05-10 13:50:13 =============================================================================== ns_node_disco_conf_events config all INFO REPORT <6040.85.0> 2011-05-10 13:50:13 =============================================================================== Pushing config INFO REPORT <6040.85.0> 2011-05-10 13:50:13 =============================================================================== Pushing config done INFO REPORT <6040.64.0> 2011-05-10 13:50:13 =============================================================================== config change: rest_creds -> ******** INFO REPORT <6040.64.0> 2011-05-10 13:50:13 =============================================================================== ns_node_disco_conf_events config all INFO REPORT <6040.85.0> 2011-05-10 13:50:13 =============================================================================== Pushing config INFO REPORT <6040.85.0> 2011-05-10 13:50:13 =============================================================================== Pushing config done INFO REPORT <6040.96.0> 2011-05-10 13:50:13 =============================================================================== menelaus_web streaming socket closed by client INFO REPORT <6040.159.0> 2011-05-10 13:50:14 =============================================================================== moxi<0.159.0>: 2011-05-10 13:50:13: (cproxy_config.c.316) env: MOXI_SASL_PLAIN_USR (13) moxi<0.159.0>: 2011-05-10 13:50:13: (cproxy_config.c.325) env: MOXI_SASL_PLAIN_PWD (8) INFO REPORT <6040.92.0> 2011-05-10 13:51:08 =============================================================================== ns_1@127.0.0.1:ns_doctor:82: Current node statuses: [{'ns_1@127.0.0.1', [{last_heard,{1305,60668,411608}}, {active_buckets,[]}, {memory, [{total,22468568}, {processes,4665416}, {processes_used,4652568}, {system,17803152}, {atom,689673}, {atom_used,685452}, {binary,26824}, {code,7338362}, {ets,523160}]}, {cluster_compatibility_version,1}, {version, [{kernel,"2.13.4"}, {mnesia,"4.4.12"}, {stdlib,"1.16.4"}, {os_mon,"2.2.4"}, {ns_server,"1.6.5.4r"}, {menelaus,"1.6.5.4r"}, {sasl,"2.1.8"}]}, {system_arch,"x86_64-unknown-linux-gnu"}, {wall_clock,60}, {memory_data,{2058731520,569856000,{<6040.70.0>,972712}}}, {disk_data, [{"/",11204616,49},{"/boot",101086,14},{"/dev/shm",1005240,0}]}, {meminfo, <<"MemTotal: 2010480 kB\nMemFree: 1453732 kB\nBuffers: 58768 kB\nCached: 333344 kB\nSwapCached: 20084 kB\nActive: 249080 kB\nInactive: 196040 kB\nHighTotal: 0 kB\nHighFree: 0 kB\nLowTotal: 2010480 kB\nLowFree: 1453732 kB\nSwapTotal: 4030456 kB\nSwapFree: 3952288 kB\nDirty: 68684 kB\nWriteback: 0 kB\nAnonPages: 50780 kB\nMapped: 13916 kB\nSlab: 86232 kB\nPageTables: 5872 kB\nNFS_Unstable: 0 kB\nBounce: 0 kB\nCommitLimit: 5035696 kB\nCommitted_AS: 390136 kB\nVmallocTotal: 34359738367 kB\nVmallocUsed: 265484 kB\nVmallocChunk: 34359471011 kB\nHugePages_Total: 0\nHugePages_Free: 0\nHugePages_Rsvd: 0\nHugepagesize: 2048 kB\n">>}, {replication,[]}, {system_memory_data, [{system_total_memory,2058731520}, {free_swap,4047142912}, {total_swap,4127186944}, {cached_memory,341344256}, {buffered_memory,60178432}, {free_memory,1488621568}, {total_memory,2058731520}]}, {statistics, [{wall_clock,{50579,1}}, {context_switches,{11765,0}}, {garbage_collection,{2088,5134474,0}}, {io,{{input,3976917},{output,258850}}}, {reductions,{1937846,9718}}, {run_queue,0}, {runtime,{530,0}}]}]}] INFO REPORT <6040.92.0> 2011-05-10 13:52:08 =============================================================================== ns_1@127.0.0.1:ns_doctor:82: Current node statuses: [{'ns_1@127.0.0.1', [{last_heard,{1305,60728,410632}}, {active_buckets,[]}, {memory, [{total,22235608}, {processes,4428264}, {processes_used,4415416}, {system,17807344}, {atom,689673}, {atom_used,685452}, {binary,30680}, {code,7338362}, {ets,523160}]}, {cluster_compatibility_version,1}, {version, [{kernel,"2.13.4"}, {mnesia,"4.4.12"}, {stdlib,"1.16.4"}, {os_mon,"2.2.4"}, {ns_server,"1.6.5.4r"}, {menelaus,"1.6.5.4r"}, {sasl,"2.1.8"}]}, {system_arch,"x86_64-unknown-linux-gnu"}, {wall_clock,120}, {memory_data,{2058731520,569602048,{<6040.70.0>,972712}}}, {disk_data, [{"/",11204616,49},{"/boot",101086,14},{"/dev/shm",1005240,0}]}, {meminfo, <<"MemTotal: 2010480 kB\nMemFree: 1454228 kB\nBuffers: 58808 kB\nCached: 333348 kB\nSwapCached: 20084 kB\nActive: 249240 kB\nInactive: 196080 kB\nHighTotal: 0 kB\nHighFree: 0 kB\nLowTotal: 2010480 kB\nLowFree: 1454228 kB\nSwapTotal: 4030456 kB\nSwapFree: 3952288 kB\nDirty: 46460 kB\nWriteback: 0 kB\nAnonPages: 50936 kB\nMapped: 13916 kB\nSlab: 85596 kB\nPageTables: 5872 kB\nNFS_Unstable: 0 kB\nBounce: 0 kB\nCommitLimit: 5035696 kB\nCommitted_AS: 393004 kB\nVmallocTotal: 34359738367 kB\nVmallocUsed: 265484 kB\nVmallocChunk: 34359471011 kB\nHugePages_Total: 0\nHugePages_Free: 0\nHugePages_Rsvd: 0\nHugepagesize: 2048 kB\n">>}, {replication,[]}, {system_memory_data, [{system_total_memory,2058731520}, {free_swap,4047142912}, {total_swap,4127186944}, {cached_memory,341348352}, {buffered_memory,60219392}, {free_memory,1489129472}, {total_memory,2058731520}]}, {statistics, [{wall_clock,{110579,1}}, {context_switches,{14190,0}}, {garbage_collection,{2497,6252788,0}}, {io,{{input,3983312},{output,265888}}}, {reductions,{2028234,9366}}, {run_queue,0}, {runtime,{560,10}}]}]}] INFO REPORT <6040.92.0> 2011-05-10 13:53:08 =============================================================================== ns_1@127.0.0.1:ns_doctor:82: Current node statuses: [{'ns_1@127.0.0.1', [{last_heard,{1305,60788,411496}}, {active_buckets,[]}, {memory, [{total,22520256}, {processes,4714664}, {processes_used,4701816}, {system,17805592}, {atom,689673}, {atom_used,685452}, {binary,28592}, {code,7338362}, {ets,523160}]}, {cluster_compatibility_version,1}, {version, [{kernel,"2.13.4"}, {mnesia,"4.4.12"}, {stdlib,"1.16.4"}, {os_mon,"2.2.4"}, {ns_server,"1.6.5.4r"}, {menelaus,"1.6.5.4r"}, {sasl,"2.1.8"}]}, {system_arch,"x86_64-unknown-linux-gnu"}, {wall_clock,180}, {memory_data,{2058731520,569475072,{<6040.70.0>,972712}}}, {disk_data, [{"/",11204616,49},{"/boot",101086,14},{"/dev/shm",1005240,0}]}, {meminfo, <<"MemTotal: 2010480 kB\nMemFree: 1454352 kB\nBuffers: 58848 kB\nCached: 333348 kB\nSwapCached: 20084 kB\nActive: 249116 kB\nInactive: 196104 kB\nHighTotal: 0 kB\nHighFree: 0 kB\nLowTotal: 2010480 kB\nLowFree: 1454352 kB\nSwapTotal: 4030456 kB\nSwapFree: 3952288 kB\nDirty: 38172 kB\nWriteback: 0 kB\nAnonPages: 50796 kB\nMapped: 13916 kB\nSlab: 85576 kB\nPageTables: 5872 kB\nNFS_Unstable: 0 kB\nBounce: 0 kB\nCommitLimit: 5035696 kB\nCommitted_AS: 390136 kB\nVmallocTotal: 34359738367 kB\nVmallocUsed: 265484 kB\nVmallocChunk: 34359471011 kB\nHugePages_Total: 0\nHugePages_Free: 0\nHugePages_Rsvd: 0\nHugepagesize: 2048 kB\n">>}, {replication,[]}, {system_memory_data, [{system_total_memory,2058731520}, {free_swap,4047142912}, {total_swap,4127186944}, {cached_memory,341348352}, {buffered_memory,60260352}, {free_memory,1489256448}, {total_memory,2058731520}]}, {statistics, [{wall_clock,{170579,1}}, {context_switches,{16619,0}}, {garbage_collection,{2926,7411319,0}}, {io,{{input,3989707},{output,272929}}}, {reductions,{2125333,9369}}, {run_queue,0}, {runtime,{620,10}}]}]}] INFO REPORT <6040.293.0> 2011-05-10 13:53:27 =============================================================================== ns_1@127.0.0.1:ns_storage_conf:279: Result of deleting file "/var/opt/membase/1.6.5.4r/data/ns_1/default": {error, enoent} INFO REPORT <6040.293.0> 2011-05-10 13:53:27 =============================================================================== ns_1@127.0.0.1:ns_storage_conf:279: Result of deleting file "/var/opt/membase/1.6.5.4r/data/ns_1/default-0.mb": {error, enoent} INFO REPORT <6040.293.0> 2011-05-10 13:53:27 =============================================================================== ns_1@127.0.0.1:ns_storage_conf:279: Result of deleting file "/var/opt/membase/1.6.5.4r/data/ns_1/default-1.mb": {error, enoent} INFO REPORT <6040.293.0> 2011-05-10 13:53:27 =============================================================================== ns_1@127.0.0.1:ns_storage_conf:279: Result of deleting file "/var/opt/membase/1.6.5.4r/data/ns_1/default-2.mb": {error, enoent} INFO REPORT <6040.293.0> 2011-05-10 13:53:27 =============================================================================== ns_1@127.0.0.1:ns_storage_conf:279: Result of deleting file "/var/opt/membase/1.6.5.4r/data/ns_1/default-3.mb": {error, enoent} INFO REPORT <6040.290.0> 2011-05-10 13:53:27 =============================================================================== ns_log: logging menelaus_web:12:Created bucket "default" of type: membase INFO REPORT <6040.64.0> 2011-05-10 13:53:27 =============================================================================== config change: buckets -> [{configs,[{"default", [{num_replicas,1}, {ram_quota,1096810496}, {auth_type,sasl}, {sasl_password,[]}, {type,membase}, {num_vbuckets,1024}, {db_shards,4}, {ht_size,3079}, {tap_keepalive,0}, {tap_noop_interval,20}, {max_txn_size,1000}, {ht_locks,5}, {servers,[]}, {map,undefined}]}]}] INFO REPORT <6040.64.0> 2011-05-10 13:53:27 =============================================================================== Writing isasl passwd file: "/var/opt/membase/1.6.5.4r/data/ns_1/isasl.pw" INFO REPORT <6040.64.0> 2011-05-10 13:53:27 =============================================================================== ns_node_disco_conf_events config all INFO REPORT <6040.85.0> 2011-05-10 13:53:27 =============================================================================== Pushing config INFO REPORT <6040.85.0> 2011-05-10 13:53:27 =============================================================================== Pushing config done INFO REPORT <6040.148.0> 2011-05-10 13:53:27 =============================================================================== ns_1@127.0.0.1:ns_bucket_sup:75: ns_bad_bucket_sup: Starting new child: {{stats_collector, "default"}, {stats_collector, start_link, ["default"]}, permanent, 10, worker, [stats_collector]} INFO REPORT <6040.107.0> 2011-05-10 13:53:27 =============================================================================== ns_1@127.0.0.1:ns_bucket_sup:75: ns_good_bucket_sup: Starting new child: {{ns_vbm_sup, "default"}, {ns_vbm_sup, start_link, ["default"]}, permanent, 1000, worker, [ns_vbm_sup]} INFO REPORT <6040.64.0> 2011-05-10 13:53:27 =============================================================================== config change: buckets -> [{configs,[{"default", [{num_replicas,1}, {ram_quota,1096810496}, {auth_type,sasl}, {sasl_password,[]}, {type,membase}, {num_vbuckets,1024}, {db_shards,4}, {ht_size,3079}, {tap_keepalive,0}, {tap_noop_interval,20}, {max_txn_size,1000}, {ht_locks,5}, {servers,['ns_1@127.0.0.1']}, {map,undefined}]}]}] INFO REPORT <6040.64.0> 2011-05-10 13:53:27 =============================================================================== ns_node_disco_conf_events config all INFO REPORT <6040.85.0> 2011-05-10 13:53:27 =============================================================================== Pushing config INFO REPORT <6040.85.0> 2011-05-10 13:53:27 =============================================================================== Pushing config done INFO REPORT <6040.64.0> 2011-05-10 13:53:27 =============================================================================== config change: buckets -> [{configs,[{"default", [{num_replicas,1}, {ram_quota,1096810496}, {auth_type,sasl}, {sasl_password,[]}, {type,membase}, {num_vbuckets,1024}, {db_shards,4}, {ht_size,3079}, {tap_keepalive,0}, {tap_noop_interval,20}, {max_txn_size,1000}, {ht_locks,5}, {servers,['ns_1@127.0.0.1']}, {map,[['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1'|...], [...]|...]}]}]}] INFO REPORT <6040.64.0> 2011-05-10 13:53:27 =============================================================================== ns_node_disco_conf_events config all INFO REPORT <6040.85.0> 2011-05-10 13:53:27 =============================================================================== Pushing config INFO REPORT <6040.85.0> 2011-05-10 13:53:27 =============================================================================== Pushing config done PROGRESS REPORT <6040.149.0> 2011-05-10 13:53:27 =============================================================================== supervisor {local,ns_bad_bucket_sup} started [{pid,<6040.300.0>}, {name,{stats_collector,"default"}}, {mfa,{stats_collector,start_link,["default"]}}, {restart_type,permanent}, {shutdown,10}, {child_type,worker}] INFO REPORT <6040.148.0> 2011-05-10 13:53:27 =============================================================================== ns_1@127.0.0.1:ns_bucket_sup:75: ns_bad_bucket_sup: Starting new child: {{stats_archiver, "default"}, {stats_archiver, start_link, ["default"]}, permanent, 10, worker, [stats_archiver]} PROGRESS REPORT <6040.108.0> 2011-05-10 13:53:27 =============================================================================== supervisor {local,ns_good_bucket_sup} started [{pid,<6040.302.0>}, {name,{ns_vbm_sup,"default"}}, {mfa,{ns_vbm_sup,start_link,["default"]}}, {restart_type,permanent}, {shutdown,1000}, {child_type,worker}] INFO REPORT <6040.107.0> 2011-05-10 13:53:27 =============================================================================== ns_1@127.0.0.1:ns_bucket_sup:75: ns_good_bucket_sup: Starting new child: {{ns_memcached, "default"}, {ns_memcached, start_link, ["default"]}, permanent, 86400000, worker, [ns_memcached]} PROGRESS REPORT <6040.108.0> 2011-05-10 13:53:27 =============================================================================== supervisor {local,ns_good_bucket_sup} started [{pid,<6040.303.0>}, {name,{ns_memcached,"default"}}, {mfa,{ns_memcached,start_link,["default"]}}, {restart_type,permanent}, {shutdown,86400000}, {child_type,worker}] PROGRESS REPORT <6040.304.0> 2011-05-10 13:53:27 =============================================================================== supervisor {local,inet_gethost_native_sup} started [{pid,<6040.305.0>},{mfa,{inet_gethost_native,init,[[]]}}] PROGRESS REPORT <6040.27.0> 2011-05-10 13:53:27 =============================================================================== supervisor {local,kernel_safe_sup} started [{pid,<6040.304.0>}, {name,inet_gethost_native_sup}, {mfa,{inet_gethost_native,start_link,[]}}, {restart_type,temporary}, {shutdown,1000}, {child_type,worker}] ERROR REPORT <6040.173.0> 2011-05-10 13:53:27 =============================================================================== ns_1@127.0.0.1:stats_reader:176: Some nodes didn't respond: ['ns_1@127.0.0.1'] INFO REPORT <6040.110.0> 2011-05-10 13:53:27 =============================================================================== ns_1@127.0.0.1:ns_mnesia:120: Created table 'stats_archiver-default-minute' INFO REPORT <6040.110.0> 2011-05-10 13:53:27 =============================================================================== ns_1@127.0.0.1:ns_mnesia:170: Mnesia table event: {write,schema, {schema,'stats_archiver-default-minute', [{name,'stats_archiver-default-minute'}, {type,ordered_set}, {ram_copies,[]}, {disc_copies,['ns_1@127.0.0.1']}, {disc_only_copies,[]}, {load_order,0}, {access_mode,read_write}, {index,[]}, {snmp,[]}, {local_content,true}, {record_name,stat_entry}, {attributes,[timestamp,values]}, {user_properties,[]}, {frag_properties,[]}, {cookie,{{1305,60807,542374},'ns_1@127.0.0.1'}}, {version,{{2,0},[]}}]}, [], {tid,4,<6040.307.0>}} INFO REPORT <6040.110.0> 2011-05-10 13:53:27 =============================================================================== ns_1@127.0.0.1:ns_mnesia:170: Mnesia table event: {write,schema, {schema,'stats_archiver-default-minute', [{name,'stats_archiver-default-minute'}, {type,ordered_set}, {ram_copies,[]}, {disc_copies,['ns_1@127.0.0.1']}, {disc_only_copies,[]}, {load_order,0}, {access_mode,read_write}, {index,[]}, {snmp,[]}, {local_content,true}, {record_name,stat_entry}, {attributes,[timestamp,values]}, {user_properties,[]}, {frag_properties,[]}, {cookie,{{1305,60807,542374},'ns_1@127.0.0.1'}}, {version,{{2,0},[]}}]}, [{schema,'stats_archiver-default-minute', [{name,'stats_archiver-default-minute'}, {type,ordered_set}, {ram_copies,[]}, {disc_copies,['ns_1@127.0.0.1']}, {disc_only_copies,[]}, {load_order,0}, {access_mode,read_write}, {index,[]}, {snmp,[]}, {local_content,true}, {record_name,stat_entry}, {attributes,[timestamp,values]}, {user_properties,[]}, {frag_properties,[]}, {cookie,{{1305,60807,542374},'ns_1@127.0.0.1'}}, {version,{{2,0},[]}}]}], {tid,4,<6040.307.0>}} INFO REPORT <6040.110.0> 2011-05-10 13:53:27 =============================================================================== ns_1@127.0.0.1:ns_mnesia:120: Created table 'stats_archiver-default-hour' INFO REPORT <6040.110.0> 2011-05-10 13:53:27 =============================================================================== ns_1@127.0.0.1:ns_mnesia:170: Mnesia table event: {write,schema, {schema,'stats_archiver-default-hour', [{name,'stats_archiver-default-hour'}, {type,ordered_set}, {ram_copies,[]}, {disc_copies,['ns_1@127.0.0.1']}, {disc_only_copies,[]}, {load_order,0}, {access_mode,read_write}, {index,[]}, {snmp,[]}, {local_content,true}, {record_name,stat_entry}, {attributes,[timestamp,values]}, {user_properties,[]}, {frag_properties,[]}, {cookie,{{1305,60807,821627},'ns_1@127.0.0.1'}}, {version,{{2,0},[]}}]}, [], {tid,5,<6040.317.0>}} INFO REPORT <6040.110.0> 2011-05-10 13:53:27 =============================================================================== ns_1@127.0.0.1:ns_mnesia:170: Mnesia table event: {write,schema, {schema,'stats_archiver-default-hour', [{name,'stats_archiver-default-hour'}, {type,ordered_set}, {ram_copies,[]}, {disc_copies,['ns_1@127.0.0.1']}, {disc_only_copies,[]}, {load_order,0}, {access_mode,read_write}, {index,[]}, {snmp,[]}, {local_content,true}, {record_name,stat_entry}, {attributes,[timestamp,values]}, {user_properties,[]}, {frag_properties,[]}, {cookie,{{1305,60807,821627},'ns_1@127.0.0.1'}}, {version,{{2,0},[]}}]}, [{schema,'stats_archiver-default-hour', [{name,'stats_archiver-default-hour'}, {type,ordered_set}, {ram_copies,[]}, {disc_copies,['ns_1@127.0.0.1']}, {disc_only_copies,[]}, {load_order,0}, {access_mode,read_write}, {index,[]}, {snmp,[]}, {local_content,true}, {record_name,stat_entry}, {attributes,[timestamp,values]}, {user_properties,[]}, {frag_properties,[]}, {cookie,{{1305,60807,821627},'ns_1@127.0.0.1'}}, {version,{{2,0},[]}}]}], {tid,5,<6040.317.0>}} INFO REPORT <6040.110.0> 2011-05-10 13:53:27 =============================================================================== ns_1@127.0.0.1:ns_mnesia:120: Created table 'stats_archiver-default-day' INFO REPORT <6040.110.0> 2011-05-10 13:53:27 =============================================================================== ns_1@127.0.0.1:ns_mnesia:170: Mnesia table event: {write,schema, {schema,'stats_archiver-default-day', [{name,'stats_archiver-default-day'}, {type,ordered_set}, {ram_copies,[]}, {disc_copies,['ns_1@127.0.0.1']}, {disc_only_copies,[]}, {load_order,0}, {access_mode,read_write}, {index,[]}, {snmp,[]}, {local_content,true}, {record_name,stat_entry}, {attributes,[timestamp,values]}, {user_properties,[]}, {frag_properties,[]}, {cookie,{{1305,60807,873496},'ns_1@127.0.0.1'}}, {version,{{2,0},[]}}]}, [], {tid,6,<6040.323.0>}} INFO REPORT <6040.110.0> 2011-05-10 13:53:27 =============================================================================== ns_1@127.0.0.1:ns_mnesia:170: Mnesia table event: {write,schema, {schema,'stats_archiver-default-day', [{name,'stats_archiver-default-day'}, {type,ordered_set}, {ram_copies,[]}, {disc_copies,['ns_1@127.0.0.1']}, {disc_only_copies,[]}, {load_order,0}, {access_mode,read_write}, {index,[]}, {snmp,[]}, {local_content,true}, {record_name,stat_entry}, {attributes,[timestamp,values]}, {user_properties,[]}, {frag_properties,[]}, {cookie,{{1305,60807,873496},'ns_1@127.0.0.1'}}, {version,{{2,0},[]}}]}, [{schema,'stats_archiver-default-day', [{name,'stats_archiver-default-day'}, {type,ordered_set}, {ram_copies,[]}, {disc_copies,['ns_1@127.0.0.1']}, {disc_only_copies,[]}, {load_order,0}, {access_mode,read_write}, {index,[]}, {snmp,[]}, {local_content,true}, {record_name,stat_entry}, {attributes,[timestamp,values]}, {user_properties,[]}, {frag_properties,[]}, {cookie,{{1305,60807,873496},'ns_1@127.0.0.1'}}, {version,{{2,0},[]}}]}], {tid,6,<6040.323.0>}} INFO REPORT <6040.110.0> 2011-05-10 13:53:27 =============================================================================== ns_1@127.0.0.1:ns_mnesia:120: Created table 'stats_archiver-default-week' INFO REPORT <6040.110.0> 2011-05-10 13:53:27 =============================================================================== ns_1@127.0.0.1:ns_mnesia:170: Mnesia table event: {write,schema, {schema,'stats_archiver-default-week', [{name,'stats_archiver-default-week'}, {type,ordered_set}, {ram_copies,[]}, {disc_copies,['ns_1@127.0.0.1']}, {disc_only_copies,[]}, {load_order,0}, {access_mode,read_write}, {index,[]}, {snmp,[]}, {local_content,true}, {record_name,stat_entry}, {attributes,[timestamp,values]}, {user_properties,[]}, {frag_properties,[]}, {cookie,{{1305,60807,917315},'ns_1@127.0.0.1'}}, {version,{{2,0},[]}}]}, [], {tid,7,<6040.329.0>}} INFO REPORT <6040.110.0> 2011-05-10 13:53:27 =============================================================================== ns_1@127.0.0.1:ns_mnesia:170: Mnesia table event: {write,schema, {schema,'stats_archiver-default-week', [{name,'stats_archiver-default-week'}, {type,ordered_set}, {ram_copies,[]}, {disc_copies,['ns_1@127.0.0.1']}, {disc_only_copies,[]}, {load_order,0}, {access_mode,read_write}, {index,[]}, {snmp,[]}, {local_content,true}, {record_name,stat_entry}, {attributes,[timestamp,values]}, {user_properties,[]}, {frag_properties,[]}, {cookie,{{1305,60807,917315},'ns_1@127.0.0.1'}}, {version,{{2,0},[]}}]}, [{schema,'stats_archiver-default-week', [{name,'stats_archiver-default-week'}, {type,ordered_set}, {ram_copies,[]}, {disc_copies,['ns_1@127.0.0.1']}, {disc_only_copies,[]}, {load_order,0}, {access_mode,read_write}, {index,[]}, {snmp,[]}, {local_content,true}, {record_name,stat_entry}, {attributes,[timestamp,values]}, {user_properties,[]}, {frag_properties,[]}, {cookie,{{1305,60807,917315},'ns_1@127.0.0.1'}}, {version,{{2,0},[]}}]}], {tid,7,<6040.329.0>}} INFO REPORT <6040.110.0> 2011-05-10 13:53:27 =============================================================================== ns_1@127.0.0.1:ns_mnesia:120: Created table 'stats_archiver-default-month' INFO REPORT <6040.110.0> 2011-05-10 13:53:27 =============================================================================== ns_1@127.0.0.1:ns_mnesia:170: Mnesia table event: {write,schema, {schema,'stats_archiver-default-month', [{name,'stats_archiver-default-month'}, {type,ordered_set}, {ram_copies,[]}, {disc_copies,['ns_1@127.0.0.1']}, {disc_only_copies,[]}, {load_order,0}, {access_mode,read_write}, {index,[]}, {snmp,[]}, {local_content,true}, {record_name,stat_entry}, {attributes,[timestamp,values]}, {user_properties,[]}, {frag_properties,[]}, {cookie,{{1305,60807,947680},'ns_1@127.0.0.1'}}, {version,{{2,0},[]}}]}, [], {tid,8,<6040.335.0>}} INFO REPORT <6040.110.0> 2011-05-10 13:53:27 =============================================================================== ns_1@127.0.0.1:ns_mnesia:170: Mnesia table event: {write,schema, {schema,'stats_archiver-default-month', [{name,'stats_archiver-default-month'}, {type,ordered_set}, {ram_copies,[]}, {disc_copies,['ns_1@127.0.0.1']}, {disc_only_copies,[]}, {load_order,0}, {access_mode,read_write}, {index,[]}, {snmp,[]}, {local_content,true}, {record_name,stat_entry}, {attributes,[timestamp,values]}, {user_properties,[]}, {frag_properties,[]}, {cookie,{{1305,60807,947680},'ns_1@127.0.0.1'}}, {version,{{2,0},[]}}]}, [{schema,'stats_archiver-default-month', [{name,'stats_archiver-default-month'}, {type,ordered_set}, {ram_copies,[]}, {disc_copies,['ns_1@127.0.0.1']}, {disc_only_copies,[]}, {load_order,0}, {access_mode,read_write}, {index,[]}, {snmp,[]}, {local_content,true}, {record_name,stat_entry}, {attributes,[timestamp,values]}, {user_properties,[]}, {frag_properties,[]}, {cookie,{{1305,60807,947680},'ns_1@127.0.0.1'}}, {version,{{2,0},[]}}]}], {tid,8,<6040.335.0>}} INFO REPORT <6040.110.0> 2011-05-10 13:53:28 =============================================================================== ns_1@127.0.0.1:ns_mnesia:120: Created table 'stats_archiver-default-year' INFO REPORT <6040.110.0> 2011-05-10 13:53:28 =============================================================================== ns_1@127.0.0.1:ns_mnesia:170: Mnesia table event: {write,schema, {schema,'stats_archiver-default-year', [{name,'stats_archiver-default-year'}, {type,ordered_set}, {ram_copies,[]}, {disc_copies,['ns_1@127.0.0.1']}, {disc_only_copies,[]}, {load_order,0}, {access_mode,read_write}, {index,[]}, {snmp,[]}, {local_content,true}, {record_name,stat_entry}, {attributes,[timestamp,values]}, {user_properties,[]}, {frag_properties,[]}, {cookie,{{1305,60807,974400},'ns_1@127.0.0.1'}}, {version,{{2,0},[]}}]}, [], {tid,9,<6040.341.0>}} INFO REPORT <6040.110.0> 2011-05-10 13:53:28 =============================================================================== ns_1@127.0.0.1:ns_mnesia:170: Mnesia table event: {write,schema, {schema,'stats_archiver-default-year', [{name,'stats_archiver-default-year'}, {type,ordered_set}, {ram_copies,[]}, {disc_copies,['ns_1@127.0.0.1']}, {disc_only_copies,[]}, {load_order,0}, {access_mode,read_write}, {index,[]}, {snmp,[]}, {local_content,true}, {record_name,stat_entry}, {attributes,[timestamp,values]}, {user_properties,[]}, {frag_properties,[]}, {cookie,{{1305,60807,974400},'ns_1@127.0.0.1'}}, {version,{{2,0},[]}}]}, [{schema,'stats_archiver-default-year', [{name,'stats_archiver-default-year'}, {type,ordered_set}, {ram_copies,[]}, {disc_copies,['ns_1@127.0.0.1']}, {disc_only_copies,[]}, {load_order,0}, {access_mode,read_write}, {index,[]}, {snmp,[]}, {local_content,true}, {record_name,stat_entry}, {attributes,[timestamp,values]}, {user_properties,[]}, {frag_properties,[]}, {cookie,{{1305,60807,974400},'ns_1@127.0.0.1'}}, {version,{{2,0},[]}}]}], {tid,9,<6040.341.0>}} PROGRESS REPORT <6040.149.0> 2011-05-10 13:53:28 =============================================================================== supervisor {local,ns_bad_bucket_sup} started [{pid,<6040.306.0>}, {name,{stats_archiver,"default"}}, {mfa,{stats_archiver,start_link,["default"]}}, {restart_type,permanent}, {shutdown,10}, {child_type,worker}] INFO REPORT <6040.148.0> 2011-05-10 13:53:28 =============================================================================== ns_1@127.0.0.1:ns_bucket_sup:75: ns_bad_bucket_sup: Starting new child: {{stats_reader, "default"}, {stats_reader, start_link, ["default"]}, permanent, 10, worker, [stats_reader]} PROGRESS REPORT <6040.149.0> 2011-05-10 13:53:28 =============================================================================== supervisor {local,ns_bad_bucket_sup} started [{pid,<6040.347.0>}, {name,{stats_reader,"default"}}, {mfa,{stats_reader,start_link,["default"]}}, {restart_type,permanent}, {shutdown,10}, {child_type,worker}] INFO REPORT <6040.295.0> 2011-05-10 13:53:28 =============================================================================== ns_1@127.0.0.1:ns_janitor:270: Waiting for "default" on ['ns_1@127.0.0.1'] INFO REPORT <6040.159.0> 2011-05-10 13:53:28 =============================================================================== moxi<0.159.0>: 2011-05-10 13:53:29: (agent_config.c.650) ERROR: bad JSON configuration: Empty serverList ({ moxi<0.159.0>: "name": "default", moxi<0.159.0>: "nodeLocator": "vbucket", moxi<0.159.0>: "saslPassword": "", moxi<0.159.0>: "nodes": [], moxi<0.159.0>: "vBucketServerMap": { moxi<0.159.0>: "hashAlgorithm": "CRC", moxi<0.159.0>: "numReplicas": 1, moxi<0.159.0>: "serverList": [], moxi<0.159.0>: "vBucketMap": [[-1, -1], [-1, -1], [-1, -1], [-1, -1], [-1, -1], [-1, -1], [-1, -1], [-1, -1], [-1, -1], [-1, -1], [-1, -1], [-1, -1], [-1, -1], [-1, -1], [-1, -1], [-1, -1], [-1, -1], [-1, -1], [-1, -1], [-1, -1], [-1, -1], [-1, -1], [-1, -1], [-1, -1], [-1, -1], [-1, -1], [-1, -1], [-1, -1], [-1, -1], [-1, -1], [-1, -1], [-1, -1], [-1, -1], [-1, -1], [-1, -1], [-1, -1], [-1, -1], [-1, -1], [-1, -1], [-1, -1], [-1, -1], [-1, -1], [-1, -1], [-1, -1], [-1, -1], [-1, -1], [-1, -1], [-1, -1], [-1, -1], [-1, -1], [-1, -1], [-1, -1], [-1, -1], [-1, -1], [-1, -1], [-1, -1], [-1, -1], [-1, -1], [-1, -1], [-1, -1], [-1, -1], [-1, -1], [-1, -1], [-1, -1], [-1, -1], [-1, -1], [-1, -1], [-1, -1], [-1, -1], [-1, -1], [-1, -1], [-1, -1 SUPERVISOR REPORT <6040.94.0> 2011-05-10 13:53:28 =============================================================================== Reporting supervisor {local,menelaus_sup} Child process errorContext child_terminated reason {noproc,{gen_server,call,['ns_memcached-default',topkeys,30000]}} pid <6040.98.0> name hot_keys_keeper start_function {hot_keys_keeper,start_link,[]} restart_type permanent shutdown 5000 child_type worker PROGRESS REPORT <6040.94.0> 2011-05-10 13:53:28 =============================================================================== supervisor {local,menelaus_sup} started [{pid,<6040.349.0>}, {name,hot_keys_keeper}, {mfa,{hot_keys_keeper,start_link,[]}}, {restart_type,permanent}, {shutdown,5000}, {child_type,worker}] INFO REPORT <6040.109.0> 2011-05-10 13:53:28 =============================================================================== ns_1@127.0.0.1:ns_orchestrator:164: Skipping janitor in state janitor_running: {janitor_state, ["default"], <6040.295.0>} INFO REPORT <6040.104.0> 2011-05-10 13:53:28 =============================================================================== memcached<0.104.0>: WARNING: Found duplicate entry for "tap_keepalive" INFO REPORT <6040.303.0> 2011-05-10 13:53:29 =============================================================================== ns_1@127.0.0.1:ns_memcached:390: Created bucket "default" with config string "vb0=false;waitforwarmup=false;ht_size=3079;ht_locks=5;failpartialwarmup=false;db_shards=4;shardpattern=%d/%b-%i.mb;db_strategy=multiMTVBDB;tap_keepalive=0;tap_noop_interval=20;max_txn_size=1000;max_size=1096810496;initfile=/etc/opt/membase/1.6.5.4r/init.sql;tap_keepalive=300;dbname=/var/opt/membase/1.6.5.4r/data/ns_1/default;" INFO REPORT <6040.295.0> 2011-05-10 13:53:29 =============================================================================== ns_1@127.0.0.1:ns_janitor:270: Waiting for "default" on ['ns_1@127.0.0.1'] INFO REPORT <6040.303.0> 2011-05-10 13:53:30 =============================================================================== ns_log: logging ns_memcached:1:Bucket "default" loaded on node 'ns_1@127.0.0.1' in 2 seconds. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:270: Waiting for "default" on ['ns_1@127.0.0.1'] INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 0 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 1 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 2 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 3 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 4 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 5 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 6 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 7 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 8 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 9 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 10 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 11 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 12 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 13 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 14 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 15 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 16 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 17 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 18 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 19 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 20 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 21 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 22 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 23 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 24 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 25 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 26 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 27 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 28 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 29 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 30 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 31 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 32 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 33 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 34 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 35 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 36 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 37 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 38 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 39 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 40 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 41 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 42 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 43 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 44 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 45 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 46 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 47 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 48 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 49 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 50 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 51 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 52 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 53 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 54 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 55 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 56 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 57 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 58 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 59 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 60 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 61 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 62 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 63 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 64 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 65 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 66 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 67 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 68 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 69 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 70 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 71 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 72 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 73 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 74 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 75 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 76 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 77 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 78 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 79 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 80 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 81 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 82 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 83 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 84 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 85 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 86 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 87 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 88 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 89 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 90 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 91 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 92 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 93 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 94 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 95 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 96 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 97 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 98 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 99 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 100 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 101 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 102 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 103 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 104 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 105 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 106 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 107 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 108 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 109 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 110 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 111 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 112 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 113 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 114 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 115 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 116 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 117 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 118 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 119 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 120 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 121 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 122 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 123 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 124 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 125 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 126 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 127 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 128 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 129 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 130 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 131 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 132 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 133 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 134 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 135 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 136 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 137 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 138 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 139 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 140 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 141 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 142 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 143 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 144 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 145 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 146 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 147 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 148 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 149 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 150 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 151 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 152 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 153 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 154 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 155 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 156 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 157 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 158 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 159 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 160 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 161 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 162 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 163 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 164 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 165 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 166 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 167 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 168 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 169 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 170 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 171 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 172 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 173 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 174 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 175 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 176 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 177 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 178 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 179 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 180 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 181 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 182 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 183 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 184 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 185 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 186 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 187 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 188 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 189 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 190 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 191 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 192 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 193 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 194 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 195 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 196 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 197 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 198 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 199 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 200 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 201 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 202 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 203 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 204 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 205 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 206 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 207 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 208 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 209 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 210 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 211 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 212 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 213 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 214 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 215 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 216 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 217 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 218 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 219 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 220 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 221 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 222 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 223 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 224 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 225 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 226 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 227 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 228 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 229 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 230 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 231 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 232 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 233 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 234 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 235 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 236 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 237 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 238 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 239 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 240 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 241 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 242 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 243 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 244 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 245 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 246 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 247 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 248 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 249 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 250 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 251 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 252 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 253 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 254 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 255 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 256 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 257 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 258 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 259 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 260 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 261 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 262 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 263 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 264 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 265 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 266 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 267 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 268 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 269 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 270 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 271 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 272 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 273 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 274 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 275 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 276 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 277 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 278 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 279 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 280 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 281 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 282 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 283 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 284 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 285 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 286 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 287 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 288 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 289 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 290 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 291 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 292 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 293 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 294 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 295 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 296 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 297 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 298 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 299 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 300 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 301 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 302 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 303 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 304 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 305 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 306 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 307 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 308 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 309 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 310 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 311 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 312 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 313 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 314 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 315 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 316 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 317 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 318 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 319 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 320 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 321 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 322 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 323 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 324 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 325 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 326 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 327 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 328 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 329 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 330 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 331 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 332 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 333 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 334 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 335 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 336 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 337 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 338 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 339 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 340 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 341 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 342 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 343 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 344 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 345 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 346 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 347 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 348 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 349 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 350 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 351 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 352 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 353 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 354 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 355 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 356 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 357 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 358 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 359 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 360 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 361 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 362 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 363 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 364 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 365 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 366 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 367 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 368 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 369 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 370 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 371 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 372 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 373 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 374 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 375 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 376 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 377 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 378 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 379 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 380 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 381 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 382 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 383 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 384 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 385 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 386 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 387 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 388 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 389 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 390 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 391 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 392 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 393 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 394 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 395 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 396 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 397 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 398 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 399 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 400 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 401 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 402 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 403 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 404 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 405 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 406 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 407 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 408 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 409 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 410 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 411 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 412 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 413 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 414 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 415 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 416 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 417 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 418 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 419 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 420 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 421 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 422 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 423 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 424 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 425 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 426 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 427 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 428 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 429 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 430 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 431 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 432 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 433 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 434 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 435 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 436 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 437 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 438 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 439 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 440 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 441 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 442 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 443 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 444 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 445 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 446 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 447 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 448 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 449 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 450 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 451 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 452 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 453 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 454 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 455 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 456 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 457 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 458 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 459 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 460 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 461 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 462 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 463 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 464 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 465 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 466 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 467 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 468 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 469 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 470 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 471 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 472 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 473 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 474 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 475 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 476 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 477 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 478 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 479 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 480 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 481 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 482 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 483 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 484 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 485 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 486 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 487 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 488 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 489 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 490 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 491 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 492 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 493 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 494 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 495 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 496 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 497 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 498 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 499 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 500 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 501 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 502 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 503 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 504 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 505 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 506 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 507 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 508 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 509 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 510 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 511 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 512 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 513 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 514 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 515 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 516 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 517 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 518 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 519 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 520 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 521 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 522 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 523 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 524 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 525 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 526 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 527 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 528 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 529 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 530 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 531 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 532 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 533 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 534 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 535 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 536 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 537 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 538 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 539 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 540 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 541 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 542 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 543 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 544 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 545 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 546 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 547 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 548 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 549 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 550 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 551 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 552 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 553 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 554 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 555 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 556 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 557 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 558 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 559 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 560 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 561 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 562 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 563 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 564 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 565 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 566 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 567 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 568 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 569 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 570 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 571 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 572 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 573 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 574 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 575 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 576 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 577 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 578 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 579 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 580 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 581 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 582 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 583 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 584 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 585 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 586 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 587 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 588 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 589 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 590 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 591 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 592 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 593 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 594 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 595 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 596 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 597 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 598 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 599 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 600 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 601 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 602 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 603 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 604 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 605 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 606 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 607 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 608 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 609 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 610 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 611 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 612 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 613 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 614 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 615 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 616 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 617 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 618 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 619 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 620 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 621 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 622 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 623 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 624 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 625 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 626 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 627 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 628 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 629 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 630 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 631 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 632 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 633 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 634 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 635 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 636 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 637 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 638 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 639 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 640 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 641 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 642 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 643 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 644 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 645 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 646 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 647 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 648 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 649 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 650 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 651 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 652 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 653 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 654 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 655 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 656 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 657 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 658 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 659 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 660 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 661 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 662 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 663 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 664 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 665 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 666 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 667 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 668 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 669 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 670 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 671 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 672 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 673 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 674 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 675 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 676 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 677 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 678 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 679 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 680 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 681 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 682 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 683 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 684 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 685 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 686 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 687 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 688 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 689 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 690 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 691 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 692 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 693 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 694 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 695 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 696 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 697 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 698 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 699 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 700 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 701 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 702 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 703 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 704 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 705 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 706 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 707 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 708 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 709 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 710 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 711 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 712 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 713 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 714 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 715 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 716 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 717 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 718 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 719 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 720 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 721 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 722 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 723 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 724 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 725 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 726 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 727 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 728 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 729 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 730 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 731 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 732 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 733 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 734 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 735 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 736 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 737 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 738 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 739 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 740 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 741 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 742 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 743 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 744 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 745 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 746 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 747 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 748 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 749 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 750 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 751 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 752 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 753 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 754 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 755 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 756 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 757 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 758 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 759 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 760 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 761 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 762 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 763 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 764 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 765 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 766 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 767 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 768 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 769 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 770 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 771 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 772 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 773 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 774 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 775 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 776 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 777 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 778 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 779 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 780 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 781 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 782 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 783 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 784 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 785 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 786 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 787 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 788 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 789 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 790 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 791 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 792 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 793 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 794 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 795 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 796 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 797 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 798 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 799 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 800 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 801 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 802 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 803 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 804 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 805 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 806 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 807 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 808 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 809 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 810 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 811 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 812 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 813 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 814 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 815 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 816 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 817 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 818 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 819 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 820 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 821 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 822 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 823 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 824 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 825 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 826 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 827 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 828 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 829 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 830 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 831 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 832 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 833 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 834 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 835 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 836 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 837 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 838 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 839 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 840 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 841 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 842 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 843 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 844 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 845 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 846 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 847 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 848 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 849 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 850 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 851 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 852 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 853 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 854 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 855 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 856 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 857 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 858 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 859 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 860 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 861 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 862 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 863 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 864 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 865 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 866 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 867 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 868 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 869 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 870 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 871 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 872 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 873 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 874 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 875 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 876 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 877 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 878 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 879 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 880 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 881 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 882 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 883 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 884 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 885 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 886 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 887 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 888 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 889 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 890 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 891 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 892 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 893 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 894 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 895 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 896 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 897 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 898 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 899 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 900 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 901 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 902 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 903 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 904 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 905 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 906 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 907 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 908 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 909 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 910 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 911 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 912 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 913 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 914 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 915 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 916 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 917 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 918 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 919 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 920 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 921 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 922 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 923 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 924 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 925 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 926 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 927 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 928 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 929 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 930 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 931 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 932 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 933 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 934 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 935 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 936 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 937 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 938 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 939 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 940 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 941 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 942 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 943 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 944 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 945 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 946 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 947 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 948 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 949 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 950 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 951 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 952 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 953 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 954 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 955 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 956 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 957 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 958 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 959 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 960 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 961 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 962 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 963 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 964 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 965 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 966 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 967 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 968 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 969 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 970 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 971 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 972 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 973 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 974 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 975 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 976 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 977 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 978 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 979 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 980 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 981 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 982 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 983 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 984 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 985 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 986 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 987 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 988 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 989 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 990 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 991 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 992 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 993 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 994 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 995 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 996 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 997 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 998 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 999 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 1000 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 1001 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 1002 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 1003 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 1004 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 1005 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 1006 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 1007 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 1008 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 1009 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 1010 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 1011 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 1012 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 1013 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 1014 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 1015 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 1016 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 1017 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 1018 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 1019 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 1020 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 1021 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 1022 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.295.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 1023 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6040.300.0> 2011-05-10 13:53:30 =============================================================================== ns_1@127.0.0.1:stats_collector:71: Stats for bucket "default": auth_cmds 2 auth_errors 0 bucket_conns 3 bytes_read 29059 bytes_written 34050 cas_badval 0 cas_hits 0 cas_misses 0 cmd_flush 0 cmd_get 0 cmd_set 2 conn_yields 0 connection_structures 13 curr_connections 13 curr_items 2 curr_items_tot 2 daemon_connections 10 decr_hits 0 decr_misses 0 delete_hits 0 delete_misses 0 ep_bg_fetched 0 ep_commit_num 0 ep_commit_time 0 ep_commit_time_total 0 ep_data_age 0 ep_data_age_highwat 0 ep_db_cleaner_status complete ep_db_strategy multiMTVBDB ep_dbinit 1 ep_dbname /var/opt/membase/1.6.5.4r/data/ns_1/default ep_dbshards 4 ep_expired 0 ep_flush_duration 0 ep_flush_duration_highwat 0 ep_flush_duration_total 0 ep_flush_preempts 0 ep_flusher_state running ep_flusher_todo 0 ep_io_num_read 0 ep_io_num_write 0 ep_io_read_bytes 0 ep_io_write_bytes 0 ep_item_begin_failed 0 ep_item_commit_failed 0 ep_item_flush_expired 0 ep_item_flush_failed 0 ep_kv_size 318 ep_max_data_size 1096810496 ep_max_txn_size 1000 ep_mem_high_wat 822607872 ep_mem_low_wat 658086297 ep_min_data_age 0 ep_num_active_non_resident 0 ep_num_eject_failures 0 ep_num_eject_replicas 0 ep_num_expiry_pager_runs 0 ep_num_non_resident 0 ep_num_not_my_vbuckets 0 ep_num_pager_runs 0 ep_num_value_ejects 0 ep_oom_errors 0 ep_overhead 25854584 ep_pending_ops 0 ep_pending_ops_max 0 ep_pending_ops_max_duration 0 ep_pending_ops_total 0 ep_queue_age_cap 900 ep_queue_size 2 ep_storage_age 0 ep_storage_age_highwat 0 ep_storage_type featured ep_store_max_concurrency 10 ep_store_max_readers 9 ep_store_max_readwrite 1 ep_tap_bg_fetch_requeued 0 ep_tap_bg_fetched 0 ep_tap_keepalive 300 ep_tmp_oom_errors 0 ep_too_old 0 ep_too_young 0 ep_total_cache_size 318 ep_total_del_items 0 ep_total_enqueued 2 ep_total_new_items 0 ep_total_persisted 0 ep_vbucket_del 0 ep_vbucket_del_fail 0 ep_version 1.6.5.3.1_13_g81a6cf1 ep_warmed_up 0 ep_warmup true ep_warmup_dups 0 ep_warmup_oom 0 ep_warmup_thread complete ep_warmup_time 15662 get_hits 0 get_misses 0 incr_hits 0 incr_misses 0 libevent 1.4.13-stable limit_maxbytes 67108864 mem_used 25854902 pid 16183 pointer_size 64 rejected_conns 0 rusage_system 0.260960 rusage_user 0.617906 threads 4 time 1305060810 total_connections 13 uptime 204 version 1.4.4_382_g9df3289 INFO REPORT <6040.62.0> 2011-05-10 13:53:42 =============================================================================== ns_1@127.0.0.1:ns_cluster:90: handling add_node("10.1.5.229", 8091, ..) INFO REPORT <6040.62.0> 2011-05-10 13:53:42 =============================================================================== ns_1@127.0.0.1:ns_cluster:247: Decided to change address to "10.1.5.227" INFO REPORT <6040.110.0> 2011-05-10 13:53:42 =============================================================================== ns_1@127.0.0.1:ns_mnesia:144: Info from Mnesia: {mnesia_checkpoint,{{1305,60822,586696},'ns_1@127.0.0.1'}} starting: <6040.382.0> INFO REPORT <6040.110.0> 2011-05-10 13:53:42 =============================================================================== ns_1@127.0.0.1:ns_mnesia:144: Info from Mnesia: {mnesia_checkpoint,{{1305,60822,586696},'ns_1@127.0.0.1'}} terminated: shutdown INFO REPORT <6040.62.0> 2011-05-10 13:53:42 =============================================================================== ns_1@127.0.0.1:ns_cluster:249: prepared mnesia. INFO REPORT <6040.62.0> 2011-05-10 13:53:42 =============================================================================== ns_server_sup plug pulled. Killing [ns_mnesia,ns_bad_bucket_worker, ns_bad_bucket_sup,ns_moxi_sup,ns_tick], keeping [ns_log, ns_log_events, ns_mail_sup, ns_node_disco_sup, ns_heart, ns_doctor, menelaus, ns_port_sup, ns_tick_event, ns_stats_event, ns_good_bucket_worker, ns_good_bucket_sup, ns_orchestrator] INFO REPORT <6040.7.0> 2011-05-10 13:53:42 =============================================================================== application mnesia exited stopped type temporary INFO REPORT <3.110.0> 2011-05-10 13:53:42 =============================================================================== ns_1@127.0.0.1:ns_mnesia:202: Shut Mnesia down: shutdown. Exiting. INFO REPORT <3.78.0> 2011-05-10 13:53:42 =============================================================================== ns_log: logging ns_node_disco:5:Node nonode@nohost saw that node 'ns_1@127.0.0.1' went down. INFO REPORT <3.55.0> 2011-05-10 13:53:42 =============================================================================== Adjusted IP to "10.1.5.227" PROGRESS REPORT <3.388.0> 2011-05-10 13:53:42 =============================================================================== supervisor {local,net_sup} started [{pid,<3.389.0>}, {name,erl_epmd}, {mfa,{erl_epmd,start_link,[]}}, {restart_type,permanent}, {shutdown,2000}, {child_type,worker}] PROGRESS REPORT <3.388.0> 2011-05-10 13:53:42 =============================================================================== supervisor {local,net_sup} started [{pid,<3.390.0>}, {name,auth}, {mfa,{auth,start_link,[]}}, {restart_type,permanent}, {shutdown,2000}, {child_type,worker}] INFO REPORT <0.78.0> 2011-05-10 13:53:42 =============================================================================== ns_log: logging ns_node_disco:4:Node 'ns_1@10.1.5.227' saw that node 'ns_1@10.1.5.227' came up. PROGRESS REPORT <0.388.0> 2011-05-10 13:53:42 =============================================================================== supervisor {local,net_sup} started [{pid,<0.391.0>}, {name,net_kernel}, {mfa,{net_kernel,start_link,[['ns_1@10.1.5.227',longnames]]}}, {restart_type,permanent}, {shutdown,2000}, {child_type,worker}] PROGRESS REPORT <0.11.0> 2011-05-10 13:53:42 =============================================================================== supervisor {local,kernel_sup} started [{pid,<0.388.0>}, {name,net_sup_dynamic}, {mfa,{erl_distribution,start_link,[['ns_1@10.1.5.227',longnames]]}}, {restart_type,permanent}, {shutdown,1000}, {child_type,supervisor}] INFO REPORT <0.55.0> 2011-05-10 13:53:42 =============================================================================== Re-setting cookie {bujipmggocppuqmx,'ns_1@10.1.5.227'} INFO REPORT <0.55.0> 2011-05-10 13:53:42 =============================================================================== saving ip config to "/etc/opt/membase/1.6.5.4r/ip" INFO REPORT <0.55.0> 2011-05-10 13:53:42 =============================================================================== save_address_config: ok INFO REPORT <0.62.0> 2011-05-10 13:53:42 =============================================================================== ns_1@10.1.5.227:ns_mnesia:69: Renaming node from 'ns_1@127.0.0.1' to 'ns_1@10.1.5.227'. INFO REPORT <0.62.0> 2011-05-10 13:53:42 =============================================================================== ns_1@10.1.5.227:ns_mnesia:73: Deleting old schema. INFO REPORT <0.62.0> 2011-05-10 13:53:42 =============================================================================== ns_1@10.1.5.227:ns_mnesia:75: Installing new backup as fallback. ERROR REPORT <0.6.0> 2011-05-10 13:53:42 =============================================================================== Truncated log event: {info_msg,<0.51.0>, {<0.65.0>,"renaming node conf ~p -> ~p:~n ~p ->~n ~p~n", [buckets,buckets, [{configs,[{"default", [{num_replicas,1}, {ram_quota,1096810496}, {auth_type,sasl}, {sasl_password,[]}, {type,membase}, {num_vbuckets,1024}, {db_shards,4}, {ht_size,3079}, {tap_keepalive,0}, {tap_noop_interval,20}, {max_txn_size,1000}, {ht_locks,5}, {servers,['ns_1@127.0.0.1']}, {map,[['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1'|...], [...]|...]}]}]}], [{configs,[{"default", [{num_replicas,1}, {ram_quota,1096810496}, {auth_type,sasl}, {sasl_password,[]}, {type,membase}, {num_vbuckets,1024}, {db_shards,4}, {ht_size,3079}, {tap_keepalive,0}, {tap_noop_interval,20}, {max_txn_size,1000}, {ht_locks,5}, {servers,['ns_1@10.1.5.227']}, {map,[['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227'|...], [...]|...]}]}]}]]}} INFO REPORT <0.65.0> 2011-05-10 13:53:42 =============================================================================== renaming node conf nodes_wanted -> nodes_wanted: ['ns_1@127.0.0.1'] -> ['ns_1@10.1.5.227'] INFO REPORT <0.65.0> 2011-05-10 13:53:42 =============================================================================== renaming node conf {node,'ns_1@127.0.0.1',isasl} -> {node,'ns_1@10.1.5.227', isasl}: [{path,"/var/opt/membase/1.6.5.4r/data/ns_1/isasl.pw"}] -> [{path,"/var/opt/membase/1.6.5.4r/data/ns_1/isasl.pw"}] INFO REPORT <0.65.0> 2011-05-10 13:53:42 =============================================================================== renaming node conf {node,'ns_1@127.0.0.1',membership} -> {node, 'ns_1@10.1.5.227', membership}: active -> active INFO REPORT <0.65.0> 2011-05-10 13:53:42 =============================================================================== renaming node conf {node,'ns_1@127.0.0.1',memcached} -> {node, 'ns_1@10.1.5.227', memcached}: [{port,11210}, {dbdir,"/var/opt/membase/1.6.5.4r/data/ns_1"}, {admin_user,"_admin"}, {admin_pass,"_admin"}, {bucket_engine,"./bin/bucket_engine/bucket_engine.so"}, {engines,[{membase,[{engine,"bin/ep_engine/ep.so"}, {initfile,"/etc/opt/membase/1.6.5.4r/init.sql"}]}, {memcached,[{engine,"bin/memcached/default_engine.so"}]}]}, {verbosity,[]}] -> [{port,11210}, {dbdir,"/var/opt/membase/1.6.5.4r/data/ns_1"}, {admin_user,"_admin"}, {admin_pass,"_admin"}, {bucket_engine,"./bin/bucket_engine/bucket_engine.so"}, {engines,[{membase,[{engine,"bin/ep_engine/ep.so"}, {initfile,"/etc/opt/membase/1.6.5.4r/init.sql"}]}, {memcached,[{engine,"bin/memcached/default_engine.so"}]}]}, {verbosity,[]}] INFO REPORT <0.65.0> 2011-05-10 13:53:42 =============================================================================== renaming node conf {node,'ns_1@127.0.0.1',ns_log} -> {node,'ns_1@10.1.5.227', ns_log}: [{filename,"/var/opt/membase/1.6.5.4r/data/ns_1/ns_log"}] -> [{filename,"/var/opt/membase/1.6.5.4r/data/ns_1/ns_log"}] INFO REPORT <0.62.0> 2011-05-10 13:53:42 =============================================================================== ns_1@10.1.5.227:ns_cluster:260: Renamed node. New name is 'ns_1@10.1.5.227'. INFO REPORT <0.64.0> 2011-05-10 13:53:42 =============================================================================== config change: {node,'ns_1@10.1.5.227',ns_log} -> [{filename,"/var/opt/membase/1.6.5.4r/data/ns_1/ns_log"}] PROGRESS REPORT <0.412.0> 2011-05-10 13:53:42 =============================================================================== supervisor {local,mnesia_sup} started [{pid,<0.413.0>}, {name,mnesia_event}, {mfa,{mnesia_sup,start_event,[]}}, {restart_type,permanent}, {shutdown,30000}, {child_type,worker}] INFO REPORT <0.64.0> 2011-05-10 13:53:42 =============================================================================== config change: {node,'ns_1@10.1.5.227',memcached} -> [{port,11210}, {dbdir,"/var/opt/membase/1.6.5.4r/data/ns_1"}, {admin_user,"_admin"}, {admin_pass,"_admin"}, {bucket_engine,"./bin/bucket_engine/bucket_engine.so"}, {engines,[{membase,[{engine,"bin/ep_engine/ep.so"}, {initfile,"/etc/opt/membase/1.6.5.4r/init.sql"}]}, {memcached,[{engine,"bin/memcached/default_engine.so"}]}]}, {verbosity,[]}] INFO REPORT <0.64.0> 2011-05-10 13:53:42 =============================================================================== config change: {node,'ns_1@10.1.5.227',membership} -> active INFO REPORT <0.64.0> 2011-05-10 13:53:42 =============================================================================== config change: {node,'ns_1@10.1.5.227',isasl} -> [{path,"/var/opt/membase/1.6.5.4r/data/ns_1/isasl.pw"}] INFO REPORT <0.64.0> 2011-05-10 13:53:42 =============================================================================== ns_node_disco_conf_events config on nodes_wanted INFO REPORT <0.64.0> 2011-05-10 13:53:42 =============================================================================== config change: nodes_wanted -> ['ns_1@10.1.5.227'] INFO REPORT <0.416.0> 2011-05-10 13:53:42 =============================================================================== ns_node_disco cookie_sync INFO REPORT <0.416.0> 2011-05-10 13:53:42 =============================================================================== ns_node_disco: nodes_wanted updated: ['ns_1@10.1.5.227'], with cookie: bujipmggocppuqmx INFO REPORT <0.416.0> 2011-05-10 13:53:42 =============================================================================== ns_node_disco: nodes_wanted pong: ['ns_1@10.1.5.227'], with cookie: bujipmggocppuqmx INFO REPORT <0.64.0> 2011-05-10 13:53:42 =============================================================================== config change: buckets -> [{configs,[{"default", [{num_replicas,1}, {ram_quota,1096810496}, {auth_type,sasl}, {sasl_password,[]}, {type,membase}, {num_vbuckets,1024}, {db_shards,4}, {ht_size,3079}, {tap_keepalive,0}, {tap_noop_interval,20}, {max_txn_size,1000}, {ht_locks,5}, {servers,['ns_1@10.1.5.227']}, {map,[['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227'|...], [...]|...]}]}]}] INFO REPORT <0.64.0> 2011-05-10 13:53:42 =============================================================================== ns_node_disco_conf_events config all INFO REPORT <0.85.0> 2011-05-10 13:53:42 =============================================================================== Pushing config INFO REPORT <0.85.0> 2011-05-10 13:53:42 =============================================================================== Pushing config done PROGRESS REPORT <0.414.0> 2011-05-10 13:53:42 =============================================================================== supervisor {local,mnesia_kernel_sup} started [{pid,<0.415.0>}, {name,mnesia_monitor}, {mfa,{mnesia_monitor,start,[]}}, {restart_type,permanent}, {shutdown,3000}, {child_type,worker}] PROGRESS REPORT <0.414.0> 2011-05-10 13:53:42 =============================================================================== supervisor {local,mnesia_kernel_sup} started [{pid,<0.418.0>}, {name,mnesia_subscr}, {mfa,{mnesia_subscr,start,[]}}, {restart_type,permanent}, {shutdown,3000}, {child_type,worker}] PROGRESS REPORT <0.414.0> 2011-05-10 13:53:42 =============================================================================== supervisor {local,mnesia_kernel_sup} started [{pid,<0.419.0>}, {name,mnesia_locker}, {mfa,{mnesia_locker,start,[]}}, {restart_type,permanent}, {shutdown,3000}, {child_type,worker}] PROGRESS REPORT <0.414.0> 2011-05-10 13:53:42 =============================================================================== supervisor {local,mnesia_kernel_sup} started [{pid,<0.420.0>}, {name,mnesia_recover}, {mfa,{mnesia_recover,start,[]}}, {restart_type,permanent}, {shutdown,180000}, {child_type,worker}] PROGRESS REPORT <0.414.0> 2011-05-10 13:53:43 =============================================================================== supervisor {local,mnesia_kernel_sup} started [{pid,<0.421.0>}, {name,mnesia_tm}, {mfa,{mnesia_tm,start,[]}}, {restart_type,permanent}, {shutdown,30000}, {child_type,worker}] PROGRESS REPORT <0.414.0> 2011-05-10 13:53:43 =============================================================================== supervisor {local,mnesia_kernel_sup} started [{pid,<0.458.0>}, {name,mnesia_checkpoint_sup}, {mfa,{mnesia_checkpoint_sup,start,[]}}, {restart_type,permanent}, {shutdown,infinity}, {child_type,supervisor}] PROGRESS REPORT <0.414.0> 2011-05-10 13:53:43 =============================================================================== supervisor {local,mnesia_kernel_sup} started [{pid,<0.459.0>}, {name,mnesia_snmp_sup}, {mfa,{mnesia_snmp_sup,start,[]}}, {restart_type,permanent}, {shutdown,infinity}, {child_type,supervisor}] PROGRESS REPORT <0.414.0> 2011-05-10 13:53:43 =============================================================================== supervisor {local,mnesia_kernel_sup} started [{pid,<0.460.0>}, {name,mnesia_controller}, {mfa,{mnesia_controller,start,[]}}, {restart_type,permanent}, {shutdown,3000}, {child_type,worker}] PROGRESS REPORT <0.414.0> 2011-05-10 13:53:43 =============================================================================== supervisor {local,mnesia_kernel_sup} started [{pid,<0.461.0>}, {name,mnesia_late_loader}, {mfa,{mnesia_late_loader,start,[]}}, {restart_type,permanent}, {shutdown,3000}, {child_type,worker}] PROGRESS REPORT <0.412.0> 2011-05-10 13:53:43 =============================================================================== supervisor {local,mnesia_sup} started [{pid,<0.414.0>}, {name,mnesia_kernel_sup}, {mfa,{mnesia_kernel_sup,start,[]}}, {restart_type,permanent}, {shutdown,infinity}, {child_type,supervisor}] PROGRESS REPORT <0.7.0> 2011-05-10 13:53:43 =============================================================================== application mnesia started_at 'ns_1@10.1.5.227' INFO REPORT <0.408.0> 2011-05-10 13:53:43 =============================================================================== ns_1@10.1.5.227:ns_mnesia:258: Using existing disk schema on ['ns_1@10.1.5.227']. INFO REPORT <0.408.0> 2011-05-10 13:53:43 =============================================================================== ns_1@10.1.5.227:ns_mnesia:196: Current config: [{access_module,mnesia}, {auto_repair,true}, {backup_module,mnesia_backup}, {checkpoints,[]}, {db_nodes,['ns_1@10.1.5.227']}, {debug,verbose}, {directory, "/var/opt/membase/1.6.5.4r/mnesia"}, {dump_log_load_regulation, false}, {dump_log_time_threshold, 180000}, {dump_log_update_in_place, true}, {dump_log_write_threshold, 1000}, {embedded_mnemosyne,false}, {event_module,mnesia_event}, {extra_db_nodes,[]}, {fallback_activated,false}, {held_locks,[]}, {ignore_fallback_at_startup, false}, {fallback_error_function, {mnesia,lkill}}, {is_running,yes}, {local_tables, ['stats_archiver-default-day', 'stats_archiver-default-month', 'stats_archiver-default-week', 'stats_archiver-default-year', schema, 'stats_archiver-default-minute', 'stats_archiver-default-hour']}, {lock_queue,[]}, {log_version,"4.3"}, {master_node_tables,[]}, {max_wait_for_decision,10000}, {protocol_version,{7,6}}, {running_db_nodes, ['ns_1@10.1.5.227']}, {schema_location,opt_disc}, {schema_version,{3,0}}, {subscribers, [<0.413.0>,<0.408.0>]}, {tables, ['stats_archiver-default-day', 'stats_archiver-default-month', 'stats_archiver-default-week', 'stats_archiver-default-year', schema, 'stats_archiver-default-minute', 'stats_archiver-default-hour']}, {transaction_commits,2}, {transaction_failures,0}, {transaction_log_writes,0}, {transaction_restarts,0}, {transactions,[]}, {use_dir,true}, {core_dir,false}, {no_table_loaders,2}, {dc_dump_limit,4}, {version,"4.4.12"}] PROGRESS REPORT <0.70.0> 2011-05-10 13:53:43 =============================================================================== supervisor {local,ns_server_sup} started [{pid,<0.408.0>}, {name,ns_mnesia}, {mfa,{ns_mnesia,start_link,[]}}, {restart_type,permanent}, {shutdown,10000}, {child_type,worker}] INFO REPORT <0.62.0> 2011-05-10 13:53:43 =============================================================================== Restarting ns_mnesia: {ok,<0.408.0>} PROGRESS REPORT <0.70.0> 2011-05-10 13:53:43 =============================================================================== supervisor {local,ns_server_sup} started [{pid,<0.467.0>}, {name,ns_bad_bucket_worker}, {mfa,{work_queue,start_link,[ns_bad_bucket_worker]}}, {restart_type,permanent}, {shutdown,10}, {child_type,worker}] INFO REPORT <0.62.0> 2011-05-10 13:53:43 =============================================================================== Restarting ns_bad_bucket_worker: {ok,<0.467.0>} PROGRESS REPORT <0.468.0> 2011-05-10 13:53:43 =============================================================================== supervisor {local,ns_bad_bucket_sup} started [{pid,<0.469.0>}, {name,{stats_collector,"default"}}, {mfa,{stats_collector,start_link,["default"]}}, {restart_type,permanent}, {shutdown,10}, {child_type,worker}] PROGRESS REPORT <0.468.0> 2011-05-10 13:53:43 =============================================================================== supervisor {local,ns_bad_bucket_sup} started [{pid,<0.470.0>}, {name,{stats_archiver,"default"}}, {mfa,{stats_archiver,start_link,["default"]}}, {restart_type,permanent}, {shutdown,10}, {child_type,worker}] PROGRESS REPORT <0.468.0> 2011-05-10 13:53:43 =============================================================================== supervisor {local,ns_bad_bucket_sup} started [{pid,<0.471.0>}, {name,{stats_reader,"default"}}, {mfa,{stats_reader,start_link,["default"]}}, {restart_type,permanent}, {shutdown,10}, {child_type,worker}] PROGRESS REPORT <0.70.0> 2011-05-10 13:53:43 =============================================================================== supervisor {local,ns_server_sup} started [{pid,<0.468.0>}, {name,ns_bad_bucket_sup}, {mfa,{ns_bucket_sup,start_link, [ns_bad_bucket_sup, #Fun, ns_bad_bucket_worker]}}, {restart_type,permanent}, {shutdown,infinity}, {child_type,supervisor}] INFO REPORT <0.62.0> 2011-05-10 13:53:43 =============================================================================== Restarting ns_bad_bucket_sup: {ok,<0.468.0>} PROGRESS REPORT <0.70.0> 2011-05-10 13:53:43 =============================================================================== supervisor {local,ns_server_sup} started [{pid,<0.472.0>}, {name,ns_moxi_sup}, {mfa,{ns_moxi_sup,start_link,[]}}, {restart_type,permanent}, {shutdown,infinity}, {child_type,supervisor}] INFO REPORT <0.62.0> 2011-05-10 13:53:43 =============================================================================== Restarting ns_moxi_sup: {ok,<0.472.0>} INFO REPORT <0.70.0> 2011-05-10 13:53:43 =============================================================================== ns_1@10.1.5.227:misc:725: start_singleton(gen_server, ns_tick, [], []): started as <0.473.0> on 'ns_1@10.1.5.227' PROGRESS REPORT <0.70.0> 2011-05-10 13:53:43 =============================================================================== supervisor {local,ns_server_sup} started [{pid,<0.473.0>}, {name,ns_tick}, {mfa,{ns_tick,start_link,[]}}, {restart_type,permanent}, {shutdown,10}, {child_type,worker}] INFO REPORT <0.62.0> 2011-05-10 13:53:43 =============================================================================== Restarting ns_tick: {ok,<0.473.0>} PROGRESS REPORT <0.70.0> 2011-05-10 13:53:43 =============================================================================== supervisor {local,ns_server_sup} started [{pid,<0.474.0>}, {name,ns_doctor}, {mfa,{ns_doctor,start_link,[]}}, {restart_type,permanent}, {shutdown,10}, {child_type,worker}] INFO REPORT <0.62.0> 2011-05-10 13:53:43 =============================================================================== ns_1@10.1.5.227:ns_server_sup:140: Restarted grey child ns_doctor: {ok, <0.474.0>} INFO REPORT <0.62.0> 2011-05-10 13:53:43 =============================================================================== ns_1@10.1.5.227:ns_cluster:264: Started ns_server_sup childs back. INFO REPORT <0.474.0> 2011-05-10 13:53:43 =============================================================================== ns_1@10.1.5.227:ns_doctor:78: Got initial status [{'ns_1@10.1.5.227', [{last_heard, {1305,60823,207165}}, {active_buckets, ["default"]}, {memory, [{total,28765624}, {processes,10145176}, {processes_used, 10133528}, {system,18620448}, {atom,782161}, {atom_used,760029}, {binary,106808}, {code,7891711}, {ets,584880}]}, {cluster_compatibility_version, 1}, {version, [{kernel,"2.13.4"}, {mnesia,"4.4.12"}, {stdlib,"1.16.4"}, {os_mon,"2.2.4"}, {ns_server,"1.6.5.4r"}, {menelaus,"1.6.5.4r"}, {sasl,"2.1.8"}]}, {system_arch, "x86_64-unknown-linux-gnu"}, {wall_clock,215}, {memory_data, {2058731520,569475072, {<0.70.0>,972712}}}, {disk_data, [{"/",11204616,49}, {"/boot",101086,14}, {"/dev/shm",1005240,0}]}, {meminfo, <<"MemTotal: 2010480 kB\nMemFree: 1376608 kB\nBuffers: 60152 kB\nCached: 339560 kB\nSwapCached: 20084 kB\nActive: 323928 kB\nInactive: 198764 kB\nHighTotal: 0 kB\nHighFree: 0 kB\nLowTotal: 2010480 kB\nLowFree: 1376608 kB\nSwapTotal: 4030456 kB\nSwapFree: 3952288 kB\nDirty: 30504 kB\nWriteback: 164 kB\nAnonPages: 120760 kB\nMapped: 15548 kB\nSlab: 85956 kB\nPageTables: 6164 kB\nNFS_Unstable: 0 kB\nBounce: 0 kB\nCommitLimit: 5035696 kB\nCommitted_AS: 501728 kB\nVmallocTotal: 34359738367 kB\nVmallocUsed: 265484 kB\nVmallocChunk: 34359471011 kB\nHugePages_Total: 0\nHugePages_Free: 0\nHugePages_Rsvd: 0\nHugepagesize: 2048 kB\n">>}, {replication, [{"default",1.0}]}, {system_memory_data, [{system_total_memory, 2058731520}, {free_swap,4047142912}, {total_swap,4127186944}, {cached_memory, 347709440}, {buffered_memory, 61595648}, {free_memory,1409646592}, {total_memory, 2058731520}]}, {statistics, [{wall_clock,{210583,5}}, {context_switches, {25878,0}}, {garbage_collection, {4299,11531733,0}}, {io, {{input,4191992}, {output,707856}}}, {reductions, {3221605,599999}}, {run_queue,0}, {runtime,{880,150}}]}]}] ERROR REPORT <0.471.0> 2011-05-10 13:53:43 =============================================================================== ** Generic server 'stats_reader-default' terminating ** Last message in was {latest,minute,1} ** When Server state == {state,"default"} ** Reason for termination == ** {aborted,{no_exists,['stats_archiver-default-minute']}} CRASH REPORT <0.471.0> 2011-05-10 13:53:43 =============================================================================== Crashing process initial_call {stats_reader,init,['Argument__1']} pid <0.471.0> registered_name 'stats_reader-default' error_info {exit,{aborted,{no_exists,['stats_archiver-default-minute']}}, [{gen_server,terminate,6},{proc_lib,init_p_do_apply,3}]} ancestors [ns_bad_bucket_sup,ns_server_sup,ns_server_cluster_sup,<0.52.0>] messages [] links [<0.468.0>] dictionary [] trap_exit false status running heap_size 377 stack_size 24 reductions 300 SUPERVISOR REPORT <0.468.0> 2011-05-10 13:53:43 =============================================================================== Reporting supervisor {local,ns_bad_bucket_sup} Child process errorContext child_terminated reason {aborted,{no_exists,['stats_archiver-default-minute']}} pid <0.471.0> name {stats_reader,"default"} start_function {stats_reader,start_link,["default"]} restart_type permanent shutdown 10 child_type worker ERROR REPORT <0.62.0> 2011-05-10 13:53:43 =============================================================================== ns_1@10.1.5.227:stats_reader:176: Some nodes didn't respond: ['ns_1@10.1.5.227'] PROGRESS REPORT <0.468.0> 2011-05-10 13:53:43 =============================================================================== supervisor {local,ns_bad_bucket_sup} started [{pid,<0.477.0>}, {name,{stats_reader,"default"}}, {mfa,{stats_reader,start_link,["default"]}}, {restart_type,permanent}, {shutdown,10}, {child_type,worker}] INFO REPORT <0.62.0> 2011-05-10 13:53:43 =============================================================================== ns_1@10.1.5.227:ns_cluster:300: Posting node info to engage_cluster on {"10.1.5.229", 8091}: {struct, [{availableStorage, {struct, [{hdd, [{struct, [{path,<<"/">>}, {sizeKBytes,11204616}, {usagePercent,49}]}, {struct, [{path,<<"/boot">>}, {sizeKBytes,101086}, {usagePercent,14}]}, {struct, [{path,<<"/dev/shm">>}, {sizeKBytes,1005240}, {usagePercent,0}]}]}]}}, {memoryQuota,1570}, {storageTotals, {struct, [{ram, {struct, [{usedByData,0}, {total,2058731520}, {quotaTotal,1646264320}, {used,569475072}]}}, {hdd, {struct, [{usedByData,1240064}, {total,11473526784}, {quotaTotal,11473526784}, {used,5622028124}, {free,5851498660}]}}]}}, {storage, {struct, [{ssd,[]}, {hdd, [{struct, [{path,<<"/var/opt/membase/1.6.5.4r/data/ns_1">>}, {quotaMb,none}, {state,ok}]}]}]}}, {uptime,<<"215">>}, {memoryTotal,2058731520}, {memoryFree,1489256448}, {mcdMemoryReserved,1570}, {mcdMemoryAllocated,1570}, {otpNode,<<"ns_1@10.1.5.227">>}, {otpCookie,<<"bujipmggocppuqmx">>}, {clusterMembership,<<"active">>}, {status,<<"healthy">>}, {hostname,<<"10.1.5.227:8091">>}, {clusterCompatibility,1}, {version,<<"1.6.5.4r">>}, {os,<<"x86_64-unknown-linux-gnu">>}, {ports,{struct,[{proxy,11211},{direct,11210}]}}]} PROGRESS REPORT <0.515.0> 2011-05-10 13:53:43 =============================================================================== supervisor {local,inets_sup} started [{pid,<0.516.0>}, {name,ftp_sup}, {mfa,{ftp_sup,start_link,[]}}, {restart_type,permanent}, {shutdown,infinity}, {child_type,supervisor}] PROGRESS REPORT <0.518.0> 2011-05-10 13:53:43 =============================================================================== supervisor {local,httpc_profile_sup} started [{pid,<0.519.0>}, {name,httpc_manager}, {mfa,{httpc_manager,start_link,[{default,only_session_cookies}]}}, {restart_type,permanent}, {shutdown,4000}, {child_type,worker}] PROGRESS REPORT <0.517.0> 2011-05-10 13:53:43 =============================================================================== supervisor {local,httpc_sup} started [{pid,<0.518.0>}, {name,httpc_profile_sup}, {mfa,{httpc_profile_sup,start_link, [[{httpc,{default,only_session_cookies}}]]}}, {restart_type,permanent}, {shutdown,infinity}, {child_type,supervisor}] PROGRESS REPORT <0.517.0> 2011-05-10 13:53:43 =============================================================================== supervisor {local,httpc_sup} started [{pid,<0.520.0>}, {name,httpc_handler_sup}, {mfa,{httpc_handler_sup,start_link,[]}}, {restart_type,permanent}, {shutdown,infinity}, {child_type,supervisor}] PROGRESS REPORT <0.515.0> 2011-05-10 13:53:43 =============================================================================== supervisor {local,inets_sup} started [{pid,<0.517.0>}, {name,httpc_sup}, {mfa,{httpc_sup,start_link, [[{httpc,{default,only_session_cookies}}]]}}, {restart_type,permanent}, {shutdown,infinity}, {child_type,supervisor}] PROGRESS REPORT <0.515.0> 2011-05-10 13:53:43 =============================================================================== supervisor {local,inets_sup} started [{pid,<0.521.0>}, {name,httpd_sup}, {mfa,{httpd_sup,start_link,[[]]}}, {restart_type,permanent}, {shutdown,infinity}, {child_type,supervisor}] PROGRESS REPORT <0.515.0> 2011-05-10 13:53:43 =============================================================================== supervisor {local,inets_sup} started [{pid,<0.522.0>}, {name,tftp_sup}, {mfa,{tftp_sup,start_link,[[]]}}, {restart_type,permanent}, {shutdown,infinity}, {child_type,supervisor}] PROGRESS REPORT <0.7.0> 2011-05-10 13:53:43 =============================================================================== application inets started_at 'ns_1@10.1.5.227' INFO REPORT <0.62.0> 2011-05-10 13:53:43 =============================================================================== ns_1@10.1.5.227:ns_cluster:306: Reply from engage_cluster on {"10.1.5.229", 8091}: {ok,{struct,[{<<"availableStorage">>, {struct,[{<<"hdd">>, [{struct,[{<<"path">>,<<"/">>}, {<<"sizeKBytes">>,11204616}, {<<"usagePercent">>,44}]}, {struct,[{<<"path">>,<<"/boot">>}, {<<"sizeKBytes">>,101086}, {<<"usagePercent">>,14}]}, {struct,[{<<"path">>,<<"/dev/shm">>}, {<<"sizeKBytes">>,1005240}, {<<"usagePercent">>,0}]}]}]}}, {<<"memoryQuota">>,1570}, {<<"storageTotals">>, {struct,[{<<"ram">>, {struct,[{<<"usedByData">>,0}, {<<"total">>,2058731520}, {<<"quotaTotal">>,1646264320}, {<<"used">>,653295616}]}}, {<<"hdd">>, {struct,[{<<"usedByData">>,0}, {<<"total">>,11473526784.0}, {<<"quotaTotal">>,11473526784.0}, {<<"used">>,5048351784.0}, {<<"free">>,6.425175e9}]}}]}}, {<<"storage">>, {struct,[{<<"ssd">>,[]}, {<<"hdd">>, [{struct,[{<<"path">>, <<"/var/opt/membase/1.6.5.4r/data/ns_1">>}, {<<"quotaMb">>,<<"none">>}, {<<"state">>,<<"ok">>}]}]}]}}, {<<"uptime">>,<<"179">>}, {<<"memoryTotal">>,2058731520}, {<<"memoryFree">>,1405435904}, {<<"mcdMemoryReserved">>,1570}, {<<"mcdMemoryAllocated">>,1570}, {<<"otpNode">>,<<"ns_1@10.1.5.229">>}, {<<"otpCookie">>,<<"awyobxdafzhifeho">>}, {<<"clusterMembership">>,<<"active">>}, {<<"status">>,<<"healthy">>}, {<<"hostname">>,<<"10.1.5.229:8091">>}, {<<"clusterCompatibility">>,1}, {<<"version">>,<<"1.6.5.4r">>}, {<<"os">>,<<"x86_64-unknown-linux-gnu">>}, {<<"ports">>, {struct,[{<<"proxy">>,11211},{<<"direct">>,11210}]}}]}} INFO REPORT <0.62.0> 2011-05-10 13:53:43 =============================================================================== ns_1@10.1.5.227:ns_cluster:371: port_please("ns_1", "10.1.5.229") = 21100 INFO REPORT <0.62.0> 2011-05-10 13:53:43 =============================================================================== ns_1@10.1.5.227:ns_cluster:461: Started node add transaction by adding node 'ns_1@10.1.5.229' to nodes_wanted INFO REPORT <0.64.0> 2011-05-10 13:53:43 =============================================================================== ns_node_disco_conf_events config on nodes_wanted INFO REPORT <0.64.0> 2011-05-10 13:53:43 =============================================================================== config change: nodes_wanted -> ['ns_1@10.1.5.227','ns_1@10.1.5.229'] INFO REPORT <0.528.0> 2011-05-10 13:53:43 =============================================================================== ns_node_disco cookie_sync INFO REPORT <0.528.0> 2011-05-10 13:53:43 =============================================================================== ns_node_disco: nodes_wanted updated: ['ns_1@10.1.5.227','ns_1@10.1.5.229'], with cookie: bujipmggocppuqmx INFO REPORT <0.64.0> 2011-05-10 13:53:43 =============================================================================== ns_node_disco_conf_events config all INFO REPORT <0.85.0> 2011-05-10 13:53:43 =============================================================================== Pushing config INFO REPORT <0.85.0> 2011-05-10 13:53:43 =============================================================================== Pushing config done INFO REPORT <0.64.0> 2011-05-10 13:53:43 =============================================================================== config change: {node,'ns_1@10.1.5.229',membership} -> inactiveAdded INFO REPORT <0.64.0> 2011-05-10 13:53:43 =============================================================================== ns_node_disco_conf_events config all INFO REPORT <0.85.0> 2011-05-10 13:53:43 =============================================================================== Pushing config INFO REPORT <0.85.0> 2011-05-10 13:53:43 =============================================================================== Pushing config done INFO REPORT <0.62.0> 2011-05-10 13:53:43 =============================================================================== ns_1@10.1.5.227:ns_cluster:431: Posting the following to complete_join on "10.1.5.229:8091": {struct, [{<<"targetNode">>,'ns_1@10.1.5.229'}, {availableStorage, {struct, [{hdd, [{struct, [{path,<<"/">>}, {sizeKBytes,11204616}, {usagePercent,49}]}, {struct, [{path,<<"/boot">>}, {sizeKBytes,101086}, {usagePercent,14}]}, {struct, [{path,<<"/dev/shm">>}, {sizeKBytes,1005240}, {usagePercent,0}]}]}]}}, {memoryQuota,1570}, {storageTotals, {struct, [{ram, {struct, [{usedByData,26082544}, {total,2058731520}, {quotaTotal,1646264320}, {used,569475072}]}}, {hdd, {struct, [{usedByData,1240064}, {total,11473526784}, {quotaTotal,11473526784}, {used,5622028124}, {free,5851498660}]}}]}}, {storage, {struct, [{ssd,[]}, {hdd, [{struct, [{path,<<"/var/opt/membase/1.6.5.4r/data/ns_1">>}, {quotaMb,none}, {state,ok}]}]}]}}, {uptime,<<"215">>}, {memoryTotal,2058731520}, {memoryFree,1489256448}, {mcdMemoryReserved,1570}, {mcdMemoryAllocated,1570}, {otpNode,<<"ns_1@10.1.5.227">>}, {otpCookie,<<"bujipmggocppuqmx">>}, {clusterMembership,<<"active">>}, {status,<<"healthy">>}, {hostname,<<"10.1.5.227:8091">>}, {clusterCompatibility,1}, {version,<<"1.6.5.4r">>}, {os,<<"x86_64-unknown-linux-gnu">>}, {ports,{struct,[{proxy,11211},{direct,11210}]}}]} INFO REPORT <0.528.0> 2011-05-10 13:53:43 =============================================================================== ns_node_disco: nodes_wanted pong: ['ns_1@10.1.5.227'], with cookie: bujipmggocppuqmx INFO REPORT <0.78.0> 2011-05-10 13:53:44 =============================================================================== ns_log: logging ns_node_disco:4:Node 'ns_1@10.1.5.227' saw that node 'ns_1@10.1.5.229' came up. INFO REPORT <0.469.0> 2011-05-10 13:53:44 =============================================================================== ns_1@10.1.5.227:stats_collector:71: Stats for bucket "default": auth_cmds 16 auth_errors 0 bucket_conns 17 bytes_read 1215254 bytes_written 492994 cas_badval 0 cas_hits 0 cas_misses 0 cmd_flush 1 cmd_get 0 cmd_set 11984 conn_yields 0 connection_structures 27 curr_connections 27 curr_items 4020 curr_items_tot 4020 daemon_connections 10 decr_hits 0 decr_misses 0 delete_hits 0 delete_misses 0 ep_bg_fetched 0 ep_commit_num 38 ep_commit_time 0 ep_commit_time_total 3 ep_data_age 2 ep_data_age_highwat 5 ep_db_cleaner_status complete ep_db_strategy multiMTVBDB ep_dbinit 1 ep_dbname /var/opt/membase/1.6.5.4r/data/ns_1/default ep_dbshards 4 ep_expired 0 ep_flush_duration 0 ep_flush_duration_highwat 1 ep_flush_duration_total 3 ep_flush_preempts 0 ep_flusher_state running ep_flusher_todo 2585 ep_io_num_read 0 ep_io_num_write 7701 ep_io_read_bytes 0 ep_io_write_bytes 554472 ep_item_begin_failed 0 ep_item_commit_failed 0 ep_item_flush_expired 0 ep_item_flush_failed 0 ep_kv_size 590700 ep_max_data_size 1096810496 ep_max_txn_size 1000 ep_mem_high_wat 822607872 ep_mem_low_wat 658086297 ep_min_data_age 0 ep_num_active_non_resident 0 ep_num_eject_failures 0 ep_num_eject_replicas 0 ep_num_expiry_pager_runs 0 ep_num_non_resident 0 ep_num_not_my_vbuckets 0 ep_num_pager_runs 0 ep_num_value_ejects 0 ep_oom_errors 0 ep_overhead 26151664 ep_pending_ops 0 ep_pending_ops_max 0 ep_pending_ops_max_duration 0 ep_pending_ops_total 0 ep_queue_age_cap 900 ep_queue_size 1439 ep_storage_age 1 ep_storage_age_highwat 2 ep_storage_type featured ep_store_max_concurrency 10 ep_store_max_readers 9 ep_store_max_readwrite 1 ep_tap_bg_fetch_requeued 0 ep_tap_bg_fetched 0 ep_tap_keepalive 300 ep_tmp_oom_errors 0 ep_too_old 0 ep_too_young 0 ep_total_cache_size 590700 ep_total_del_items 0 ep_total_enqueued 12012 ep_total_new_items 7701 ep_total_persisted 7701 ep_vbucket_del 0 ep_vbucket_del_fail 0 ep_version 1.6.5.3.1_13_g81a6cf1 ep_warmed_up 0 ep_warmup true ep_warmup_dups 0 ep_warmup_oom 0 ep_warmup_thread complete ep_warmup_time 15662 get_hits 0 get_misses 0 incr_hits 0 incr_misses 0 libevent 1.4.13-stable limit_maxbytes 67108864 mem_used 26742364 pid 16183 pointer_size 64 rejected_conns 0 rusage_system 0.747886 rusage_user 1.414784 threads 4 time 1305060823 total_connections 27 uptime 217 version 1.4.4_382_g9df3289 INFO REPORT <0.64.0> 2011-05-10 13:53:44 =============================================================================== ns_node_disco_conf_events config on otp INFO REPORT <0.64.0> 2011-05-10 13:53:44 =============================================================================== config change: otp -> [{cookie,bujipmggocppuqmx}] INFO REPORT <0.558.0> 2011-05-10 13:53:44 =============================================================================== ns_node_disco cookie_sync INFO REPORT <0.558.0> 2011-05-10 13:53:44 =============================================================================== ns_node_disco: nodes_wanted updated: ['ns_1@10.1.5.227','ns_1@10.1.5.229'], with cookie: bujipmggocppuqmx INFO REPORT <0.64.0> 2011-05-10 13:53:44 =============================================================================== config change: {node,'ns_1@127.0.0.1',ns_log} -> [{filename,"/var/opt/membase/1.6.5.4r/data/ns_1/ns_log"}] INFO REPORT <0.64.0> 2011-05-10 13:53:44 =============================================================================== config change: {node,'ns_1@127.0.0.1',membership} -> active INFO REPORT <0.558.0> 2011-05-10 13:53:44 =============================================================================== ns_node_disco: nodes_wanted pong: ['ns_1@10.1.5.227','ns_1@10.1.5.229'], with cookie: bujipmggocppuqmx INFO REPORT <0.64.0> 2011-05-10 13:53:44 =============================================================================== config change: {node,'ns_1@127.0.0.1',isasl} -> [{path,"/var/opt/membase/1.6.5.4r/data/ns_1/isasl.pw"}] INFO REPORT <0.64.0> 2011-05-10 13:53:44 =============================================================================== config change: {node,'ns_1@10.1.5.229',memcached} -> [{port,11210}, {dbdir,"/var/opt/membase/1.6.5.4r/data/ns_1"}, {admin_user,"_admin"}, {admin_pass,"_admin"}, {bucket_engine,"./bin/bucket_engine/bucket_engine.so"}, {engines,[{membase,[{engine,"bin/ep_engine/ep.so"}, {initfile,"/etc/opt/membase/1.6.5.4r/init.sql"}]}, {memcached,[{engine,"bin/memcached/default_engine.so"}]}]}, {verbosity,[]}] INFO REPORT <0.64.0> 2011-05-10 13:53:44 =============================================================================== config change: {node,'ns_1@10.1.5.229',ns_log} -> [{filename,"/var/opt/membase/1.6.5.4r/data/ns_1/ns_log"}] INFO REPORT <0.64.0> 2011-05-10 13:53:44 =============================================================================== config change: {node,'ns_1@127.0.0.1',memcached} -> [{port,11210}, {dbdir,"/var/opt/membase/1.6.5.4r/data/ns_1"}, {admin_user,"_admin"}, {admin_pass,"_admin"}, {bucket_engine,"./bin/bucket_engine/bucket_engine.so"}, {engines,[{membase,[{engine,"bin/ep_engine/ep.so"}, {initfile,"/etc/opt/membase/1.6.5.4r/init.sql"}]}, {memcached,[{engine,"bin/memcached/default_engine.so"}]}]}, {verbosity,[]}] INFO REPORT <0.64.0> 2011-05-10 13:53:44 =============================================================================== config change: {node,'ns_1@10.1.5.229',isasl} -> [{path,"/var/opt/membase/1.6.5.4r/data/ns_1/isasl.pw"}] INFO REPORT <0.64.0> 2011-05-10 13:53:44 =============================================================================== ns_node_disco_conf_events config all INFO REPORT <0.85.0> 2011-05-10 13:53:44 =============================================================================== Pushing config INFO REPORT <0.85.0> 2011-05-10 13:53:44 =============================================================================== Pushing config done INFO REPORT <0.62.0> 2011-05-10 13:53:44 =============================================================================== ns_1@10.1.5.227:ns_cluster:437: Reply from complete_join on "10.1.5.229:8091": {ok,[]} INFO REPORT <0.62.0> 2011-05-10 13:53:44 =============================================================================== ns_1@10.1.5.227:ns_cluster:92: add_node("10.1.5.229", 8091, ..) -> {ok, 'ns_1@10.1.5.229'} INFO REPORT <0.109.0> 2011-05-10 13:53:45 =============================================================================== ns_log: logging ns_orchestrator:4:Starting rebalance, KeepNodes = ['ns_1@10.1.5.227','ns_1@10.1.5.229'], EjectNodes = [] INFO REPORT <0.64.0> 2011-05-10 13:53:45 =============================================================================== config change: {node,'ns_1@10.1.5.229',membership} -> active INFO REPORT <0.64.0> 2011-05-10 13:53:45 =============================================================================== config change: {node,'ns_1@10.1.5.227',membership} -> active INFO REPORT <0.64.0> 2011-05-10 13:53:45 =============================================================================== ns_node_disco_conf_events config all INFO REPORT <0.85.0> 2011-05-10 13:53:45 =============================================================================== Pushing config INFO REPORT <0.85.0> 2011-05-10 13:53:45 =============================================================================== Pushing config done INFO REPORT <0.64.0> 2011-05-10 13:53:45 =============================================================================== config change: rebalance_status -> running INFO REPORT <0.64.0> 2011-05-10 13:53:45 =============================================================================== ns_node_disco_conf_events config all INFO REPORT <0.85.0> 2011-05-10 13:53:45 =============================================================================== Pushing config INFO REPORT <0.85.0> 2011-05-10 13:53:45 =============================================================================== Pushing config done INFO REPORT <0.64.0> 2011-05-10 13:53:45 =============================================================================== config change: buckets -> [{configs,[{"default", [{num_replicas,1}, {ram_quota,1096810496}, {auth_type,sasl}, {sasl_password,[]}, {type,membase}, {num_vbuckets,1024}, {db_shards,4}, {ht_size,3079}, {tap_keepalive,0}, {tap_noop_interval,20}, {max_txn_size,1000}, {ht_locks,5}, {servers,['ns_1@10.1.5.227','ns_1@10.1.5.229']}, {map,[['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227',undefined], ['ns_1@10.1.5.227'|...], [...]|...]}]}]}] INFO REPORT <0.64.0> 2011-05-10 13:53:45 =============================================================================== ns_node_disco_conf_events config all INFO REPORT <0.85.0> 2011-05-10 13:53:45 =============================================================================== Pushing config INFO REPORT <0.85.0> 2011-05-10 13:53:45 =============================================================================== Pushing config done ERROR REPORT <0.362.0> 2011-05-10 13:53:46 =============================================================================== ns_1@10.1.5.227:stats_reader:176: Some nodes didn't respond: ['ns_1@10.1.5.229'] ERROR REPORT <0.360.0> 2011-05-10 13:53:46 =============================================================================== ns_1@10.1.5.227:stats_reader:176: Some nodes didn't respond: ['ns_1@10.1.5.229'] INFO REPORT <0.577.0> 2011-05-10 13:53:46 =============================================================================== ns_1@10.1.5.227:ns_rebalancer:420: Waiting for ['ns_1@10.1.5.229'] INFO REPORT <0.77.0> 2011-05-10 13:53:47 =============================================================================== Detected a new node (from node 'ns_1@10.1.5.227'). Moving config around. INFO REPORT <0.77.0> 2011-05-10 13:53:47 =============================================================================== ns_node_disco_log: nodes changed: ['ns_1@10.1.5.227','ns_1@10.1.5.229'] INFO REPORT <0.85.0> 2011-05-10 13:53:47 =============================================================================== Pulling config INFO REPORT <0.85.0> 2011-05-10 13:53:47 =============================================================================== Pulling config from: 'ns_1@10.1.5.229' INFO REPORT <0.85.0> 2011-05-10 13:53:47 =============================================================================== Pulling config done INFO REPORT <0.577.0> 2011-05-10 13:53:47 =============================================================================== ns_1@10.1.5.227:ns_rebalancer:420: Waiting for ['ns_1@10.1.5.229'] INFO REPORT <0.109.0> 2011-05-10 13:53:48 =============================================================================== ns_1@10.1.5.227:ns_orchestrator:164: Skipping janitor in state rebalancing: {rebalancing_state, <0.577.0>, {dict, 2, 16, 16, 8, 80, 48, {[], [], [], [], [], [], [], [], [], [], [], [], [], [], [], []}, {{[], [], [], [], [], [], [], [['ns_1@10.1.5.227'| 0.0]], [], [['ns_1@10.1.5.229'| 0.0]], [], [], [], [], [], []}}}} INFO REPORT <0.577.0> 2011-05-10 13:53:48 =============================================================================== ns_1@10.1.5.227:ns_rebalancer:420: Waiting for ['ns_1@10.1.5.229'] INFO REPORT <0.577.0> 2011-05-10 13:53:49 =============================================================================== ns_1@10.1.5.227:ns_rebalancer:420: Waiting for ['ns_1@10.1.5.229'] INFO REPORT <0.577.0> 2011-05-10 13:53:50 =============================================================================== ns_1@10.1.5.227:ns_rebalancer:420: Waiting for ['ns_1@10.1.5.229'] INFO REPORT <0.577.0> 2011-05-10 13:53:51 =============================================================================== ns_1@10.1.5.227:ns_rebalancer:420: Waiting for ['ns_1@10.1.5.229'] INFO REPORT <0.577.0> 2011-05-10 13:53:52 =============================================================================== ns_1@10.1.5.227:ns_rebalancer:420: Waiting for ['ns_1@10.1.5.229'] INFO REPORT <0.616.0> 2011-05-10 13:53:53 =============================================================================== vbucketmigrator<0.616.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.616.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.616.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.616.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.616.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.616.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.616.0>: Starting to move bucket 0 INFO REPORT <0.104.0> 2011-05-10 13:53:53 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.109.0> 2011-05-10 13:53:58 =============================================================================== ns_1@10.1.5.227:ns_orchestrator:164: Skipping janitor in state rebalancing: {rebalancing_state, <0.577.0>, {dict, 2, 16, 16, 8, 80, 48, {[], [], [], [], [], [], [], [], [], [], [], [], [], [], [], []}, {{[], [], [], [], [], [], [], [['ns_1@10.1.5.227'| 0.0]], [], [['ns_1@10.1.5.229'| 0.0]], [], [], [], [], [], []}}}} INFO REPORT <0.616.0> 2011-05-10 13:53:58 =============================================================================== vbucketmigrator<0.616.0>: Bucket 0 moved to the next server vbucketmigrator<0.616.0>: Validate bucket states vbucketmigrator<0.616.0>: 0 ok INFO REPORT <0.626.0> 2011-05-10 13:53:58 =============================================================================== vbucketmigrator<0.626.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.626.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.626.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.626.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.626.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.626.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.626.0>: Starting to move bucket 1 vbucketmigrator<0.626.0>: Bucket 1 moved to the next server vbucketmigrator<0.626.0>: Validate bucket states vbucketmigrator<0.626.0>: 1 ok INFO REPORT <0.104.0> 2011-05-10 13:53:59 =============================================================================== memcached<0.104.0>: Vbucket is going dead. memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.628.0> 2011-05-10 13:53:59 =============================================================================== vbucketmigrator<0.628.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.628.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.628.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.628.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.628.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.628.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.628.0>: Starting to move bucket 2 vbucketmigrator<0.628.0>: Bucket 2 moved to the next server INFO REPORT <0.85.0> 2011-05-10 13:54:02 =============================================================================== Pulling config from: 'ns_1@10.1.5.229' INFO REPORT <0.628.0> 2011-05-10 13:54:04 =============================================================================== vbucketmigrator<0.628.0>: Validate bucket states vbucketmigrator<0.628.0>: 2 ok INFO REPORT <0.635.0> 2011-05-10 13:54:04 =============================================================================== vbucketmigrator<0.635.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.635.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.635.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.635.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.635.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.635.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.635.0>: Starting to move bucket 3 vbucketmigrator<0.635.0>: Bucket 3 moved to the next server vbucketmigrator<0.635.0>: Validate bucket states vbucketmigrator<0.635.0>: 3 ok INFO REPORT <0.104.0> 2011-05-10 13:54:05 =============================================================================== memcached<0.104.0>: Vbucket is going dead. memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.637.0> 2011-05-10 13:54:05 =============================================================================== vbucketmigrator<0.637.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.637.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.637.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.637.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.637.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.637.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.637.0>: Starting to move bucket 4 vbucketmigrator<0.637.0>: Bucket 4 moved to the next server INFO REPORT <0.109.0> 2011-05-10 13:54:08 =============================================================================== ns_1@10.1.5.227:ns_orchestrator:164: Skipping janitor in state rebalancing: {rebalancing_state, <0.577.0>, {dict, 2, 16, 16, 8, 80, 48, {[], [], [], [], [], [], [], [], [], [], [], [], [], [], [], []}, {{[], [], [], [], [], [], [], [['ns_1@10.1.5.227'| 0.0078125]], [], [['ns_1@10.1.5.229'| 0.0078125]], [], [], [], [], [], []}}}} INFO REPORT <0.637.0> 2011-05-10 13:54:10 =============================================================================== vbucketmigrator<0.637.0>: Validate bucket states vbucketmigrator<0.637.0>: 4 ok INFO REPORT <0.655.0> 2011-05-10 13:54:11 =============================================================================== vbucketmigrator<0.655.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.655.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.655.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.655.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.655.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.655.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.655.0>: Starting to move bucket 5 vbucketmigrator<0.655.0>: Bucket 5 moved to the next server INFO REPORT <0.104.0> 2011-05-10 13:54:11 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.655.0> 2011-05-10 13:54:16 =============================================================================== vbucketmigrator<0.655.0>: Validate bucket states vbucketmigrator<0.655.0>: 5 ok INFO REPORT <0.662.0> 2011-05-10 13:54:17 =============================================================================== vbucketmigrator<0.662.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.662.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.662.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.662.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.662.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.662.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.662.0>: Starting to move bucket 6 vbucketmigrator<0.662.0>: Bucket 6 moved to the next server INFO REPORT <0.104.0> 2011-05-10 13:54:17 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.109.0> 2011-05-10 13:54:18 =============================================================================== ns_1@10.1.5.227:ns_orchestrator:164: Skipping janitor in state rebalancing: {rebalancing_state, <0.577.0>, {dict, 2, 16, 16, 8, 80, 48, {[], [], [], [], [], [], [], [], [], [], [], [], [], [], [], []}, {{[], [], [], [], [], [], [], [['ns_1@10.1.5.227'| 0.01171875]], [], [['ns_1@10.1.5.229'| 0.01171875]], [], [], [], [], [], []}}}} INFO REPORT <0.662.0> 2011-05-10 13:54:22 =============================================================================== vbucketmigrator<0.662.0>: Validate bucket states vbucketmigrator<0.662.0>: 6 ok INFO REPORT <0.673.0> 2011-05-10 13:54:23 =============================================================================== vbucketmigrator<0.673.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.673.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.673.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.673.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.673.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.673.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.673.0>: Starting to move bucket 7 vbucketmigrator<0.673.0>: Bucket 7 moved to the next server INFO REPORT <0.104.0> 2011-05-10 13:54:23 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.109.0> 2011-05-10 13:54:28 =============================================================================== ns_1@10.1.5.227:ns_orchestrator:164: Skipping janitor in state rebalancing: {rebalancing_state, <0.577.0>, {dict, 2, 16, 16, 8, 80, 48, {[], [], [], [], [], [], [], [], [], [], [], [], [], [], [], []}, {{[], [], [], [], [], [], [], [['ns_1@10.1.5.227'| 0.013671875]], [], [['ns_1@10.1.5.229'| 0.013671875]], [], [], [], [], [], []}}}} INFO REPORT <0.673.0> 2011-05-10 13:54:28 =============================================================================== vbucketmigrator<0.673.0>: Validate bucket states vbucketmigrator<0.673.0>: 7 ok INFO REPORT <0.689.0> 2011-05-10 13:54:29 =============================================================================== vbucketmigrator<0.689.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.689.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.689.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.689.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.689.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.689.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.689.0>: Starting to move bucket 8 vbucketmigrator<0.689.0>: Bucket 8 moved to the next server INFO REPORT <0.104.0> 2011-05-10 13:54:29 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.689.0> 2011-05-10 13:54:34 =============================================================================== vbucketmigrator<0.689.0>: Validate bucket states vbucketmigrator<0.689.0>: 8 ok INFO REPORT <0.693.0> 2011-05-10 13:54:35 =============================================================================== vbucketmigrator<0.693.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.693.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.693.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.693.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.693.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.693.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.693.0>: Starting to move bucket 9 vbucketmigrator<0.693.0>: Bucket 9 moved to the next server INFO REPORT <0.104.0> 2011-05-10 13:54:35 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.109.0> 2011-05-10 13:54:38 =============================================================================== ns_1@10.1.5.227:ns_orchestrator:164: Skipping janitor in state rebalancing: {rebalancing_state, <0.577.0>, {dict, 2, 16, 16, 8, 80, 48, {[], [], [], [], [], [], [], [], [], [], [], [], [], [], [], []}, {{[], [], [], [], [], [], [], [['ns_1@10.1.5.227'| 0.017578125]], [], [['ns_1@10.1.5.229'| 0.017578125]], [], [], [], [], [], []}}}} INFO REPORT <0.693.0> 2011-05-10 13:54:40 =============================================================================== vbucketmigrator<0.693.0>: Validate bucket states vbucketmigrator<0.693.0>: 9 ok INFO REPORT <0.698.0> 2011-05-10 13:54:41 =============================================================================== vbucketmigrator<0.698.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.698.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.698.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.698.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.698.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.698.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.698.0>: Starting to move bucket 10 vbucketmigrator<0.698.0>: Bucket 10 moved to the next server INFO REPORT <0.104.0> 2011-05-10 13:54:41 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.474.0> 2011-05-10 13:54:43 =============================================================================== ns_1@10.1.5.227:ns_doctor:82: Current node statuses: [{'ns_1@10.1.5.227', [{last_heard,{1305,60882,411240}}, {active_buckets,["default"]}, {memory, [{total,33499920}, {processes,14019448}, {processes_used,14008472}, {system,19480472}, {atom,814593}, {atom_used,800106}, {binary,115480}, {code,8389348}, {ets,822696}]}, {cluster_compatibility_version,1}, {version, [{kernel,"2.13.4"}, {mnesia,"4.4.12"}, {stdlib,"1.16.4"}, {os_mon,"2.2.4"}, {inets,"5.2"}, {ns_server,"1.6.5.4r"}, {menelaus,"1.6.5.4r"}, {sasl,"2.1.8"}]}, {system_arch,"x86_64-unknown-linux-gnu"}, {wall_clock,274}, {memory_data,{2058731520,663089152,{<0.65.0>,1572192}}}, {disk_data, [{"/",11204616,49},{"/boot",101086,14},{"/dev/shm",1005240,0}]}, {meminfo, <<"MemTotal: 2010480 kB\nMemFree: 1359584 kB\nBuffers: 61224 kB\nCached: 340800 kB\nSwapCached: 20084 kB\nActive: 342480 kB\nInactive: 196228 kB\nHighTotal: 0 kB\nHighFree: 0 kB\nLowTotal: 2010480 kB\nLowFree: 1359584 kB\nSwapTotal: 4030456 kB\nSwapFree: 3952288 kB\nDirty: 22168 kB\nWriteback: 0 kB\nAnonPages: 134508 kB\nMapped: 15732 kB\nSlab: 86332 kB\nPageTables: 6580 kB\nNFS_Unstable: 0 kB\nBounce: 0 kB\nCommitLimit: 5035696 kB\nCommitted_AS: 540396 kB\nVmallocTotal: 34359738367 kB\nVmallocUsed: 265484 kB\nVmallocChunk: 34359471011 kB\nHugePages_Total: 0\nHugePages_Free: 0\nHugePages_Rsvd: 0\nHugepagesize: 2048 kB\n">>}, {replication,[{"default",1.0}]}, {system_memory_data, [{system_total_memory,2058731520}, {free_swap,4047142912}, {total_swap,4127186944}, {cached_memory,348979200}, {buffered_memory,62693376}, {free_memory,1392214016}, {total_memory,2058731520}]}, {statistics, [{wall_clock,{270579,1}}, {context_switches,{45576,0}}, {garbage_collection,{8878,41515427,0}}, {io,{{input,5609535},{output,2479658}}}, {reductions,{12073361,632897}}, {run_queue,0}, {runtime,{1940,110}}]}]}, {'ns_1@10.1.5.229', [{last_heard,{1305,60882,242050}}, {active_buckets,["default"]}, {memory, [{total,27314832}, {processes,8504368}, {processes_used,8497328}, {system,18810464}, {atom,778121}, {atom_used,752745}, {binary,227504}, {code,7791241}, {ets,735344}]}, {cluster_compatibility_version,1}, {version, [{kernel,"2.13.4"}, {mnesia,"4.4.12"}, {stdlib,"1.16.4"}, {os_mon,"2.2.4"}, {ns_server,"1.6.5.4r"}, {menelaus,"1.6.5.4r"}, {sasl,"2.1.8"}]}, {system_arch,"x86_64-unknown-linux-gnu"}, {wall_clock,237}, {memory_data,{2058731520,657874944,{<9352.365.0>,1343120}}}, {disk_data, [{"/",11204616,44},{"/boot",101086,14},{"/dev/shm",1005240,0}]}, {meminfo, <<"MemTotal: 2010480 kB\nMemFree: 1327504 kB\nBuffers: 75596 kB\nCached: 338452 kB\nSwapCached: 0 kB\nActive: 355324 kB\nInactive: 210312 kB\nHighTotal: 0 kB\nHighFree: 0 kB\nLowTotal: 2010480 kB\nLowFree: 1327504 kB\nSwapTotal: 4030456 kB\nSwapFree: 4030388 kB\nDirty: 25576 kB\nWriteback: 0 kB\nAnonPages: 151580 kB\nMapped: 28104 kB\nSlab: 91944 kB\nPageTables: 5840 kB\nNFS_Unstable: 0 kB\nBounce: 0 kB\nCommitLimit: 5035696 kB\nCommitted_AS: 473428 kB\nVmallocTotal: 34359738367 kB\nVmallocUsed: 265484 kB\nVmallocChunk: 34359471011 kB\nHugePages_Total: 0\nHugePages_Free: 0\nHugePages_Rsvd: 0\nHugepagesize: 2048 kB\n">>}, {replication,[{"default",1.0}]}, {system_memory_data, [{system_total_memory,2058731520}, {free_swap,4127117312}, {total_swap,4127186944}, {cached_memory,346574848}, {buffered_memory,77410304}, {free_memory,1359364096}, {total_memory,2058731520}]}, {statistics, [{wall_clock,{229806,1}}, {context_switches,{24602,0}}, {garbage_collection,{5026,12597596,0}}, {io,{{input,4547110},{output,867902}}}, {reductions,{4052550,233983}}, {run_queue,0}, {runtime,{800,30}}]}]}] PROGRESS REPORT <0.34.0> 2011-05-10 13:54:45 =============================================================================== supervisor {local,sasl_sup} started [{pid,<0.703.0>}, {name,rb_server}, {mfa,{rb,start_link, [[{report_dir,"/var/opt/membase/1.6.5.4r/logs"}, {start_log,"/opt/membase/1.6.5.4r/tmp/nslogs700583_16139.log"}, {type,all}, {max,all}, {report_dir}]]}}, {restart_type,temporary}, {shutdown,brutal_kill}, {child_type,worker}] INFO REPORT <0.698.0> 2011-05-10 13:54:46 =============================================================================== vbucketmigrator<0.698.0>: Validate bucket states vbucketmigrator<0.698.0>: 10 ok INFO REPORT <0.2105.0> 2011-05-10 13:54:47 =============================================================================== vbucketmigrator<0.2105.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.2105.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.2105.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.2105.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.2105.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.2105.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.2105.0>: Starting to move bucket 11 vbucketmigrator<0.2105.0>: Bucket 11 moved to the next server INFO REPORT <0.104.0> 2011-05-10 13:54:47 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.109.0> 2011-05-10 13:54:48 =============================================================================== ns_1@10.1.5.227:ns_orchestrator:164: Skipping janitor in state rebalancing: {rebalancing_state, <0.577.0>, {dict, 2, 16, 16, 8, 80, 48, {[], [], [], [], [], [], [], [], [], [], [], [], [], [], [], []}, {{[], [], [], [], [], [], [], [['ns_1@10.1.5.227'| 0.021484375]], [], [['ns_1@10.1.5.229'| 0.021484375]], [], [], [], [], [], []}}}} INFO REPORT <0.2105.0> 2011-05-10 13:54:53 =============================================================================== vbucketmigrator<0.2105.0>: Validate bucket states vbucketmigrator<0.2105.0>: 11 ok INFO REPORT <0.2118.0> 2011-05-10 13:54:54 =============================================================================== vbucketmigrator<0.2118.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.2118.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.2118.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.2118.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.2118.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.2118.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.2118.0>: Starting to move bucket 12 vbucketmigrator<0.2118.0>: Bucket 12 moved to the next server INFO REPORT <0.104.0> 2011-05-10 13:54:54 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.85.0> 2011-05-10 13:54:57 =============================================================================== Pulling config from: 'ns_1@10.1.5.229' INFO REPORT <0.109.0> 2011-05-10 13:54:58 =============================================================================== ns_1@10.1.5.227:ns_orchestrator:164: Skipping janitor in state rebalancing: {rebalancing_state, <0.577.0>, {dict, 2, 16, 16, 8, 80, 48, {[], [], [], [], [], [], [], [], [], [], [], [], [], [], [], []}, {{[], [], [], [], [], [], [], [['ns_1@10.1.5.227'| 0.0234375]], [], [['ns_1@10.1.5.229'| 0.0234375]], [], [], [], [], [], []}}}} INFO REPORT <0.2118.0> 2011-05-10 13:54:58 =============================================================================== vbucketmigrator<0.2118.0>: Validate bucket states vbucketmigrator<0.2118.0>: 12 ok INFO REPORT <0.2125.0> 2011-05-10 13:54:59 =============================================================================== vbucketmigrator<0.2125.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.2125.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.2125.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.2125.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.2125.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.2125.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.2125.0>: Starting to move bucket 13 vbucketmigrator<0.2125.0>: Bucket 13 moved to the next server INFO REPORT <0.104.0> 2011-05-10 13:54:59 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.2125.0> 2011-05-10 13:55:04 =============================================================================== vbucketmigrator<0.2125.0>: Validate bucket states vbucketmigrator<0.2125.0>: 13 ok INFO REPORT <0.2129.0> 2011-05-10 13:55:05 =============================================================================== vbucketmigrator<0.2129.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.2129.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.2129.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.2129.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.2129.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.2129.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.2129.0>: Starting to move bucket 14 vbucketmigrator<0.2129.0>: Bucket 14 moved to the next server INFO REPORT <0.104.0> 2011-05-10 13:55:06 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.109.0> 2011-05-10 13:55:08 =============================================================================== ns_1@10.1.5.227:ns_orchestrator:164: Skipping janitor in state rebalancing: {rebalancing_state, <0.577.0>, {dict, 2, 16, 16, 8, 80, 48, {[], [], [], [], [], [], [], [], [], [], [], [], [], [], [], []}, {{[], [], [], [], [], [], [], [['ns_1@10.1.5.227'| 0.02734375]], [], [['ns_1@10.1.5.229'| 0.02734375]], [], [], [], [], [], []}}}} INFO REPORT <0.2129.0> 2011-05-10 13:55:11 =============================================================================== vbucketmigrator<0.2129.0>: Validate bucket states vbucketmigrator<0.2129.0>: 14 ok INFO REPORT <0.2143.0> 2011-05-10 13:55:12 =============================================================================== vbucketmigrator<0.2143.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.2143.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.2143.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.2143.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.2143.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.2143.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.2143.0>: Starting to move bucket 15 vbucketmigrator<0.2143.0>: Bucket 15 moved to the next server INFO REPORT <0.104.0> 2011-05-10 13:55:12 =============================================================================== memcached<0.104.0>: Vbucket is going dead. PROGRESS REPORT <0.34.0> 2011-05-10 13:55:15 =============================================================================== supervisor {local,sasl_sup} started [{pid,<0.2149.0>}, {name,rb_server}, {mfa,{rb,start_link, [[{report_dir,"/var/opt/membase/1.6.5.4r/logs"}, {start_log,"/opt/membase/1.6.5.4r/tmp/nslogs741910_16139.log"}, {type,all}, {max,all}, {report_dir}]]}}, {restart_type,temporary}, {shutdown,brutal_kill}, {child_type,worker}] INFO REPORT <0.2143.0> 2011-05-10 13:55:16 =============================================================================== vbucketmigrator<0.2143.0>: Validate bucket states vbucketmigrator<0.2143.0>: 15 ok INFO REPORT <0.3571.0> 2011-05-10 13:55:17 =============================================================================== vbucketmigrator<0.3571.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.3571.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.3571.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.3571.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.3571.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.3571.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.3571.0>: Starting to move bucket 16 vbucketmigrator<0.3571.0>: Bucket 16 moved to the next server INFO REPORT <0.104.0> 2011-05-10 13:55:17 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.109.0> 2011-05-10 13:55:18 =============================================================================== ns_1@10.1.5.227:ns_orchestrator:164: Skipping janitor in state rebalancing: {rebalancing_state, <0.577.0>, {dict, 2, 16, 16, 8, 80, 48, {[], [], [], [], [], [], [], [], [], [], [], [], [], [], [], []}, {{[], [], [], [], [], [], [], [['ns_1@10.1.5.227'| 0.03125]], [], [['ns_1@10.1.5.229'| 0.03125]], [], [], [], [], [], []}}}} INFO REPORT <0.3571.0> 2011-05-10 13:55:22 =============================================================================== vbucketmigrator<0.3571.0>: Validate bucket states vbucketmigrator<0.3571.0>: 16 ok INFO REPORT <0.3576.0> 2011-05-10 13:55:23 =============================================================================== vbucketmigrator<0.3576.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.3576.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.3576.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.3576.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.3576.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.3576.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.3576.0>: Starting to move bucket 17 vbucketmigrator<0.3576.0>: Bucket 17 moved to the next server INFO REPORT <0.104.0> 2011-05-10 13:55:23 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.469.0> 2011-05-10 13:55:24 =============================================================================== ns_1@10.1.5.227:stats_collector:71: Stats for bucket "default": auth_cmds 50 auth_errors 0 bucket_conns 34 bytes_read 1580033 bytes_written 1663025 cas_badval 0 cas_hits 0 cas_misses 0 cmd_flush 1 cmd_get 0 cmd_set 16004 conn_yields 0 connection_structures 44 curr_connections 44 curr_items 7898 curr_items_tot 8040 daemon_connections 10 decr_hits 0 decr_misses 0 delete_hits 0 delete_misses 0 ep_bg_fetched 0 ep_commit_num 53 ep_commit_time 0 ep_commit_time_total 4 ep_data_age 13 ep_data_age_highwat 13 ep_db_cleaner_status complete ep_db_strategy multiMTVBDB ep_dbinit 1 ep_dbname /var/opt/membase/1.6.5.4r/data/ns_1/default ep_dbshards 4 ep_expired 0 ep_flush_duration 1 ep_flush_duration_highwat 8 ep_flush_duration_total 12 ep_flush_preempts 0 ep_flusher_state running ep_flusher_todo 0 ep_io_num_read 142 ep_io_num_write 15741 ep_io_read_bytes 8091 ep_io_write_bytes 1011432 ep_item_begin_failed 0 ep_item_commit_failed 0 ep_item_flush_expired 0 ep_item_flush_failed 0 ep_kv_size 1181400 ep_max_data_size 1096810496 ep_max_txn_size 1000 ep_mem_high_wat 822607872 ep_mem_low_wat 658086297 ep_min_data_age 0 ep_num_active_non_resident 0 ep_num_eject_failures 0 ep_num_eject_replicas 0 ep_num_expiry_pager_runs 0 ep_num_non_resident 0 ep_num_not_my_vbuckets 0 ep_num_pager_runs 0 ep_num_value_ejects 0 ep_oom_errors 0 ep_overhead 25854448 ep_pending_ops 0 ep_pending_ops_max 0 ep_pending_ops_max_duration 0 ep_pending_ops_total 0 ep_queue_age_cap 900 ep_queue_size 0 ep_storage_age 0 ep_storage_age_highwat 10 ep_storage_type featured ep_store_max_concurrency 10 ep_store_max_readers 9 ep_store_max_readwrite 1 ep_tap_bg_fetch_requeued 0 ep_tap_bg_fetched 142 ep_tap_keepalive 300 ep_tmp_oom_errors 0 ep_too_old 0 ep_too_young 0 ep_total_cache_size 1181400 ep_total_del_items 0 ep_total_enqueued 16035 ep_total_new_items 15741 ep_total_persisted 15741 ep_vbucket_del 0 ep_vbucket_del_fail 0 ep_version 1.6.5.3.1_13_g81a6cf1 ep_warmed_up 0 ep_warmup true ep_warmup_dups 0 ep_warmup_oom 0 ep_warmup_thread complete ep_warmup_time 15662 get_hits 0 get_misses 0 incr_hits 0 incr_misses 0 libevent 1.4.13-stable limit_maxbytes 67108864 mem_used 27035848 pid 16183 pointer_size 64 rejected_conns 0 rusage_system 1.160823 rusage_user 2.514617 tap_connect_received 18 tap_mutation_sent 284 tap_opaque_sent 18 tap_vbucket_set_sent 36 threads 4 time 1305060923 total_connections 61 uptime 317 version 1.4.4_382_g9df3289 INFO REPORT <0.109.0> 2011-05-10 13:55:28 =============================================================================== ns_1@10.1.5.227:ns_orchestrator:164: Skipping janitor in state rebalancing: {rebalancing_state, <0.577.0>, {dict, 2, 16, 16, 8, 80, 48, {[], [], [], [], [], [], [], [], [], [], [], [], [], [], [], []}, {{[], [], [], [], [], [], [], [['ns_1@10.1.5.227'| 0.033203125]], [], [['ns_1@10.1.5.229'| 0.033203125]], [], [], [], [], [], []}}}} INFO REPORT <0.85.0> 2011-05-10 13:55:28 =============================================================================== Pulling config from: 'ns_1@10.1.5.229' INFO REPORT <0.3576.0> 2011-05-10 13:55:28 =============================================================================== vbucketmigrator<0.3576.0>: Validate bucket states vbucketmigrator<0.3576.0>: 17 ok INFO REPORT <0.3591.0> 2011-05-10 13:55:30 =============================================================================== vbucketmigrator<0.3591.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.3591.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.3591.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.3591.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.3591.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.3591.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.3591.0>: Starting to move bucket 18 vbucketmigrator<0.3591.0>: Bucket 18 moved to the next server INFO REPORT <0.104.0> 2011-05-10 13:55:30 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.3591.0> 2011-05-10 13:55:35 =============================================================================== vbucketmigrator<0.3591.0>: Validate bucket states vbucketmigrator<0.3591.0>: 18 ok INFO REPORT <0.3595.0> 2011-05-10 13:55:36 =============================================================================== vbucketmigrator<0.3595.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.3595.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.3595.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.3595.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.3595.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.3595.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.3595.0>: Starting to move bucket 19 vbucketmigrator<0.3595.0>: Bucket 19 moved to the next server INFO REPORT <0.104.0> 2011-05-10 13:55:36 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.109.0> 2011-05-10 13:55:38 =============================================================================== ns_1@10.1.5.227:ns_orchestrator:164: Skipping janitor in state rebalancing: {rebalancing_state, <0.577.0>, {dict, 2, 16, 16, 8, 80, 48, {[], [], [], [], [], [], [], [], [], [], [], [], [], [], [], []}, {{[], [], [], [], [], [], [], [['ns_1@10.1.5.227'| 0.037109375]], [], [['ns_1@10.1.5.229'| 0.037109375]], [], [], [], [], [], []}}}} INFO REPORT <0.3595.0> 2011-05-10 13:55:40 =============================================================================== vbucketmigrator<0.3595.0>: Validate bucket states vbucketmigrator<0.3595.0>: 19 ok INFO REPORT <0.85.0> 2011-05-10 13:55:41 =============================================================================== Pulling config from: 'ns_1@10.1.5.229' INFO REPORT <0.3600.0> 2011-05-10 13:55:41 =============================================================================== vbucketmigrator<0.3600.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.3600.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.3600.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.3600.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.3600.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.3600.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.3600.0>: Starting to move bucket 20 vbucketmigrator<0.3600.0>: Bucket 20 moved to the next server INFO REPORT <0.104.0> 2011-05-10 13:55:41 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.474.0> 2011-05-10 13:55:43 =============================================================================== ns_1@10.1.5.227:ns_doctor:82: Current node statuses: [{'ns_1@10.1.5.227', [{last_heard,{1305,60942,411182}}, {active_buckets,["default"]}, {memory, [{total,59128208}, {processes,39271280}, {processes_used,39261168}, {system,19856928}, {atom,819441}, {atom_used,806779}, {binary,90080}, {code,8552890}, {ets,890936}]}, {cluster_compatibility_version,1}, {version, [{kernel,"2.13.4"}, {mnesia,"4.4.12"}, {stdlib,"1.16.4"}, {os_mon,"2.2.4"}, {inets,"5.2"}, {ns_server,"1.6.5.4r"}, {menelaus,"1.6.5.4r"}, {sasl,"2.1.8"}]}, {system_arch,"x86_64-unknown-linux-gnu"}, {wall_clock,334}, {memory_data,{2058731520,674390016,{<0.639.0>,12342384}}}, {disk_data, [{"/",11204616,49},{"/boot",101086,14},{"/dev/shm",1005240,0}]}, {meminfo, <<"MemTotal: 2010480 kB\nMemFree: 1356236 kB\nBuffers: 61576 kB\nCached: 341484 kB\nSwapCached: 20084 kB\nActive: 345316 kB\nInactive: 196680 kB\nHighTotal: 0 kB\nHighFree: 0 kB\nLowTotal: 2010480 kB\nLowFree: 1356236 kB\nSwapTotal: 4030456 kB\nSwapFree: 3952288 kB\nDirty: 13796 kB\nWriteback: 0 kB\nAnonPages: 136716 kB\nMapped: 15732 kB\nSlab: 86384 kB\nPageTables: 6608 kB\nNFS_Unstable: 0 kB\nBounce: 0 kB\nCommitLimit: 5035696 kB\nCommitted_AS: 568384 kB\nVmallocTotal: 34359738367 kB\nVmallocUsed: 265484 kB\nVmallocChunk: 34359471011 kB\nHugePages_Total: 0\nHugePages_Free: 0\nHugePages_Rsvd: 0\nHugepagesize: 2048 kB\n">>}, {replication,[{"default",1.0}]}, {system_memory_data, [{system_total_memory,2058731520}, {free_swap,4047142912}, {total_swap,4127186944}, {cached_memory,349679616}, {buffered_memory,63053824}, {free_memory,1388785664}, {total_memory,2058731520}]}, {statistics, [{wall_clock,{330579,2}}, {context_switches,{166380,0}}, {garbage_collection,{30988,98251758,0}}, {io,{{input,53332577},{output,6496797}}}, {reductions,{36635086,727336}}, {run_queue,0}, {runtime,{4710,120}}]}]}, {'ns_1@10.1.5.229', [{last_heard,{1305,60942,242091}}, {active_buckets,["default"]}, {memory, [{total,31466408}, {processes,12581008}, {processes_used,12573968}, {system,18885400}, {atom,778121}, {atom_used,752794}, {binary,218608}, {code,7791241}, {ets,818560}]}, {cluster_compatibility_version,1}, {version, [{kernel,"2.13.4"}, {mnesia,"4.4.12"}, {stdlib,"1.16.4"}, {os_mon,"2.2.4"}, {ns_server,"1.6.5.4r"}, {menelaus,"1.6.5.4r"}, {sasl,"2.1.8"}]}, {system_arch,"x86_64-unknown-linux-gnu"}, {wall_clock,297}, {memory_data,{2058731520,699621376,{<9352.365.0>,1572376}}}, {disk_data, [{"/",11204616,44},{"/boot",101086,14},{"/dev/shm",1005240,0}]}, {meminfo, <<"MemTotal: 2010480 kB\nMemFree: 1323040 kB\nBuffers: 76300 kB\nCached: 338792 kB\nSwapCached: 0 kB\nActive: 359160 kB\nInactive: 210864 kB\nHighTotal: 0 kB\nHighFree: 0 kB\nLowTotal: 2010480 kB\nLowFree: 1323040 kB\nSwapTotal: 4030456 kB\nSwapFree: 4030388 kB\nDirty: 17288 kB\nWriteback: 0 kB\nAnonPages: 154920 kB\nMapped: 28104 kB\nSlab: 91968 kB\nPageTables: 5856 kB\nNFS_Unstable: 0 kB\nBounce: 0 kB\nCommitLimit: 5035696 kB\nCommitted_AS: 478036 kB\nVmallocTotal: 34359738367 kB\nVmallocUsed: 265484 kB\nVmallocChunk: 34359471011 kB\nHugePages_Total: 0\nHugePages_Free: 0\nHugePages_Rsvd: 0\nHugepagesize: 2048 kB\n">>}, {replication,[{"default",1.0}]}, {system_memory_data, [{system_total_memory,2058731520}, {free_swap,4127117312}, {total_swap,4127186944}, {cached_memory,346923008}, {buffered_memory,78131200}, {free_memory,1354792960}, {total_memory,2058731520}]}, {statistics, [{wall_clock,{289806,1}}, {context_switches,{30451,0}}, {garbage_collection,{6162,19743500,0}}, {io,{{input,4770370},{output,1127003}}}, {reductions,{7562733,236936}}, {run_queue,0}, {runtime,{1160,50}}]}]}] INFO REPORT <0.3600.0> 2011-05-10 13:55:47 =============================================================================== vbucketmigrator<0.3600.0>: Validate bucket states vbucketmigrator<0.3600.0>: 20 ok INFO REPORT <0.3604.0> 2011-05-10 13:55:48 =============================================================================== vbucketmigrator<0.3604.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.3604.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.3604.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.3604.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.3604.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.3604.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.3604.0>: Starting to move bucket 21 vbucketmigrator<0.3604.0>: Bucket 21 moved to the next server INFO REPORT <0.104.0> 2011-05-10 13:55:48 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.109.0> 2011-05-10 13:55:48 =============================================================================== ns_1@10.1.5.227:ns_orchestrator:164: Skipping janitor in state rebalancing: {rebalancing_state, <0.577.0>, {dict, 2, 16, 16, 8, 80, 48, {[], [], [], [], [], [], [], [], [], [], [], [], [], [], [], []}, {{[], [], [], [], [], [], [], [['ns_1@10.1.5.227'| 0.041015625]], [], [['ns_1@10.1.5.229'| 0.041015625]], [], [], [], [], [], []}}}} INFO REPORT <0.3604.0> 2011-05-10 13:55:53 =============================================================================== vbucketmigrator<0.3604.0>: Validate bucket states vbucketmigrator<0.3604.0>: 21 ok INFO REPORT <0.3617.0> 2011-05-10 13:55:54 =============================================================================== vbucketmigrator<0.3617.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.3617.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.3617.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.3617.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.3617.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.3617.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.3617.0>: Starting to move bucket 22 vbucketmigrator<0.3617.0>: Bucket 22 moved to the next server INFO REPORT <0.104.0> 2011-05-10 13:55:54 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.109.0> 2011-05-10 13:55:58 =============================================================================== ns_1@10.1.5.227:ns_orchestrator:164: Skipping janitor in state rebalancing: {rebalancing_state, <0.577.0>, {dict, 2, 16, 16, 8, 80, 48, {[], [], [], [], [], [], [], [], [], [], [], [], [], [], [], []}, {{[], [], [], [], [], [], [], [['ns_1@10.1.5.227'| 0.04296875]], [], [['ns_1@10.1.5.229'| 0.04296875]], [], [], [], [], [], []}}}} INFO REPORT <0.3617.0> 2011-05-10 13:55:58 =============================================================================== vbucketmigrator<0.3617.0>: Validate bucket states vbucketmigrator<0.3617.0>: 22 ok INFO REPORT <0.3624.0> 2011-05-10 13:55:59 =============================================================================== vbucketmigrator<0.3624.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.3624.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.3624.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.3624.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.3624.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.3624.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.3624.0>: Starting to move bucket 23 vbucketmigrator<0.3624.0>: Bucket 23 moved to the next server INFO REPORT <0.104.0> 2011-05-10 13:55:59 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.3624.0> 2011-05-10 13:56:05 =============================================================================== vbucketmigrator<0.3624.0>: Validate bucket states vbucketmigrator<0.3624.0>: 23 ok INFO REPORT <0.3628.0> 2011-05-10 13:56:06 =============================================================================== vbucketmigrator<0.3628.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.3628.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.3628.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.3628.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.3628.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.3628.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.3628.0>: Starting to move bucket 24 vbucketmigrator<0.3628.0>: Bucket 24 moved to the next server INFO REPORT <0.104.0> 2011-05-10 13:56:06 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.109.0> 2011-05-10 13:56:08 =============================================================================== ns_1@10.1.5.227:ns_orchestrator:164: Skipping janitor in state rebalancing: {rebalancing_state, <0.577.0>, {dict, 2, 16, 16, 8, 80, 48, {[], [], [], [], [], [], [], [], [], [], [], [], [], [], [], []}, {{[], [], [], [], [], [], [], [['ns_1@10.1.5.227'| 0.046875]], [], [['ns_1@10.1.5.229'| 0.046875]], [], [], [], [], [], []}}}} INFO REPORT <0.3628.0> 2011-05-10 13:56:11 =============================================================================== vbucketmigrator<0.3628.0>: Validate bucket states vbucketmigrator<0.3628.0>: 24 ok INFO REPORT <0.3642.0> 2011-05-10 13:56:12 =============================================================================== vbucketmigrator<0.3642.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.3642.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.3642.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.3642.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.3642.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.3642.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.3642.0>: Starting to move bucket 25 vbucketmigrator<0.3642.0>: Bucket 25 moved to the next server INFO REPORT <0.104.0> 2011-05-10 13:56:12 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.3642.0> 2011-05-10 13:56:16 =============================================================================== vbucketmigrator<0.3642.0>: Validate bucket states vbucketmigrator<0.3642.0>: 25 ok INFO REPORT <0.3646.0> 2011-05-10 13:56:18 =============================================================================== vbucketmigrator<0.3646.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.3646.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.3646.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.3646.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.3646.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.3646.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.3646.0>: Starting to move bucket 26 vbucketmigrator<0.3646.0>: Bucket 26 moved to the next server INFO REPORT <0.104.0> 2011-05-10 13:56:18 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.109.0> 2011-05-10 13:56:18 =============================================================================== ns_1@10.1.5.227:ns_orchestrator:164: Skipping janitor in state rebalancing: {rebalancing_state, <0.577.0>, {dict, 2, 16, 16, 8, 80, 48, {[], [], [], [], [], [], [], [], [], [], [], [], [], [], [], []}, {{[], [], [], [], [], [], [], [['ns_1@10.1.5.227'| 0.05078125]], [], [['ns_1@10.1.5.229'| 0.05078125]], [], [], [], [], [], []}}}} INFO REPORT <0.3646.0> 2011-05-10 13:56:23 =============================================================================== vbucketmigrator<0.3646.0>: Validate bucket states vbucketmigrator<0.3646.0>: 26 ok INFO REPORT <0.3651.0> 2011-05-10 13:56:24 =============================================================================== vbucketmigrator<0.3651.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.3651.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.3651.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.3651.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.3651.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.3651.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.3651.0>: Starting to move bucket 27 vbucketmigrator<0.3651.0>: Bucket 27 moved to the next server INFO REPORT <0.104.0> 2011-05-10 13:56:24 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.85.0> 2011-05-10 13:56:24 =============================================================================== Pulling config from: 'ns_1@10.1.5.229' INFO REPORT <0.109.0> 2011-05-10 13:56:28 =============================================================================== ns_1@10.1.5.227:ns_orchestrator:164: Skipping janitor in state rebalancing: {rebalancing_state, <0.577.0>, {dict, 2, 16, 16, 8, 80, 48, {[], [], [], [], [], [], [], [], [], [], [], [], [], [], [], []}, {{[], [], [], [], [], [], [], [['ns_1@10.1.5.227'| 0.052734375]], [], [['ns_1@10.1.5.229'| 0.052734375]], [], [], [], [], [], []}}}} INFO REPORT <0.3651.0> 2011-05-10 13:56:28 =============================================================================== vbucketmigrator<0.3651.0>: Validate bucket states vbucketmigrator<0.3651.0>: 27 ok INFO REPORT <0.3666.0> 2011-05-10 13:56:29 =============================================================================== vbucketmigrator<0.3666.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.3666.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.3666.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.3666.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.3666.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.3666.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.3666.0>: Starting to move bucket 28 vbucketmigrator<0.3666.0>: Bucket 28 moved to the next server INFO REPORT <0.104.0> 2011-05-10 13:56:29 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.3666.0> 2011-05-10 13:56:35 =============================================================================== vbucketmigrator<0.3666.0>: Validate bucket states vbucketmigrator<0.3666.0>: 28 ok INFO REPORT <0.3670.0> 2011-05-10 13:56:36 =============================================================================== vbucketmigrator<0.3670.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.3670.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.3670.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.3670.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.3670.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.3670.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.3670.0>: Starting to move bucket 29 vbucketmigrator<0.3670.0>: Bucket 29 moved to the next server INFO REPORT <0.104.0> 2011-05-10 13:56:36 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.109.0> 2011-05-10 13:56:38 =============================================================================== ns_1@10.1.5.227:ns_orchestrator:164: Skipping janitor in state rebalancing: {rebalancing_state, <0.577.0>, {dict, 2, 16, 16, 8, 80, 48, {[], [], [], [], [], [], [], [], [], [], [], [], [], [], [], []}, {{[], [], [], [], [], [], [], [['ns_1@10.1.5.227'| 0.056640625]], [], [['ns_1@10.1.5.229'| 0.056640625]], [], [], [], [], [], []}}}} INFO REPORT <0.3670.0> 2011-05-10 13:56:41 =============================================================================== vbucketmigrator<0.3670.0>: Validate bucket states vbucketmigrator<0.3670.0>: 29 ok INFO REPORT <0.3675.0> 2011-05-10 13:56:42 =============================================================================== vbucketmigrator<0.3675.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.3675.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.3675.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.3675.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.3675.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.3675.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.3675.0>: Starting to move bucket 30 vbucketmigrator<0.3675.0>: Bucket 30 moved to the next server INFO REPORT <0.104.0> 2011-05-10 13:56:42 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.474.0> 2011-05-10 13:56:43 =============================================================================== ns_1@10.1.5.227:ns_doctor:82: Current node statuses: [{'ns_1@10.1.5.227', [{last_heard,{1305,61002,411213}}, {active_buckets,["default"]}, {memory, [{total,58862288}, {processes,38897304}, {processes_used,38886328}, {system,19964984}, {atom,819441}, {atom_used,806779}, {binary,134952}, {code,8552890}, {ets,953784}]}, {cluster_compatibility_version,1}, {version, [{kernel,"2.13.4"}, {mnesia,"4.4.12"}, {stdlib,"1.16.4"}, {os_mon,"2.2.4"}, {inets,"5.2"}, {ns_server,"1.6.5.4r"}, {menelaus,"1.6.5.4r"}, {sasl,"2.1.8"}]}, {system_arch,"x86_64-unknown-linux-gnu"}, {wall_clock,394}, {memory_data,{2058731520,672485376,{<0.639.0>,19371696}}}, {disk_data, [{"/",11204616,49},{"/boot",101086,14},{"/dev/shm",1005240,0}]}, {meminfo, <<"MemTotal: 2010480 kB\nMemFree: 1353880 kB\nBuffers: 61852 kB\nCached: 341748 kB\nSwapCached: 20084 kB\nActive: 347556 kB\nInactive: 196644 kB\nHighTotal: 0 kB\nHighFree: 0 kB\nLowTotal: 2010480 kB\nLowFree: 1353880 kB\nSwapTotal: 4030456 kB\nSwapFree: 3952288 kB\nDirty: 5592 kB\nWriteback: 0 kB\nAnonPages: 138424 kB\nMapped: 15732 kB\nSlab: 86440 kB\nPageTables: 6612 kB\nNFS_Unstable: 0 kB\nBounce: 0 kB\nCommitLimit: 5035696 kB\nCommitted_AS: 566488 kB\nVmallocTotal: 34359738367 kB\nVmallocUsed: 265484 kB\nVmallocChunk: 34359471011 kB\nHugePages_Total: 0\nHugePages_Free: 0\nHugePages_Rsvd: 0\nHugepagesize: 2048 kB\n">>}, {replication,[{"default",1.0}]}, {system_memory_data, [{system_total_memory,2058731520}, {free_swap,4047142912}, {total_swap,4127186944}, {cached_memory,349949952}, {buffered_memory,63336448}, {free_memory,1386373120}, {total_memory,2058731520}]}, {statistics, [{wall_clock,{390579,1}}, {context_switches,{176406,0}}, {garbage_collection,{33213,116861666,0}}, {io,{{input,53476586},{output,6913144}}}, {reductions,{41125821,737474}}, {run_queue,0}, {runtime,{5360,100}}]}]}, {'ns_1@10.1.5.229', [{last_heard,{1305,61002,242086}}, {active_buckets,["default"]}, {memory, [{total,32173344}, {processes,13223280}, {processes_used,13216240}, {system,18950064}, {atom,778121}, {atom_used,752794}, {binary,220488}, {code,7791241}, {ets,881008}]}, {cluster_compatibility_version,1}, {version, [{kernel,"2.13.4"}, {mnesia,"4.4.12"}, {stdlib,"1.16.4"}, {os_mon,"2.2.4"}, {ns_server,"1.6.5.4r"}, {menelaus,"1.6.5.4r"}, {sasl,"2.1.8"}]}, {system_arch,"x86_64-unknown-linux-gnu"}, {wall_clock,357}, {memory_data,{2058731520,704192512,{<9352.12.0>,4114680}}}, {disk_data, [{"/",11204616,44},{"/boot",101086,14},{"/dev/shm",1005240,0}]}, {meminfo, <<"MemTotal: 2010480 kB\nMemFree: 1321428 kB\nBuffers: 77036 kB\nCached: 339128 kB\nSwapCached: 0 kB\nActive: 360148 kB\nInactive: 211436 kB\nHighTotal: 0 kB\nHighFree: 0 kB\nLowTotal: 2010480 kB\nLowFree: 1321428 kB\nSwapTotal: 4030456 kB\nSwapFree: 4030388 kB\nDirty: 9252 kB\nWriteback: 0 kB\nAnonPages: 155404 kB\nMapped: 28104 kB\nSlab: 91984 kB\nPageTables: 5860 kB\nNFS_Unstable: 0 kB\nBounce: 0 kB\nCommitLimit: 5035696 kB\nCommitted_AS: 480132 kB\nVmallocTotal: 34359738367 kB\nVmallocUsed: 265484 kB\nVmallocChunk: 34359471011 kB\nHugePages_Total: 0\nHugePages_Free: 0\nHugePages_Rsvd: 0\nHugepagesize: 2048 kB\n">>}, {replication,[{"default",1.0}]}, {system_memory_data, [{system_total_memory,2058731520}, {free_swap,4127117312}, {total_swap,4127186944}, {cached_memory,347267072}, {buffered_memory,78884864}, {free_memory,1353142272}, {total_memory,2058731520}]}, {statistics, [{wall_clock,{349806,1}}, {context_switches,{35226,0}}, {garbage_collection,{7205,24174331,0}}, {io,{{input,4976765},{output,1383422}}}, {reductions,{8926329,231682}}, {run_queue,0}, {runtime,{1380,40}}]}]}] INFO REPORT <0.3675.0> 2011-05-10 13:56:46 =============================================================================== vbucketmigrator<0.3675.0>: Validate bucket states vbucketmigrator<0.3675.0>: 30 ok INFO REPORT <0.3692.0> 2011-05-10 13:56:47 =============================================================================== vbucketmigrator<0.3692.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.3692.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.3692.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.3692.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.3692.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.3692.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.3692.0>: Starting to move bucket 31 vbucketmigrator<0.3692.0>: Bucket 31 moved to the next server INFO REPORT <0.104.0> 2011-05-10 13:56:48 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.109.0> 2011-05-10 13:56:48 =============================================================================== ns_1@10.1.5.227:ns_orchestrator:164: Skipping janitor in state rebalancing: {rebalancing_state, <0.577.0>, {dict, 2, 16, 16, 8, 80, 48, {[], [], [], [], [], [], [], [], [], [], [], [], [], [], [], []}, {{[], [], [], [], [], [], [], [['ns_1@10.1.5.227'| 0.060546875]], [], [['ns_1@10.1.5.229'| 0.060546875]], [], [], [], [], [], []}}}} INFO REPORT <0.3692.0> 2011-05-10 13:56:53 =============================================================================== vbucketmigrator<0.3692.0>: Validate bucket states vbucketmigrator<0.3692.0>: 31 ok INFO REPORT <0.3705.0> 2011-05-10 13:56:54 =============================================================================== vbucketmigrator<0.3705.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.3705.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.3705.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.3705.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.3705.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.3705.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.3705.0>: Starting to move bucket 32 vbucketmigrator<0.3705.0>: Bucket 32 moved to the next server INFO REPORT <0.104.0> 2011-05-10 13:56:54 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.109.0> 2011-05-10 13:56:58 =============================================================================== ns_1@10.1.5.227:ns_orchestrator:164: Skipping janitor in state rebalancing: {rebalancing_state, <0.577.0>, {dict, 2, 16, 16, 8, 80, 48, {[], [], [], [], [], [], [], [], [], [], [], [], [], [], [], []}, {{[], [], [], [], [], [], [], [['ns_1@10.1.5.227'| 0.0625]], [], [['ns_1@10.1.5.229'| 0.0625]], [], [], [], [], [], []}}}} INFO REPORT <0.3705.0> 2011-05-10 13:56:59 =============================================================================== vbucketmigrator<0.3705.0>: Validate bucket states vbucketmigrator<0.3705.0>: 32 ok INFO REPORT <0.3712.0> 2011-05-10 13:57:00 =============================================================================== vbucketmigrator<0.3712.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.3712.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.3712.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.3712.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.3712.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.3712.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.3712.0>: Starting to move bucket 33 vbucketmigrator<0.3712.0>: Bucket 33 moved to the next server INFO REPORT <0.104.0> 2011-05-10 13:57:00 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.85.0> 2011-05-10 13:57:00 =============================================================================== Pulling config from: 'ns_1@10.1.5.229' INFO REPORT <0.469.0> 2011-05-10 13:57:04 =============================================================================== ns_1@10.1.5.227:stats_collector:71: Stats for bucket "default": auth_cmds 66 auth_errors 0 bucket_conns 34 bytes_read 1586691 bytes_written 2705649 cas_badval 0 cas_hits 0 cas_misses 0 cmd_flush 1 cmd_get 0 cmd_set 16004 conn_yields 0 connection_structures 44 curr_connections 44 curr_items 7763 curr_items_tot 8040 daemon_connections 10 decr_hits 0 decr_misses 0 delete_hits 0 delete_misses 0 ep_bg_fetched 0 ep_commit_num 53 ep_commit_time 0 ep_commit_time_total 4 ep_data_age 13 ep_data_age_highwat 13 ep_db_cleaner_status complete ep_db_strategy multiMTVBDB ep_dbinit 1 ep_dbname /var/opt/membase/1.6.5.4r/data/ns_1/default ep_dbshards 4 ep_expired 0 ep_flush_duration 1 ep_flush_duration_highwat 8 ep_flush_duration_total 12 ep_flush_preempts 0 ep_flusher_state running ep_flusher_todo 0 ep_io_num_read 277 ep_io_num_write 15741 ep_io_read_bytes 15771 ep_io_write_bytes 1011432 ep_item_begin_failed 0 ep_item_commit_failed 0 ep_item_flush_expired 0 ep_item_flush_failed 0 ep_kv_size 1181400 ep_max_data_size 1096810496 ep_max_txn_size 1000 ep_mem_high_wat 822607872 ep_mem_low_wat 658086297 ep_min_data_age 0 ep_num_active_non_resident 0 ep_num_eject_failures 0 ep_num_eject_replicas 0 ep_num_expiry_pager_runs 0 ep_num_non_resident 0 ep_num_not_my_vbuckets 0 ep_num_pager_runs 0 ep_num_value_ejects 0 ep_oom_errors 0 ep_overhead 25854448 ep_pending_ops 0 ep_pending_ops_max 0 ep_pending_ops_max_duration 0 ep_pending_ops_total 0 ep_queue_age_cap 900 ep_queue_size 0 ep_storage_age 0 ep_storage_age_highwat 10 ep_storage_type featured ep_store_max_concurrency 10 ep_store_max_readers 9 ep_store_max_readwrite 1 ep_tap_bg_fetch_requeued 0 ep_tap_bg_fetched 277 ep_tap_keepalive 300 ep_tmp_oom_errors 0 ep_too_old 0 ep_too_young 0 ep_total_cache_size 1181400 ep_total_del_items 0 ep_total_enqueued 16035 ep_total_new_items 15741 ep_total_persisted 15741 ep_vbucket_del 0 ep_vbucket_del_fail 0 ep_version 1.6.5.3.1_13_g81a6cf1 ep_warmed_up 0 ep_warmup true ep_warmup_dups 0 ep_warmup_oom 0 ep_warmup_thread complete ep_warmup_time 15662 get_hits 0 get_misses 0 incr_hits 0 incr_misses 0 libevent 1.4.13-stable limit_maxbytes 67108864 mem_used 27035848 pid 16183 pointer_size 64 rejected_conns 0 rusage_system 1.177820 rusage_user 2.698589 tap_connect_received 34 tap_mutation_sent 554 tap_opaque_sent 34 tap_vbucket_set_sent 68 threads 4 time 1305061023 total_connections 77 uptime 417 version 1.4.4_382_g9df3289 INFO REPORT <0.3712.0> 2011-05-10 13:57:04 =============================================================================== vbucketmigrator<0.3712.0>: Validate bucket states vbucketmigrator<0.3712.0>: 33 ok INFO REPORT <0.3716.0> 2011-05-10 13:57:06 =============================================================================== vbucketmigrator<0.3716.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.3716.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.3716.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.3716.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.3716.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.3716.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.3716.0>: Starting to move bucket 34 vbucketmigrator<0.3716.0>: Bucket 34 moved to the next server INFO REPORT <0.104.0> 2011-05-10 13:57:06 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.109.0> 2011-05-10 13:57:08 =============================================================================== ns_1@10.1.5.227:ns_orchestrator:164: Skipping janitor in state rebalancing: {rebalancing_state, <0.577.0>, {dict, 2, 16, 16, 8, 80, 48, {[], [], [], [], [], [], [], [], [], [], [], [], [], [], [], []}, {{[], [], [], [], [], [], [], [['ns_1@10.1.5.227'| 0.06640625]], [], [['ns_1@10.1.5.229'| 0.06640625]], [], [], [], [], [], []}}}} INFO REPORT <0.3716.0> 2011-05-10 13:57:11 =============================================================================== vbucketmigrator<0.3716.0>: Validate bucket states vbucketmigrator<0.3716.0>: 34 ok INFO REPORT <0.3730.0> 2011-05-10 13:57:12 =============================================================================== vbucketmigrator<0.3730.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.3730.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.3730.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.3730.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.3730.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.3730.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.3730.0>: Starting to move bucket 35 vbucketmigrator<0.3730.0>: Bucket 35 moved to the next server INFO REPORT <0.104.0> 2011-05-10 13:57:12 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.3730.0> 2011-05-10 13:57:17 =============================================================================== vbucketmigrator<0.3730.0>: Validate bucket states vbucketmigrator<0.3730.0>: 35 ok INFO REPORT <0.3734.0> 2011-05-10 13:57:18 =============================================================================== vbucketmigrator<0.3734.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.3734.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.3734.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.3734.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.3734.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.3734.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.3734.0>: Starting to move bucket 36 vbucketmigrator<0.3734.0>: Bucket 36 moved to the next server INFO REPORT <0.104.0> 2011-05-10 13:57:18 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.109.0> 2011-05-10 13:57:18 =============================================================================== ns_1@10.1.5.227:ns_orchestrator:164: Skipping janitor in state rebalancing: {rebalancing_state, <0.577.0>, {dict, 2, 16, 16, 8, 80, 48, {[], [], [], [], [], [], [], [], [], [], [], [], [], [], [], []}, {{[], [], [], [], [], [], [], [['ns_1@10.1.5.227'| 0.0703125]], [], [['ns_1@10.1.5.229'| 0.0703125]], [], [], [], [], [], []}}}} INFO REPORT <0.3734.0> 2011-05-10 13:57:23 =============================================================================== vbucketmigrator<0.3734.0>: Validate bucket states vbucketmigrator<0.3734.0>: 36 ok INFO REPORT <0.3739.0> 2011-05-10 13:57:24 =============================================================================== vbucketmigrator<0.3739.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.3739.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.3739.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.3739.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.3739.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.3739.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.3739.0>: Starting to move bucket 37 vbucketmigrator<0.3739.0>: Bucket 37 moved to the next server INFO REPORT <0.104.0> 2011-05-10 13:57:24 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.109.0> 2011-05-10 13:57:28 =============================================================================== ns_1@10.1.5.227:ns_orchestrator:164: Skipping janitor in state rebalancing: {rebalancing_state, <0.577.0>, {dict, 2, 16, 16, 8, 80, 48, {[], [], [], [], [], [], [], [], [], [], [], [], [], [], [], []}, {{[], [], [], [], [], [], [], [['ns_1@10.1.5.227'| 0.072265625]], [], [['ns_1@10.1.5.229'| 0.072265625]], [], [], [], [], [], []}}}} INFO REPORT <0.3739.0> 2011-05-10 13:57:29 =============================================================================== vbucketmigrator<0.3739.0>: Validate bucket states vbucketmigrator<0.3739.0>: 37 ok INFO REPORT <0.3754.0> 2011-05-10 13:57:30 =============================================================================== vbucketmigrator<0.3754.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.3754.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.3754.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.3754.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.3754.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.3754.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.3754.0>: Starting to move bucket 38 vbucketmigrator<0.3754.0>: Bucket 38 moved to the next server INFO REPORT <0.104.0> 2011-05-10 13:57:30 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.85.0> 2011-05-10 13:57:30 =============================================================================== Pulling config from: 'ns_1@10.1.5.229' INFO REPORT <0.3754.0> 2011-05-10 13:57:35 =============================================================================== vbucketmigrator<0.3754.0>: Validate bucket states vbucketmigrator<0.3754.0>: 38 ok INFO REPORT <0.3758.0> 2011-05-10 13:57:36 =============================================================================== vbucketmigrator<0.3758.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.3758.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.3758.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.3758.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.3758.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.3758.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.3758.0>: Starting to move bucket 39 vbucketmigrator<0.3758.0>: Bucket 39 moved to the next server INFO REPORT <0.104.0> 2011-05-10 13:57:36 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.85.0> 2011-05-10 13:57:36 =============================================================================== Pulling config from: 'ns_1@10.1.5.229' INFO REPORT <0.109.0> 2011-05-10 13:57:38 =============================================================================== ns_1@10.1.5.227:ns_orchestrator:164: Skipping janitor in state rebalancing: {rebalancing_state, <0.577.0>, {dict, 2, 16, 16, 8, 80, 48, {[], [], [], [], [], [], [], [], [], [], [], [], [], [], [], []}, {{[], [], [], [], [], [], [], [['ns_1@10.1.5.227'| 0.076171875]], [], [['ns_1@10.1.5.229'| 0.076171875]], [], [], [], [], [], []}}}} INFO REPORT <0.3758.0> 2011-05-10 13:57:40 =============================================================================== vbucketmigrator<0.3758.0>: Validate bucket states vbucketmigrator<0.3758.0>: 39 ok INFO REPORT <0.3763.0> 2011-05-10 13:57:41 =============================================================================== vbucketmigrator<0.3763.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.3763.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.3763.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.3763.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.3763.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.3763.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.3763.0>: Starting to move bucket 40 vbucketmigrator<0.3763.0>: Bucket 40 moved to the next server INFO REPORT <0.104.0> 2011-05-10 13:57:41 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.474.0> 2011-05-10 13:57:43 =============================================================================== ns_1@10.1.5.227:ns_doctor:82: Current node statuses: [{'ns_1@10.1.5.227', [{last_heard,{1305,61062,411210}}, {active_buckets,["default"]}, {memory, [{total,59118512}, {processes,39086728}, {processes_used,39074952}, {system,20031784}, {atom,819441}, {atom_used,806779}, {binary,136592}, {code,8552890}, {ets,1017072}]}, {cluster_compatibility_version,1}, {version, [{kernel,"2.13.4"}, {mnesia,"4.4.12"}, {stdlib,"1.16.4"}, {os_mon,"2.2.4"}, {inets,"5.2"}, {ns_server,"1.6.5.4r"}, {menelaus,"1.6.5.4r"}, {sasl,"2.1.8"}]}, {system_arch,"x86_64-unknown-linux-gnu"}, {wall_clock,454}, {memory_data,{2058731520,672231424,{<0.639.0>,19371696}}}, {disk_data, [{"/",11204616,49},{"/boot",101086,14},{"/dev/shm",1005240,0}]}, {meminfo, <<"MemTotal: 2010480 kB\nMemFree: 1353632 kB\nBuffers: 62148 kB\nCached: 341996 kB\nSwapCached: 20084 kB\nActive: 347120 kB\nInactive: 196892 kB\nHighTotal: 0 kB\nHighFree: 0 kB\nLowTotal: 2010480 kB\nLowFree: 1353632 kB\nSwapTotal: 4030456 kB\nSwapFree: 3952288 kB\nDirty: 280 kB\nWriteback: 0 kB\nAnonPages: 137652 kB\nMapped: 15732 kB\nSlab: 86412 kB\nPageTables: 6612 kB\nNFS_Unstable: 0 kB\nBounce: 0 kB\nCommitLimit: 5035696 kB\nCommitted_AS: 568972 kB\nVmallocTotal: 34359738367 kB\nVmallocUsed: 265484 kB\nVmallocChunk: 34359471011 kB\nHugePages_Total: 0\nHugePages_Free: 0\nHugePages_Rsvd: 0\nHugepagesize: 2048 kB\n">>}, {replication,[{"default",1.0}]}, {system_memory_data, [{system_total_memory,2058731520}, {free_swap,4047142912}, {total_swap,4127186944}, {cached_memory,350203904}, {buffered_memory,63639552}, {free_memory,1386119168}, {total_memory,2058731520}]}, {statistics, [{wall_clock,{450579,1}}, {context_switches,{186561,0}}, {garbage_collection,{35504,135140348,0}}, {io,{{input,54373194},{output,8027353}}}, {reductions,{45563127,697029}}, {run_queue,0}, {runtime,{6120,120}}]}]}, {'ns_1@10.1.5.229', [{last_heard,{1305,61062,242307}}, {active_buckets,["default"]}, {memory, [{total,32117312}, {processes,13105768}, {processes_used,13082248}, {system,19011544}, {atom,778121}, {atom_used,752794}, {binary,217120}, {code,7791241}, {ets,944424}]}, {cluster_compatibility_version,1}, {version, [{kernel,"2.13.4"}, {mnesia,"4.4.12"}, {stdlib,"1.16.4"}, {os_mon,"2.2.4"}, {ns_server,"1.6.5.4r"}, {menelaus,"1.6.5.4r"}, {sasl,"2.1.8"}]}, {system_arch,"x86_64-unknown-linux-gnu"}, {wall_clock,417}, {memory_data,{2058731520,706224128,{<9352.12.0>,4114680}}}, {disk_data, [{"/",11204616,44},{"/boot",101086,14},{"/dev/shm",1005240,0}]}, {meminfo, <<"MemTotal: 2010480 kB\nMemFree: 1319940 kB\nBuffers: 77788 kB\nCached: 339548 kB\nSwapCached: 0 kB\nActive: 360924 kB\nInactive: 212156 kB\nHighTotal: 0 kB\nHighFree: 0 kB\nLowTotal: 2010480 kB\nLowFree: 1319940 kB\nSwapTotal: 4030456 kB\nSwapFree: 4030388 kB\nDirty: 1016 kB\nWriteback: 0 kB\nAnonPages: 155744 kB\nMapped: 28104 kB\nSlab: 91952 kB\nPageTables: 5860 kB\nNFS_Unstable: 0 kB\nBounce: 0 kB\nCommitLimit: 5035696 kB\nCommitted_AS: 479648 kB\nVmallocTotal: 34359738367 kB\nVmallocUsed: 265484 kB\nVmallocChunk: 34359471011 kB\nHugePages_Total: 0\nHugePages_Free: 0\nHugePages_Rsvd: 0\nHugepagesize: 2048 kB\n">>}, {replication,[{"default",1.0}]}, {system_memory_data, [{system_total_memory,2058731520}, {free_swap,4127117312}, {total_swap,4127186944}, {cached_memory,347697152}, {buffered_memory,79654912}, {free_memory,1351618560}, {total_memory,2058731520}]}, {statistics, [{wall_clock,{409806,1}}, {context_switches,{40409,0}}, {garbage_collection,{8558,28784916,0}}, {io,{{input,5911840},{output,2316635}}}, {reductions,{10354836,236352}}, {run_queue,0}, {runtime,{1700,50}}]}]}] INFO REPORT <0.3763.0> 2011-05-10 13:57:46 =============================================================================== vbucketmigrator<0.3763.0>: Validate bucket states vbucketmigrator<0.3763.0>: 40 ok INFO REPORT <0.3767.0> 2011-05-10 13:57:46 =============================================================================== vbucketmigrator<0.3767.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.3767.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.3767.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.3767.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.3767.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.3767.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.3767.0>: Starting to move bucket 41 vbucketmigrator<0.3767.0>: Bucket 41 moved to the next server vbucketmigrator<0.3767.0>: Validate bucket states vbucketmigrator<0.3767.0>: 41 ok INFO REPORT <0.104.0> 2011-05-10 13:57:47 =============================================================================== memcached<0.104.0>: Vbucket is going dead. memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.3769.0> 2011-05-10 13:57:47 =============================================================================== vbucketmigrator<0.3769.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.3769.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.3769.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.3769.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.3769.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.3769.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.3769.0>: Starting to move bucket 42 vbucketmigrator<0.3769.0>: Bucket 42 moved to the next server INFO REPORT <0.109.0> 2011-05-10 13:57:48 =============================================================================== ns_1@10.1.5.227:ns_orchestrator:164: Skipping janitor in state rebalancing: {rebalancing_state, <0.577.0>, {dict, 2, 16, 16, 8, 80, 48, {[], [], [], [], [], [], [], [], [], [], [], [], [], [], [], []}, {{[], [], [], [], [], [], [], [['ns_1@10.1.5.227'| 0.08203125]], [], [['ns_1@10.1.5.229'| 0.08203125]], [], [], [], [], [], []}}}} INFO REPORT <0.3769.0> 2011-05-10 13:57:52 =============================================================================== vbucketmigrator<0.3769.0>: Validate bucket states vbucketmigrator<0.3769.0>: 42 ok INFO REPORT <0.3782.0> 2011-05-10 13:57:53 =============================================================================== vbucketmigrator<0.3782.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.3782.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.3782.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.3782.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.3782.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.3782.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.3782.0>: Starting to move bucket 43 vbucketmigrator<0.3782.0>: Bucket 43 moved to the next server INFO REPORT <0.104.0> 2011-05-10 13:57:53 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.3782.0> 2011-05-10 13:57:58 =============================================================================== vbucketmigrator<0.3782.0>: Validate bucket states vbucketmigrator<0.3782.0>: 43 ok INFO REPORT <0.109.0> 2011-05-10 13:57:58 =============================================================================== ns_1@10.1.5.227:ns_orchestrator:164: Skipping janitor in state rebalancing: {rebalancing_state, <0.577.0>, {dict, 2, 16, 16, 8, 80, 48, {[], [], [], [], [], [], [], [], [], [], [], [], [], [], [], []}, {{[], [], [], [], [], [], [], [['ns_1@10.1.5.227'| 0.0859375]], [], [['ns_1@10.1.5.229'| 0.0859375]], [], [], [], [], [], []}}}} INFO REPORT <0.3786.0> 2011-05-10 13:57:59 =============================================================================== vbucketmigrator<0.3786.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.3786.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.3786.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.3786.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.3786.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.3786.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.3786.0>: Starting to move bucket 44 vbucketmigrator<0.3786.0>: Bucket 44 moved to the next server INFO REPORT <0.104.0> 2011-05-10 13:57:59 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.3786.0> 2011-05-10 13:58:04 =============================================================================== vbucketmigrator<0.3786.0>: Validate bucket states vbucketmigrator<0.3786.0>: 44 ok INFO REPORT <0.3793.0> 2011-05-10 13:58:04 =============================================================================== vbucketmigrator<0.3793.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.3793.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.3793.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.3793.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.3793.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.3793.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.3793.0>: Starting to move bucket 45 vbucketmigrator<0.3793.0>: Bucket 45 moved to the next server vbucketmigrator<0.3793.0>: Validate bucket states vbucketmigrator<0.3793.0>: 45 ok INFO REPORT <0.104.0> 2011-05-10 13:58:05 =============================================================================== memcached<0.104.0>: Vbucket is going dead. memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.3795.0> 2011-05-10 13:58:05 =============================================================================== vbucketmigrator<0.3795.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.3795.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.3795.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.3795.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.3795.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.3795.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.3795.0>: Starting to move bucket 46 vbucketmigrator<0.3795.0>: Bucket 46 moved to the next server INFO REPORT <0.85.0> 2011-05-10 13:58:07 =============================================================================== Pulling config from: 'ns_1@10.1.5.229' INFO REPORT <0.109.0> 2011-05-10 13:58:08 =============================================================================== ns_1@10.1.5.227:ns_orchestrator:164: Skipping janitor in state rebalancing: {rebalancing_state, <0.577.0>, {dict, 2, 16, 16, 8, 80, 48, {[], [], [], [], [], [], [], [], [], [], [], [], [], [], [], []}, {{[], [], [], [], [], [], [], [['ns_1@10.1.5.227'| 0.08984375]], [], [['ns_1@10.1.5.229'| 0.08984375]], [], [], [], [], [], []}}}} INFO REPORT <0.3795.0> 2011-05-10 13:58:10 =============================================================================== vbucketmigrator<0.3795.0>: Validate bucket states vbucketmigrator<0.3795.0>: 46 ok INFO REPORT <0.3809.0> 2011-05-10 13:58:11 =============================================================================== vbucketmigrator<0.3809.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.3809.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.3809.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.3809.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.3809.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.3809.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.3809.0>: Starting to move bucket 47 vbucketmigrator<0.3809.0>: Bucket 47 moved to the next server INFO REPORT <0.104.0> 2011-05-10 13:58:11 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.3809.0> 2011-05-10 13:58:16 =============================================================================== vbucketmigrator<0.3809.0>: Validate bucket states vbucketmigrator<0.3809.0>: 47 ok INFO REPORT <0.3813.0> 2011-05-10 13:58:17 =============================================================================== vbucketmigrator<0.3813.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.3813.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.3813.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.3813.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.3813.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.3813.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.3813.0>: Starting to move bucket 48 vbucketmigrator<0.3813.0>: Bucket 48 moved to the next server INFO REPORT <0.104.0> 2011-05-10 13:58:17 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.109.0> 2011-05-10 13:58:18 =============================================================================== ns_1@10.1.5.227:ns_orchestrator:164: Skipping janitor in state rebalancing: {rebalancing_state, <0.577.0>, {dict, 2, 16, 16, 8, 80, 48, {[], [], [], [], [], [], [], [], [], [], [], [], [], [], [], []}, {{[], [], [], [], [], [], [], [['ns_1@10.1.5.227'| 0.09375]], [], [['ns_1@10.1.5.229'| 0.09375]], [], [], [], [], [], []}}}} INFO REPORT <0.3813.0> 2011-05-10 13:58:22 =============================================================================== vbucketmigrator<0.3813.0>: Validate bucket states vbucketmigrator<0.3813.0>: 48 ok INFO REPORT <0.3818.0> 2011-05-10 13:58:23 =============================================================================== vbucketmigrator<0.3818.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.3818.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.3818.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.3818.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.3818.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.3818.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.3818.0>: Starting to move bucket 49 vbucketmigrator<0.3818.0>: Bucket 49 moved to the next server INFO REPORT <0.104.0> 2011-05-10 13:58:23 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.3818.0> 2011-05-10 13:58:28 =============================================================================== vbucketmigrator<0.3818.0>: Validate bucket states vbucketmigrator<0.3818.0>: 49 ok INFO REPORT <0.109.0> 2011-05-10 13:58:28 =============================================================================== ns_1@10.1.5.227:ns_orchestrator:164: Skipping janitor in state rebalancing: {rebalancing_state, <0.577.0>, {dict, 2, 16, 16, 8, 80, 48, {[], [], [], [], [], [], [], [], [], [], [], [], [], [], [], []}, {{[], [], [], [], [], [], [], [['ns_1@10.1.5.227'| 0.09765625]], [], [['ns_1@10.1.5.229'| 0.09765625]], [], [], [], [], [], []}}}} INFO REPORT <0.3822.0> 2011-05-10 13:58:29 =============================================================================== vbucketmigrator<0.3822.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.3822.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.3822.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.3822.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.3822.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.3822.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.3822.0>: Starting to move bucket 50 vbucketmigrator<0.3822.0>: Bucket 50 moved to the next server INFO REPORT <0.104.0> 2011-05-10 13:58:29 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.85.0> 2011-05-10 13:58:29 =============================================================================== Pulling config from: 'ns_1@10.1.5.229' INFO REPORT <0.3822.0> 2011-05-10 13:58:34 =============================================================================== vbucketmigrator<0.3822.0>: Validate bucket states vbucketmigrator<0.3822.0>: 50 ok INFO REPORT <0.3837.0> 2011-05-10 13:58:35 =============================================================================== vbucketmigrator<0.3837.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.3837.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.3837.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.3837.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.3837.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.3837.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.3837.0>: Starting to move bucket 51 vbucketmigrator<0.3837.0>: Bucket 51 moved to the next server INFO REPORT <0.104.0> 2011-05-10 13:58:35 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.109.0> 2011-05-10 13:58:38 =============================================================================== ns_1@10.1.5.227:ns_orchestrator:164: Skipping janitor in state rebalancing: {rebalancing_state, <0.577.0>, {dict, 2, 16, 16, 8, 80, 48, {[], [], [], [], [], [], [], [], [], [], [], [], [], [], [], []}, {{[], [], [], [], [], [], [], [['ns_1@10.1.5.227'| 0.099609375]], [], [['ns_1@10.1.5.229'| 0.099609375]], [], [], [], [], [], []}}}} INFO REPORT <0.3837.0> 2011-05-10 13:58:40 =============================================================================== vbucketmigrator<0.3837.0>: Validate bucket states vbucketmigrator<0.3837.0>: 51 ok INFO REPORT <0.3842.0> 2011-05-10 13:58:41 =============================================================================== vbucketmigrator<0.3842.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.3842.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.3842.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.3842.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.3842.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.3842.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.3842.0>: Starting to move bucket 52 vbucketmigrator<0.3842.0>: Bucket 52 moved to the next server INFO REPORT <0.104.0> 2011-05-10 13:58:41 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.474.0> 2011-05-10 13:58:43 =============================================================================== ns_1@10.1.5.227:ns_doctor:82: Current node statuses: [{'ns_1@10.1.5.227', [{last_heard,{1305,61122,411245}}, {active_buckets,["default"]}, {memory, [{total,59200296}, {processes,39111552}, {processes_used,39099776}, {system,20088744}, {atom,819441}, {atom_used,806779}, {binary,132288}, {code,8552890}, {ets,1077248}]}, {cluster_compatibility_version,1}, {version, [{kernel,"2.13.4"}, {mnesia,"4.4.12"}, {stdlib,"1.16.4"}, {os_mon,"2.2.4"}, {inets,"5.2"}, {ns_server,"1.6.5.4r"}, {menelaus,"1.6.5.4r"}, {sasl,"2.1.8"}]}, {system_arch,"x86_64-unknown-linux-gnu"}, {wall_clock,514}, {memory_data,{2058731520,673882112,{<0.639.0>,19371696}}}, {disk_data, [{"/",11204616,49},{"/boot",101086,14},{"/dev/shm",1005240,0}]}, {meminfo, <<"MemTotal: 2010480 kB\nMemFree: 1352392 kB\nBuffers: 62468 kB\nCached: 342260 kB\nSwapCached: 20084 kB\nActive: 348600 kB\nInactive: 197208 kB\nHighTotal: 0 kB\nHighFree: 0 kB\nLowTotal: 2010480 kB\nLowFree: 1352392 kB\nSwapTotal: 4030456 kB\nSwapFree: 3952288 kB\nDirty: 236 kB\nWriteback: 0 kB\nAnonPages: 138876 kB\nMapped: 15732 kB\nSlab: 86428 kB\nPageTables: 6608 kB\nNFS_Unstable: 0 kB\nBounce: 0 kB\nCommitLimit: 5035696 kB\nCommitted_AS: 566488 kB\nVmallocTotal: 34359738367 kB\nVmallocUsed: 265484 kB\nVmallocChunk: 34359471011 kB\nHugePages_Total: 0\nHugePages_Free: 0\nHugePages_Rsvd: 0\nHugepagesize: 2048 kB\n">>}, {replication,[{"default",1.0}]}, {system_memory_data, [{system_total_memory,2058731520}, {free_swap,4047142912}, {total_swap,4127186944}, {cached_memory,350474240}, {buffered_memory,63967232}, {free_memory,1384849408}, {total_memory,2058731520}]}, {statistics, [{wall_clock,{510579,1}}, {context_switches,{196973,0}}, {garbage_collection,{37612,154744055,0}}, {io,{{input,54518899},{output,8450496}}}, {reductions,{50093777,709941}}, {run_queue,0}, {runtime,{6880,110}}]}]}, {'ns_1@10.1.5.229', [{last_heard,{1305,61122,242156}}, {active_buckets,["default"]}, {memory, [{total,32356952}, {processes,13268360}, {processes_used,13244840}, {system,19088592}, {atom,778121}, {atom_used,752794}, {binary,233664}, {code,7791241}, {ets,1004584}]}, {cluster_compatibility_version,1}, {version, [{kernel,"2.13.4"}, {mnesia,"4.4.12"}, {stdlib,"1.16.4"}, {os_mon,"2.2.4"}, {ns_server,"1.6.5.4r"}, {menelaus,"1.6.5.4r"}, {sasl,"2.1.8"}]}, {system_arch,"x86_64-unknown-linux-gnu"}, {wall_clock,477}, {memory_data,{2058731520,706985984,{<9352.12.0>,4114680}}}, {disk_data, [{"/",11204616,44},{"/boot",101086,14},{"/dev/shm",1005240,0}]}, {meminfo, <<"MemTotal: 2010480 kB\nMemFree: 1318576 kB\nBuffers: 78700 kB\nCached: 340008 kB\nSwapCached: 0 kB\nActive: 361680 kB\nInactive: 212852 kB\nHighTotal: 0 kB\nHighFree: 0 kB\nLowTotal: 2010480 kB\nLowFree: 1318576 kB\nSwapTotal: 4030456 kB\nSwapFree: 4030388 kB\nDirty: 172 kB\nWriteback: 0 kB\nAnonPages: 155852 kB\nMapped: 28104 kB\nSlab: 91904 kB\nPageTables: 5860 kB\nNFS_Unstable: 0 kB\nBounce: 0 kB\nCommitLimit: 5035696 kB\nCommitted_AS: 475856 kB\nVmallocTotal: 34359738367 kB\nVmallocUsed: 265484 kB\nVmallocChunk: 34359471011 kB\nHugePages_Total: 0\nHugePages_Free: 0\nHugePages_Rsvd: 0\nHugepagesize: 2048 kB\n">>}, {replication,[{"default",1.0}]}, {system_memory_data, [{system_total_memory,2058731520}, {free_swap,4127117312}, {total_swap,4127186944}, {cached_memory,348168192}, {buffered_memory,80588800}, {free_memory,1350221824}, {total_memory,2058731520}]}, {statistics, [{wall_clock,{469806,1}}, {context_switches,{45229,0}}, {garbage_collection,{9633,33248892,0}}, {io,{{input,6150213},{output,2583681}}}, {reductions,{11725770,232335}}, {run_queue,0}, {runtime,{1940,30}}]}]}] INFO REPORT <0.469.0> 2011-05-10 13:58:44 =============================================================================== ns_1@10.1.5.227:stats_collector:71: Stats for bucket "default": auth_cmds 85 auth_errors 0 bucket_conns 34 bytes_read 1593928 bytes_written 3748701 cas_badval 0 cas_hits 0 cas_misses 0 cmd_flush 1 cmd_get 0 cmd_set 16004 conn_yields 0 connection_structures 44 curr_connections 44 curr_items 7611 curr_items_tot 8040 daemon_connections 10 decr_hits 0 decr_misses 0 delete_hits 0 delete_misses 0 ep_bg_fetched 0 ep_commit_num 53 ep_commit_time 0 ep_commit_time_total 4 ep_data_age 13 ep_data_age_highwat 13 ep_db_cleaner_status complete ep_db_strategy multiMTVBDB ep_dbinit 1 ep_dbname /var/opt/membase/1.6.5.4r/data/ns_1/default ep_dbshards 4 ep_expired 0 ep_flush_duration 1 ep_flush_duration_highwat 8 ep_flush_duration_total 12 ep_flush_preempts 0 ep_flusher_state running ep_flusher_todo 0 ep_io_num_read 429 ep_io_num_write 15741 ep_io_read_bytes 24346 ep_io_write_bytes 1011432 ep_item_begin_failed 0 ep_item_commit_failed 0 ep_item_flush_expired 0 ep_item_flush_failed 0 ep_kv_size 1181400 ep_max_data_size 1096810496 ep_max_txn_size 1000 ep_mem_high_wat 822607872 ep_mem_low_wat 658086297 ep_min_data_age 0 ep_num_active_non_resident 0 ep_num_eject_failures 0 ep_num_eject_replicas 0 ep_num_expiry_pager_runs 0 ep_num_non_resident 0 ep_num_not_my_vbuckets 0 ep_num_pager_runs 0 ep_num_value_ejects 0 ep_oom_errors 0 ep_overhead 25854448 ep_pending_ops 0 ep_pending_ops_max 0 ep_pending_ops_max_duration 0 ep_pending_ops_total 0 ep_queue_age_cap 900 ep_queue_size 0 ep_storage_age 0 ep_storage_age_highwat 10 ep_storage_type featured ep_store_max_concurrency 10 ep_store_max_readers 9 ep_store_max_readwrite 1 ep_tap_bg_fetch_requeued 0 ep_tap_bg_fetched 429 ep_tap_keepalive 300 ep_tmp_oom_errors 0 ep_too_old 0 ep_too_young 0 ep_total_cache_size 1181400 ep_total_del_items 0 ep_total_enqueued 16035 ep_total_new_items 15741 ep_total_persisted 15741 ep_vbucket_del 0 ep_vbucket_del_fail 0 ep_version 1.6.5.3.1_13_g81a6cf1 ep_warmed_up 0 ep_warmup true ep_warmup_dups 0 ep_warmup_oom 0 ep_warmup_thread complete ep_warmup_time 15662 get_hits 0 get_misses 0 incr_hits 0 incr_misses 0 libevent 1.4.13-stable limit_maxbytes 67108864 mem_used 27035848 pid 16183 pointer_size 64 rejected_conns 0 rusage_system 1.212815 rusage_user 2.904558 tap_connect_received 53 tap_mutation_sent 858 tap_opaque_sent 53 tap_vbucket_set_sent 106 threads 4 time 1305061124 total_connections 96 uptime 518 version 1.4.4_382_g9df3289 INFO REPORT <0.3842.0> 2011-05-10 13:58:46 =============================================================================== vbucketmigrator<0.3842.0>: Validate bucket states vbucketmigrator<0.3842.0>: 52 ok INFO REPORT <0.3846.0> 2011-05-10 13:58:47 =============================================================================== vbucketmigrator<0.3846.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.3846.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.3846.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.3846.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.3846.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.3846.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.3846.0>: Starting to move bucket 53 vbucketmigrator<0.3846.0>: Bucket 53 moved to the next server INFO REPORT <0.104.0> 2011-05-10 13:58:47 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.109.0> 2011-05-10 13:58:48 =============================================================================== ns_1@10.1.5.227:ns_orchestrator:164: Skipping janitor in state rebalancing: {rebalancing_state, <0.577.0>, {dict, 2, 16, 16, 8, 80, 48, {[], [], [], [], [], [], [], [], [], [], [], [], [], [], [], []}, {{[], [], [], [], [], [], [], [['ns_1@10.1.5.227'| 0.103515625]], [], [['ns_1@10.1.5.229'| 0.103515625]], [], [], [], [], [], []}}}} INFO REPORT <0.3846.0> 2011-05-10 13:58:52 =============================================================================== vbucketmigrator<0.3846.0>: Validate bucket states vbucketmigrator<0.3846.0>: 53 ok INFO REPORT <0.3859.0> 2011-05-10 13:58:53 =============================================================================== vbucketmigrator<0.3859.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.3859.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.3859.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.3859.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.3859.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.3859.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.3859.0>: Starting to move bucket 54 vbucketmigrator<0.3859.0>: Bucket 54 moved to the next server INFO REPORT <0.104.0> 2011-05-10 13:58:53 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.3859.0> 2011-05-10 13:58:58 =============================================================================== vbucketmigrator<0.3859.0>: Validate bucket states vbucketmigrator<0.3859.0>: 54 ok INFO REPORT <0.109.0> 2011-05-10 13:58:58 =============================================================================== ns_1@10.1.5.227:ns_orchestrator:164: Skipping janitor in state rebalancing: {rebalancing_state, <0.577.0>, {dict, 2, 16, 16, 8, 80, 48, {[], [], [], [], [], [], [], [], [], [], [], [], [], [], [], []}, {{[], [], [], [], [], [], [], [['ns_1@10.1.5.227'| 0.107421875]], [], [['ns_1@10.1.5.229'| 0.107421875]], [], [], [], [], [], []}}}} INFO REPORT <0.3863.0> 2011-05-10 13:58:59 =============================================================================== vbucketmigrator<0.3863.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.3863.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.3863.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.3863.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.3863.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.3863.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.3863.0>: Starting to move bucket 55 vbucketmigrator<0.3863.0>: Bucket 55 moved to the next server INFO REPORT <0.104.0> 2011-05-10 13:58:59 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.3863.0> 2011-05-10 13:59:04 =============================================================================== vbucketmigrator<0.3863.0>: Validate bucket states vbucketmigrator<0.3863.0>: 55 ok INFO REPORT <0.3870.0> 2011-05-10 13:59:05 =============================================================================== vbucketmigrator<0.3870.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.3870.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.3870.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.3870.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.3870.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.3870.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.3870.0>: Starting to move bucket 56 vbucketmigrator<0.3870.0>: Bucket 56 moved to the next server INFO REPORT <0.104.0> 2011-05-10 13:59:05 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.85.0> 2011-05-10 13:59:06 =============================================================================== Pulling config from: 'ns_1@10.1.5.229' INFO REPORT <0.109.0> 2011-05-10 13:59:08 =============================================================================== ns_1@10.1.5.227:ns_orchestrator:164: Skipping janitor in state rebalancing: {rebalancing_state, <0.577.0>, {dict, 2, 16, 16, 8, 80, 48, {[], [], [], [], [], [], [], [], [], [], [], [], [], [], [], []}, {{[], [], [], [], [], [], [], [['ns_1@10.1.5.227'| 0.109375]], [], [['ns_1@10.1.5.229'| 0.109375]], [], [], [], [], [], []}}}} INFO REPORT <0.3870.0> 2011-05-10 13:59:10 =============================================================================== vbucketmigrator<0.3870.0>: Validate bucket states vbucketmigrator<0.3870.0>: 56 ok INFO REPORT <0.3884.0> 2011-05-10 13:59:11 =============================================================================== vbucketmigrator<0.3884.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.3884.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.3884.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.3884.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.3884.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.3884.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.3884.0>: Starting to move bucket 57 vbucketmigrator<0.3884.0>: Bucket 57 moved to the next server INFO REPORT <0.104.0> 2011-05-10 13:59:11 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.3884.0> 2011-05-10 13:59:16 =============================================================================== vbucketmigrator<0.3884.0>: Validate bucket states vbucketmigrator<0.3884.0>: 57 ok INFO REPORT <0.3888.0> 2011-05-10 13:59:17 =============================================================================== vbucketmigrator<0.3888.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.3888.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.3888.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.3888.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.3888.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.3888.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.3888.0>: Starting to move bucket 58 vbucketmigrator<0.3888.0>: Bucket 58 moved to the next server INFO REPORT <0.104.0> 2011-05-10 13:59:17 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.109.0> 2011-05-10 13:59:18 =============================================================================== ns_1@10.1.5.227:ns_orchestrator:164: Skipping janitor in state rebalancing: {rebalancing_state, <0.577.0>, {dict, 2, 16, 16, 8, 80, 48, {[], [], [], [], [], [], [], [], [], [], [], [], [], [], [], []}, {{[], [], [], [], [], [], [], [['ns_1@10.1.5.227'| 0.11328125]], [], [['ns_1@10.1.5.229'| 0.11328125]], [], [], [], [], [], []}}}} INFO REPORT <0.3888.0> 2011-05-10 13:59:22 =============================================================================== vbucketmigrator<0.3888.0>: Validate bucket states vbucketmigrator<0.3888.0>: 58 ok INFO REPORT <0.3893.0> 2011-05-10 13:59:23 =============================================================================== vbucketmigrator<0.3893.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.3893.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.3893.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.3893.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.3893.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.3893.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.3893.0>: Starting to move bucket 59 vbucketmigrator<0.3893.0>: Bucket 59 moved to the next server INFO REPORT <0.104.0> 2011-05-10 13:59:23 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.3893.0> 2011-05-10 13:59:28 =============================================================================== vbucketmigrator<0.3893.0>: Validate bucket states vbucketmigrator<0.3893.0>: 59 ok INFO REPORT <0.109.0> 2011-05-10 13:59:28 =============================================================================== ns_1@10.1.5.227:ns_orchestrator:164: Skipping janitor in state rebalancing: {rebalancing_state, <0.577.0>, {dict, 2, 16, 16, 8, 80, 48, {[], [], [], [], [], [], [], [], [], [], [], [], [], [], [], []}, {{[], [], [], [], [], [], [], [['ns_1@10.1.5.227'| 0.1171875]], [], [['ns_1@10.1.5.229'| 0.1171875]], [], [], [], [], [], []}}}} INFO REPORT <0.3897.0> 2011-05-10 13:59:29 =============================================================================== vbucketmigrator<0.3897.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.3897.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.3897.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.3897.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.3897.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.3897.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.3897.0>: Starting to move bucket 60 vbucketmigrator<0.3897.0>: Bucket 60 moved to the next server INFO REPORT <0.104.0> 2011-05-10 13:59:29 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.85.0> 2011-05-10 13:59:29 =============================================================================== Pulling config from: 'ns_1@10.1.5.229' INFO REPORT <0.3897.0> 2011-05-10 13:59:34 =============================================================================== vbucketmigrator<0.3897.0>: Validate bucket states vbucketmigrator<0.3897.0>: 60 ok INFO REPORT <0.3912.0> 2011-05-10 13:59:35 =============================================================================== vbucketmigrator<0.3912.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.3912.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.3912.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.3912.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.3912.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.3912.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.3912.0>: Starting to move bucket 61 vbucketmigrator<0.3912.0>: Bucket 61 moved to the next server INFO REPORT <0.104.0> 2011-05-10 13:59:35 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.109.0> 2011-05-10 13:59:38 =============================================================================== ns_1@10.1.5.227:ns_orchestrator:164: Skipping janitor in state rebalancing: {rebalancing_state, <0.577.0>, {dict, 2, 16, 16, 8, 80, 48, {[], [], [], [], [], [], [], [], [], [], [], [], [], [], [], []}, {{[], [], [], [], [], [], [], [['ns_1@10.1.5.227'| 0.119140625]], [], [['ns_1@10.1.5.229'| 0.119140625]], [], [], [], [], [], []}}}} INFO REPORT <0.3912.0> 2011-05-10 13:59:40 =============================================================================== vbucketmigrator<0.3912.0>: Validate bucket states vbucketmigrator<0.3912.0>: 61 ok INFO REPORT <0.3917.0> 2011-05-10 13:59:41 =============================================================================== vbucketmigrator<0.3917.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.3917.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.3917.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.3917.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.3917.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.3917.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.3917.0>: Starting to move bucket 62 vbucketmigrator<0.3917.0>: Bucket 62 moved to the next server INFO REPORT <0.104.0> 2011-05-10 13:59:41 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.474.0> 2011-05-10 13:59:43 =============================================================================== ns_1@10.1.5.227:ns_doctor:82: Current node statuses: [{'ns_1@10.1.5.227', [{last_heard,{1305,61182,411249}}, {active_buckets,["default"]}, {memory, [{total,59016064}, {processes,38887184}, {processes_used,38875408}, {system,20128880}, {atom,819441}, {atom_used,806779}, {binary,111656}, {code,8552890}, {ets,1137056}]}, {cluster_compatibility_version,1}, {version, [{kernel,"2.13.4"}, {mnesia,"4.4.12"}, {stdlib,"1.16.4"}, {os_mon,"2.2.4"}, {inets,"5.2"}, {ns_server,"1.6.5.4r"}, {menelaus,"1.6.5.4r"}, {sasl,"2.1.8"}]}, {system_arch,"x86_64-unknown-linux-gnu"}, {wall_clock,574}, {memory_data,{2058731520,673247232,{<0.639.0>,19371696}}}, {disk_data, [{"/",11204616,49},{"/boot",101086,14},{"/dev/shm",1005240,0}]}, {meminfo, <<"MemTotal: 2010480 kB\nMemFree: 1352888 kB\nBuffers: 62752 kB\nCached: 342524 kB\nSwapCached: 20084 kB\nActive: 347604 kB\nInactive: 197636 kB\nHighTotal: 0 kB\nHighFree: 0 kB\nLowTotal: 2010480 kB\nLowFree: 1352888 kB\nSwapTotal: 4030456 kB\nSwapFree: 3952288 kB\nDirty: 132 kB\nWriteback: 0 kB\nAnonPages: 137788 kB\nMapped: 15732 kB\nSlab: 86432 kB\nPageTables: 6608 kB\nNFS_Unstable: 0 kB\nBounce: 0 kB\nCommitLimit: 5035696 kB\nCommitted_AS: 566488 kB\nVmallocTotal: 34359738367 kB\nVmallocUsed: 265484 kB\nVmallocChunk: 34359471011 kB\nHugePages_Total: 0\nHugePages_Free: 0\nHugePages_Rsvd: 0\nHugepagesize: 2048 kB\n">>}, {replication,[{"default",1.0}]}, {system_memory_data, [{system_total_memory,2058731520}, {free_swap,4047142912}, {total_swap,4127186944}, {cached_memory,350744576}, {buffered_memory,64258048}, {free_memory,1385357312}, {total_memory,2058731520}]}, {statistics, [{wall_clock,{570579,1}}, {context_switches,{207311,0}}, {garbage_collection,{39701,174503074,0}}, {io,{{input,54663169},{output,8878835}}}, {reductions,{54651693,713175}}, {run_queue,0}, {runtime,{7560,110}}]}]}, {'ns_1@10.1.5.229', [{last_heard,{1305,61182,242095}}, {active_buckets,["default"]}, {memory, [{total,32419784}, {processes,13289040}, {processes_used,13265520}, {system,19130744}, {atom,778121}, {atom_used,752794}, {binary,215544}, {code,7791241}, {ets,1064496}]}, {cluster_compatibility_version,1}, {version, [{kernel,"2.13.4"}, {mnesia,"4.4.12"}, {stdlib,"1.16.4"}, {os_mon,"2.2.4"}, {ns_server,"1.6.5.4r"}, {menelaus,"1.6.5.4r"}, {sasl,"2.1.8"}]}, {system_arch,"x86_64-unknown-linux-gnu"}, {wall_clock,537}, {memory_data,{2058731520,708636672,{<9352.12.0>,4114680}}}, {disk_data, [{"/",11204616,44},{"/boot",101086,14},{"/dev/shm",1005240,0}]}, {meminfo, <<"MemTotal: 2010480 kB\nMemFree: 1317088 kB\nBuffers: 79480 kB\nCached: 340436 kB\nSwapCached: 0 kB\nActive: 362500 kB\nInactive: 213508 kB\nHighTotal: 0 kB\nHighFree: 0 kB\nLowTotal: 2010480 kB\nLowFree: 1317088 kB\nSwapTotal: 4030456 kB\nSwapFree: 4030388 kB\nDirty: 188 kB\nWriteback: 0 kB\nAnonPages: 156040 kB\nMapped: 28104 kB\nSlab: 91916 kB\nPageTables: 5872 kB\nNFS_Unstable: 0 kB\nBounce: 0 kB\nCommitLimit: 5035696 kB\nCommitted_AS: 476880 kB\nVmallocTotal: 34359738367 kB\nVmallocUsed: 265484 kB\nVmallocChunk: 34359471011 kB\nHugePages_Total: 0\nHugePages_Free: 0\nHugePages_Rsvd: 0\nHugepagesize: 2048 kB\n">>}, {replication,[{"default",1.0}]}, {system_memory_data, [{system_total_memory,2058731520}, {free_swap,4127117312}, {total_swap,4127186944}, {cached_memory,348606464}, {buffered_memory,81387520}, {free_memory,1348698112}, {total_memory,2058731520}]}, {statistics, [{wall_clock,{529806,1}}, {context_switches,{50021,0}}, {garbage_collection,{10721,37723863,0}}, {io,{{input,6371870},{output,2833502}}}, {reductions,{13114034,231889}}, {run_queue,0}, {runtime,{2130,30}}]}]}] INFO REPORT <0.85.0> 2011-05-10 13:59:45 =============================================================================== Pulling config from: 'ns_1@10.1.5.229' INFO REPORT <0.3917.0> 2011-05-10 13:59:46 =============================================================================== vbucketmigrator<0.3917.0>: Validate bucket states vbucketmigrator<0.3917.0>: 62 ok INFO REPORT <0.3934.0> 2011-05-10 13:59:47 =============================================================================== vbucketmigrator<0.3934.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.3934.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.3934.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.3934.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.3934.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.3934.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.3934.0>: Starting to move bucket 63 vbucketmigrator<0.3934.0>: Bucket 63 moved to the next server INFO REPORT <0.104.0> 2011-05-10 13:59:47 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.109.0> 2011-05-10 13:59:48 =============================================================================== ns_1@10.1.5.227:ns_orchestrator:164: Skipping janitor in state rebalancing: {rebalancing_state, <0.577.0>, {dict, 2, 16, 16, 8, 80, 48, {[], [], [], [], [], [], [], [], [], [], [], [], [], [], [], []}, {{[], [], [], [], [], [], [], [['ns_1@10.1.5.227'| 0.123046875]], [], [['ns_1@10.1.5.229'| 0.123046875]], [], [], [], [], [], []}}}} INFO REPORT <0.3934.0> 2011-05-10 13:59:52 =============================================================================== vbucketmigrator<0.3934.0>: Validate bucket states vbucketmigrator<0.3934.0>: 63 ok INFO REPORT <0.3958.0> 2011-05-10 13:59:53 =============================================================================== vbucketmigrator<0.3958.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.3958.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.3958.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.3958.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.3958.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.3958.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.3958.0>: Starting to move bucket 64 vbucketmigrator<0.3958.0>: Bucket 64 moved to the next server INFO REPORT <0.104.0> 2011-05-10 13:59:53 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.3958.0> 2011-05-10 13:59:58 =============================================================================== vbucketmigrator<0.3958.0>: Validate bucket states vbucketmigrator<0.3958.0>: 64 ok INFO REPORT <0.109.0> 2011-05-10 13:59:58 =============================================================================== ns_1@10.1.5.227:ns_orchestrator:164: Skipping janitor in state rebalancing: {rebalancing_state, <0.577.0>, {dict, 2, 16, 16, 8, 80, 48, {[], [], [], [], [], [], [], [], [], [], [], [], [], [], [], []}, {{[], [], [], [], [], [], [], [['ns_1@10.1.5.227'| 0.126953125]], [], [['ns_1@10.1.5.229'| 0.126953125]], [], [], [], [], [], []}}}} INFO REPORT <0.3962.0> 2011-05-10 13:59:59 =============================================================================== vbucketmigrator<0.3962.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.3962.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.3962.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.3962.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.3962.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.3962.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.3962.0>: Starting to move bucket 65 vbucketmigrator<0.3962.0>: Bucket 65 moved to the next server INFO REPORT <0.104.0> 2011-05-10 13:59:59 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.3962.0> 2011-05-10 14:00:04 =============================================================================== vbucketmigrator<0.3962.0>: Validate bucket states vbucketmigrator<0.3962.0>: 65 ok INFO REPORT <0.3969.0> 2011-05-10 14:00:05 =============================================================================== vbucketmigrator<0.3969.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.3969.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.3969.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.3969.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.3969.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.3969.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.3969.0>: Starting to move bucket 66 vbucketmigrator<0.3969.0>: Bucket 66 moved to the next server INFO REPORT <0.104.0> 2011-05-10 14:00:05 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.109.0> 2011-05-10 14:00:08 =============================================================================== ns_1@10.1.5.227:ns_orchestrator:164: Skipping janitor in state rebalancing: {rebalancing_state, <0.577.0>, {dict, 2, 16, 16, 8, 80, 48, {[], [], [], [], [], [], [], [], [], [], [], [], [], [], [], []}, {{[], [], [], [], [], [], [], [['ns_1@10.1.5.227'| 0.12890625]], [], [['ns_1@10.1.5.229'| 0.12890625]], [], [], [], [], [], []}}}} INFO REPORT <0.3969.0> 2011-05-10 14:00:10 =============================================================================== vbucketmigrator<0.3969.0>: Validate bucket states vbucketmigrator<0.3969.0>: 66 ok INFO REPORT <0.3982.0> 2011-05-10 14:00:11 =============================================================================== vbucketmigrator<0.3982.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.3982.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.3982.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.3982.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.3982.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.3982.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.3982.0>: Starting to move bucket 67 vbucketmigrator<0.3982.0>: Bucket 67 moved to the next server INFO REPORT <0.104.0> 2011-05-10 14:00:11 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.3982.0> 2011-05-10 14:00:16 =============================================================================== vbucketmigrator<0.3982.0>: Validate bucket states vbucketmigrator<0.3982.0>: 67 ok INFO REPORT <0.3990.0> 2011-05-10 14:00:17 =============================================================================== vbucketmigrator<0.3990.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.3990.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.3990.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.3990.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.3990.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.3990.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.3990.0>: Starting to move bucket 68 vbucketmigrator<0.3990.0>: Bucket 68 moved to the next server INFO REPORT <0.104.0> 2011-05-10 14:00:17 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.109.0> 2011-05-10 14:00:18 =============================================================================== ns_1@10.1.5.227:ns_orchestrator:164: Skipping janitor in state rebalancing: {rebalancing_state, <0.577.0>, {dict, 2, 16, 16, 8, 80, 48, {[], [], [], [], [], [], [], [], [], [], [], [], [], [], [], []}, {{[], [], [], [], [], [], [], [['ns_1@10.1.5.227'| 0.1328125]], [], [['ns_1@10.1.5.229'| 0.1328125]], [], [], [], [], [], []}}}} INFO REPORT <0.3990.0> 2011-05-10 14:00:22 =============================================================================== vbucketmigrator<0.3990.0>: Validate bucket states vbucketmigrator<0.3990.0>: 68 ok INFO REPORT <0.3995.0> 2011-05-10 14:00:23 =============================================================================== vbucketmigrator<0.3995.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.3995.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.3995.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.3995.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.3995.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.3995.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.3995.0>: Starting to move bucket 69 vbucketmigrator<0.3995.0>: Bucket 69 moved to the next server INFO REPORT <0.104.0> 2011-05-10 14:00:23 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.469.0> 2011-05-10 14:00:24 =============================================================================== ns_1@10.1.5.227:stats_collector:71: Stats for bucket "default": auth_cmds 102 auth_errors 0 bucket_conns 34 bytes_read 1600779 bytes_written 4791741 cas_badval 0 cas_hits 0 cas_misses 0 cmd_flush 1 cmd_get 0 cmd_set 16004 conn_yields 0 connection_structures 44 curr_connections 44 curr_items 7492 curr_items_tot 8040 daemon_connections 10 decr_hits 0 decr_misses 0 delete_hits 0 delete_misses 0 ep_bg_fetched 0 ep_commit_num 53 ep_commit_time 0 ep_commit_time_total 4 ep_data_age 13 ep_data_age_highwat 13 ep_db_cleaner_status complete ep_db_strategy multiMTVBDB ep_dbinit 1 ep_dbname /var/opt/membase/1.6.5.4r/data/ns_1/default ep_dbshards 4 ep_expired 0 ep_flush_duration 1 ep_flush_duration_highwat 8 ep_flush_duration_total 12 ep_flush_preempts 0 ep_flusher_state running ep_flusher_todo 0 ep_io_num_read 548 ep_io_num_write 15741 ep_io_read_bytes 31093 ep_io_write_bytes 1011432 ep_item_begin_failed 0 ep_item_commit_failed 0 ep_item_flush_expired 0 ep_item_flush_failed 0 ep_kv_size 1181400 ep_max_data_size 1096810496 ep_max_txn_size 1000 ep_mem_high_wat 822607872 ep_mem_low_wat 658086297 ep_min_data_age 0 ep_num_active_non_resident 0 ep_num_eject_failures 0 ep_num_eject_replicas 0 ep_num_expiry_pager_runs 0 ep_num_non_resident 0 ep_num_not_my_vbuckets 0 ep_num_pager_runs 0 ep_num_value_ejects 0 ep_oom_errors 0 ep_overhead 25854448 ep_pending_ops 0 ep_pending_ops_max 0 ep_pending_ops_max_duration 0 ep_pending_ops_total 0 ep_queue_age_cap 900 ep_queue_size 0 ep_storage_age 0 ep_storage_age_highwat 10 ep_storage_type featured ep_store_max_concurrency 10 ep_store_max_readers 9 ep_store_max_readwrite 1 ep_tap_bg_fetch_requeued 0 ep_tap_bg_fetched 548 ep_tap_keepalive 300 ep_tmp_oom_errors 0 ep_too_old 0 ep_too_young 0 ep_total_cache_size 1181400 ep_total_del_items 0 ep_total_enqueued 16035 ep_total_new_items 15741 ep_total_persisted 15741 ep_vbucket_del 0 ep_vbucket_del_fail 0 ep_version 1.6.5.3.1_13_g81a6cf1 ep_warmed_up 0 ep_warmup true ep_warmup_dups 0 ep_warmup_oom 0 ep_warmup_thread complete ep_warmup_time 15662 get_hits 0 get_misses 0 incr_hits 0 incr_misses 0 libevent 1.4.13-stable limit_maxbytes 67108864 mem_used 27035848 pid 16183 pointer_size 64 rejected_conns 0 rusage_system 1.252809 rusage_user 3.071533 tap_connect_received 70 tap_mutation_sent 1096 tap_opaque_sent 70 tap_vbucket_set_sent 140 threads 4 time 1305061224 total_connections 113 uptime 618 version 1.4.4_382_g9df3289 INFO REPORT <0.3995.0> 2011-05-10 14:00:28 =============================================================================== vbucketmigrator<0.3995.0>: Validate bucket states vbucketmigrator<0.3995.0>: 69 ok INFO REPORT <0.109.0> 2011-05-10 14:00:28 =============================================================================== ns_1@10.1.5.227:ns_orchestrator:164: Skipping janitor in state rebalancing: {rebalancing_state, <0.577.0>, {dict, 2, 16, 16, 8, 80, 48, {[], [], [], [], [], [], [], [], [], [], [], [], [], [], [], []}, {{[], [], [], [], [], [], [], [['ns_1@10.1.5.227'| 0.13671875]], [], [['ns_1@10.1.5.229'| 0.13671875]], [], [], [], [], [], []}}}} INFO REPORT <0.3999.0> 2011-05-10 14:00:29 =============================================================================== vbucketmigrator<0.3999.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.3999.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.3999.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.3999.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.3999.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.3999.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.3999.0>: Starting to move bucket 70 vbucketmigrator<0.3999.0>: Bucket 70 moved to the next server INFO REPORT <0.104.0> 2011-05-10 14:00:29 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.3999.0> 2011-05-10 14:00:34 =============================================================================== vbucketmigrator<0.3999.0>: Validate bucket states vbucketmigrator<0.3999.0>: 70 ok INFO REPORT <0.4014.0> 2011-05-10 14:00:35 =============================================================================== vbucketmigrator<0.4014.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.4014.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.4014.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.4014.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.4014.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.4014.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.4014.0>: Starting to move bucket 71 vbucketmigrator<0.4014.0>: Bucket 71 moved to the next server INFO REPORT <0.104.0> 2011-05-10 14:00:35 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.109.0> 2011-05-10 14:00:38 =============================================================================== ns_1@10.1.5.227:ns_orchestrator:164: Skipping janitor in state rebalancing: {rebalancing_state, <0.577.0>, {dict, 2, 16, 16, 8, 80, 48, {[], [], [], [], [], [], [], [], [], [], [], [], [], [], [], []}, {{[], [], [], [], [], [], [], [['ns_1@10.1.5.227'| 0.138671875]], [], [['ns_1@10.1.5.229'| 0.138671875]], [], [], [], [], [], []}}}} INFO REPORT <0.85.0> 2011-05-10 14:00:39 =============================================================================== Pulling config from: 'ns_1@10.1.5.229' INFO REPORT <0.4014.0> 2011-05-10 14:00:40 =============================================================================== vbucketmigrator<0.4014.0>: Validate bucket states vbucketmigrator<0.4014.0>: 71 ok INFO REPORT <0.4019.0> 2011-05-10 14:00:41 =============================================================================== vbucketmigrator<0.4019.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.4019.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.4019.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.4019.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.4019.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.4019.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.4019.0>: Starting to move bucket 72 vbucketmigrator<0.4019.0>: Bucket 72 moved to the next server INFO REPORT <0.104.0> 2011-05-10 14:00:41 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.474.0> 2011-05-10 14:00:43 =============================================================================== ns_1@10.1.5.227:ns_doctor:82: Current node statuses: [{'ns_1@10.1.5.227', [{last_heard,{1305,61242,411241}}, {active_buckets,["default"]}, {memory, [{total,40166608}, {processes,19978592}, {processes_used,19965952}, {system,20188016}, {atom,819441}, {atom_used,806779}, {binary,98048}, {code,8552890}, {ets,1203336}]}, {cluster_compatibility_version,1}, {version, [{kernel,"2.13.4"}, {mnesia,"4.4.12"}, {stdlib,"1.16.4"}, {os_mon,"2.2.4"}, {inets,"5.2"}, {ns_server,"1.6.5.4r"}, {menelaus,"1.6.5.4r"}, {sasl,"2.1.8"}]}, {system_arch,"x86_64-unknown-linux-gnu"}, {wall_clock,634}, {memory_data,{2058731520,671469568,{<0.639.0>,19371696}}}, {disk_data, [{"/",11204616,49},{"/boot",101086,14},{"/dev/shm",1005240,0}]}, {meminfo, <<"MemTotal: 2010480 kB\nMemFree: 1354748 kB\nBuffers: 63104 kB\nCached: 341620 kB\nSwapCached: 20084 kB\nActive: 346580 kB\nInactive: 196844 kB\nHighTotal: 0 kB\nHighFree: 0 kB\nLowTotal: 2010480 kB\nLowFree: 1354748 kB\nSwapTotal: 4030456 kB\nSwapFree: 3952288 kB\nDirty: 268 kB\nWriteback: 0 kB\nAnonPages: 136508 kB\nMapped: 15732 kB\nSlab: 86420 kB\nPageTables: 6600 kB\nNFS_Unstable: 0 kB\nBounce: 0 kB\nCommitLimit: 5035696 kB\nCommitted_AS: 547564 kB\nVmallocTotal: 34359738367 kB\nVmallocUsed: 265484 kB\nVmallocChunk: 34359471011 kB\nHugePages_Total: 0\nHugePages_Free: 0\nHugePages_Rsvd: 0\nHugepagesize: 2048 kB\n">>}, {replication,[{"default",1.0}]}, {system_memory_data, [{system_total_memory,2058731520}, {free_swap,4047142912}, {total_swap,4127186944}, {cached_memory,349818880}, {buffered_memory,64618496}, {free_memory,1387261952}, {total_memory,2058731520}]}, {statistics, [{wall_clock,{630579,1}}, {context_switches,{219514,0}}, {garbage_collection,{42082,196211801,0}}, {io,{{input,55812070},{output,9893162}}}, {reductions,{59685462,746861}}, {run_queue,0}, {runtime,{8410,110}}]}]}, {'ns_1@10.1.5.229', [{last_heard,{1305,61242,242067}}, {active_buckets,["default"]}, {memory, [{total,32491528}, {processes,13293920}, {processes_used,13270400}, {system,19197608}, {atom,778121}, {atom_used,752794}, {binary,215208}, {code,7791241}, {ets,1130688}]}, {cluster_compatibility_version,1}, {version, [{kernel,"2.13.4"}, {mnesia,"4.4.12"}, {stdlib,"1.16.4"}, {os_mon,"2.2.4"}, {ns_server,"1.6.5.4r"}, {menelaus,"1.6.5.4r"}, {sasl,"2.1.8"}]}, {system_arch,"x86_64-unknown-linux-gnu"}, {wall_clock,597}, {memory_data,{2058731520,709906432,{<9352.12.0>,4114680}}}, {disk_data, [{"/",11204616,44},{"/boot",101086,14},{"/dev/shm",1005240,0}]}, {meminfo, <<"MemTotal: 2010480 kB\nMemFree: 1317088 kB\nBuffers: 80216 kB\nCached: 339776 kB\nSwapCached: 0 kB\nActive: 363020 kB\nInactive: 213316 kB\nHighTotal: 0 kB\nHighFree: 0 kB\nLowTotal: 2010480 kB\nLowFree: 1317088 kB\nSwapTotal: 4030456 kB\nSwapFree: 4030388 kB\nDirty: 240 kB\nWriteback: 0 kB\nAnonPages: 156312 kB\nMapped: 28104 kB\nSlab: 91904 kB\nPageTables: 5868 kB\nNFS_Unstable: 0 kB\nBounce: 0 kB\nCommitLimit: 5035696 kB\nCommitted_AS: 476880 kB\nVmallocTotal: 34359738367 kB\nVmallocUsed: 265484 kB\nVmallocChunk: 34359471011 kB\nHugePages_Total: 0\nHugePages_Free: 0\nHugePages_Rsvd: 0\nHugepagesize: 2048 kB\n">>}, {replication,[{"default",1.0}]}, {system_memory_data, [{system_total_memory,2058731520}, {free_swap,4127117312}, {total_swap,4127186944}, {cached_memory,347930624}, {buffered_memory,82141184}, {free_memory,1348698112}, {total_memory,2058731520}]}, {statistics, [{wall_clock,{589806,1}}, {context_switches,{55219,0}}, {garbage_collection,{11864,42434773,0}}, {io,{{input,7346775},{output,3364620}}}, {reductions,{14559910,235962}}, {run_queue,0}, {runtime,{2410,40}}]}]}] INFO REPORT <0.4019.0> 2011-05-10 14:00:46 =============================================================================== vbucketmigrator<0.4019.0>: Validate bucket states vbucketmigrator<0.4019.0>: 72 ok INFO REPORT <0.4023.0> 2011-05-10 14:00:47 =============================================================================== vbucketmigrator<0.4023.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.4023.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.4023.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.4023.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.4023.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.4023.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.4023.0>: Starting to move bucket 73 vbucketmigrator<0.4023.0>: Bucket 73 moved to the next server INFO REPORT <0.104.0> 2011-05-10 14:00:47 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.109.0> 2011-05-10 14:00:48 =============================================================================== ns_1@10.1.5.227:ns_orchestrator:164: Skipping janitor in state rebalancing: {rebalancing_state, <0.577.0>, {dict, 2, 16, 16, 8, 80, 48, {[], [], [], [], [], [], [], [], [], [], [], [], [], [], [], []}, {{[], [], [], [], [], [], [], [['ns_1@10.1.5.227'| 0.142578125]], [], [['ns_1@10.1.5.229'| 0.142578125]], [], [], [], [], [], []}}}} INFO REPORT <0.4023.0> 2011-05-10 14:00:52 =============================================================================== vbucketmigrator<0.4023.0>: Validate bucket states vbucketmigrator<0.4023.0>: 73 ok INFO REPORT <0.4036.0> 2011-05-10 14:00:53 =============================================================================== vbucketmigrator<0.4036.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.4036.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.4036.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.4036.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.4036.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.4036.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.4036.0>: Starting to move bucket 74 vbucketmigrator<0.4036.0>: Bucket 74 moved to the next server INFO REPORT <0.104.0> 2011-05-10 14:00:53 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.4036.0> 2011-05-10 14:00:58 =============================================================================== vbucketmigrator<0.4036.0>: Validate bucket states vbucketmigrator<0.4036.0>: 74 ok INFO REPORT <0.109.0> 2011-05-10 14:00:58 =============================================================================== ns_1@10.1.5.227:ns_orchestrator:164: Skipping janitor in state rebalancing: {rebalancing_state, <0.577.0>, {dict, 2, 16, 16, 8, 80, 48, {[], [], [], [], [], [], [], [], [], [], [], [], [], [], [], []}, {{[], [], [], [], [], [], [], [['ns_1@10.1.5.227'| 0.146484375]], [], [['ns_1@10.1.5.229'| 0.146484375]], [], [], [], [], [], []}}}} INFO REPORT <0.4040.0> 2011-05-10 14:00:59 =============================================================================== vbucketmigrator<0.4040.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.4040.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.4040.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.4040.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.4040.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.4040.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.4040.0>: Starting to move bucket 75 vbucketmigrator<0.4040.0>: Bucket 75 moved to the next server INFO REPORT <0.104.0> 2011-05-10 14:00:59 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.4040.0> 2011-05-10 14:01:04 =============================================================================== vbucketmigrator<0.4040.0>: Validate bucket states vbucketmigrator<0.4040.0>: 75 ok INFO REPORT <0.4047.0> 2011-05-10 14:01:05 =============================================================================== vbucketmigrator<0.4047.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.4047.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.4047.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.4047.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.4047.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.4047.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.4047.0>: Starting to move bucket 76 vbucketmigrator<0.4047.0>: Bucket 76 moved to the next server INFO REPORT <0.104.0> 2011-05-10 14:01:05 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.109.0> 2011-05-10 14:01:08 =============================================================================== ns_1@10.1.5.227:ns_orchestrator:164: Skipping janitor in state rebalancing: {rebalancing_state, <0.577.0>, {dict, 2, 16, 16, 8, 80, 48, {[], [], [], [], [], [], [], [], [], [], [], [], [], [], [], []}, {{[], [], [], [], [], [], [], [['ns_1@10.1.5.227'| 0.1484375]], [], [['ns_1@10.1.5.229'| 0.1484375]], [], [], [], [], [], []}}}} INFO REPORT <0.4047.0> 2011-05-10 14:01:10 =============================================================================== vbucketmigrator<0.4047.0>: Validate bucket states vbucketmigrator<0.4047.0>: 76 ok INFO REPORT <0.4057.0> 2011-05-10 14:01:11 =============================================================================== vbucketmigrator<0.4057.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.4057.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.4057.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.4057.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.4057.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.4057.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.4057.0>: Starting to move bucket 77 vbucketmigrator<0.4057.0>: Bucket 77 moved to the next server INFO REPORT <0.104.0> 2011-05-10 14:01:11 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.4057.0> 2011-05-10 14:01:16 =============================================================================== vbucketmigrator<0.4057.0>: Validate bucket states vbucketmigrator<0.4057.0>: 77 ok INFO REPORT <0.4066.0> 2011-05-10 14:01:17 =============================================================================== vbucketmigrator<0.4066.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.4066.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.4066.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.4066.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.4066.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.4066.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.4066.0>: Starting to move bucket 78 vbucketmigrator<0.4066.0>: Bucket 78 moved to the next server INFO REPORT <0.104.0> 2011-05-10 14:01:17 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.109.0> 2011-05-10 14:01:18 =============================================================================== ns_1@10.1.5.227:ns_orchestrator:164: Skipping janitor in state rebalancing: {rebalancing_state, <0.577.0>, {dict, 2, 16, 16, 8, 80, 48, {[], [], [], [], [], [], [], [], [], [], [], [], [], [], [], []}, {{[], [], [], [], [], [], [], [['ns_1@10.1.5.227'| 0.15234375]], [], [['ns_1@10.1.5.229'| 0.15234375]], [], [], [], [], [], []}}}} INFO REPORT <0.4066.0> 2011-05-10 14:01:22 =============================================================================== vbucketmigrator<0.4066.0>: Validate bucket states vbucketmigrator<0.4066.0>: 78 ok INFO REPORT <0.4071.0> 2011-05-10 14:01:23 =============================================================================== vbucketmigrator<0.4071.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.4071.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.4071.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.4071.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.4071.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.4071.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.4071.0>: Starting to move bucket 79 vbucketmigrator<0.4071.0>: Bucket 79 moved to the next server INFO REPORT <0.104.0> 2011-05-10 14:01:23 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.4071.0> 2011-05-10 14:01:28 =============================================================================== vbucketmigrator<0.4071.0>: Validate bucket states vbucketmigrator<0.4071.0>: 79 ok INFO REPORT <0.109.0> 2011-05-10 14:01:28 =============================================================================== ns_1@10.1.5.227:ns_orchestrator:164: Skipping janitor in state rebalancing: {rebalancing_state, <0.577.0>, {dict, 2, 16, 16, 8, 80, 48, {[], [], [], [], [], [], [], [], [], [], [], [], [], [], [], []}, {{[], [], [], [], [], [], [], [['ns_1@10.1.5.227'| 0.15625]], [], [['ns_1@10.1.5.229'| 0.15625]], [], [], [], [], [], []}}}} INFO REPORT <0.4075.0> 2011-05-10 14:01:29 =============================================================================== vbucketmigrator<0.4075.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.4075.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.4075.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.4075.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.4075.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.4075.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.4075.0>: Starting to move bucket 80 vbucketmigrator<0.4075.0>: Bucket 80 moved to the next server INFO REPORT <0.104.0> 2011-05-10 14:01:29 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.85.0> 2011-05-10 14:01:30 =============================================================================== Pulling config from: 'ns_1@10.1.5.229' INFO REPORT <0.4075.0> 2011-05-10 14:01:34 =============================================================================== vbucketmigrator<0.4075.0>: Validate bucket states vbucketmigrator<0.4075.0>: 80 ok INFO REPORT <0.4090.0> 2011-05-10 14:01:35 =============================================================================== vbucketmigrator<0.4090.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.4090.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.4090.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.4090.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.4090.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.4090.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.4090.0>: Starting to move bucket 81 vbucketmigrator<0.4090.0>: Bucket 81 moved to the next server INFO REPORT <0.104.0> 2011-05-10 14:01:35 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.109.0> 2011-05-10 14:01:38 =============================================================================== ns_1@10.1.5.227:ns_orchestrator:164: Skipping janitor in state rebalancing: {rebalancing_state, <0.577.0>, {dict, 2, 16, 16, 8, 80, 48, {[], [], [], [], [], [], [], [], [], [], [], [], [], [], [], []}, {{[], [], [], [], [], [], [], [['ns_1@10.1.5.227'| 0.158203125]], [], [['ns_1@10.1.5.229'| 0.158203125]], [], [], [], [], [], []}}}} INFO REPORT <0.4090.0> 2011-05-10 14:01:40 =============================================================================== vbucketmigrator<0.4090.0>: Validate bucket states vbucketmigrator<0.4090.0>: 81 ok INFO REPORT <0.4098.0> 2011-05-10 14:01:41 =============================================================================== vbucketmigrator<0.4098.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.4098.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.4098.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.4098.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.4098.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.4098.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.4098.0>: Starting to move bucket 82 vbucketmigrator<0.4098.0>: Bucket 82 moved to the next server INFO REPORT <0.104.0> 2011-05-10 14:01:41 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.474.0> 2011-05-10 14:01:43 =============================================================================== ns_1@10.1.5.227:ns_doctor:82: Current node statuses: [{'ns_1@10.1.5.227', [{last_heard,{1305,61302,411207}}, {active_buckets,["default"]}, {memory, [{total,41246568}, {processes,20976480}, {processes_used,20964704}, {system,20270088}, {atom,819441}, {atom_used,806779}, {binary,118272}, {code,8552890}, {ets,1263504}]}, {cluster_compatibility_version,1}, {version, [{kernel,"2.13.4"}, {mnesia,"4.4.12"}, {stdlib,"1.16.4"}, {os_mon,"2.2.4"}, {inets,"5.2"}, {ns_server,"1.6.5.4r"}, {menelaus,"1.6.5.4r"}, {sasl,"2.1.8"}]}, {system_arch,"x86_64-unknown-linux-gnu"}, {wall_clock,694}, {memory_data,{2058731520,671469568,{<0.12.0>,6657168}}}, {disk_data, [{"/",11204616,49},{"/boot",101086,14},{"/dev/shm",1005240,0}]}, {meminfo, <<"MemTotal: 2010480 kB\nMemFree: 1354128 kB\nBuffers: 63400 kB\nCached: 341876 kB\nSwapCached: 20084 kB\nActive: 346748 kB\nInactive: 197288 kB\nHighTotal: 0 kB\nHighFree: 0 kB\nLowTotal: 2010480 kB\nLowFree: 1354128 kB\nSwapTotal: 4030456 kB\nSwapFree: 3952288 kB\nDirty: 232 kB\nWriteback: 0 kB\nAnonPages: 136568 kB\nMapped: 15732 kB\nSlab: 86432 kB\nPageTables: 6604 kB\nNFS_Unstable: 0 kB\nBounce: 0 kB\nCommitLimit: 5035696 kB\nCommitted_AS: 549512 kB\nVmallocTotal: 34359738367 kB\nVmallocUsed: 265484 kB\nVmallocChunk: 34359471011 kB\nHugePages_Total: 0\nHugePages_Free: 0\nHugePages_Rsvd: 0\nHugepagesize: 2048 kB\n">>}, {replication,[{"default",1.0}]}, {system_memory_data, [{system_total_memory,2058731520}, {free_swap,4047142912}, {total_swap,4127186944}, {cached_memory,350081024}, {buffered_memory,64921600}, {free_memory,1386627072}, {total_memory,2058731520}]}, {statistics, [{wall_clock,{690579,1}}, {context_switches,{229874,0}}, {garbage_collection,{44379,215822110,0}}, {io,{{input,55956480},{output,10316663}}}, {reductions,{64245439,748461}}, {run_queue,0}, {runtime,{9080,120}}]}]}, {'ns_1@10.1.5.229', [{last_heard,{1305,61302,242146}}, {active_buckets,["default"]}, {memory, [{total,32583096}, {processes,13325216}, {processes_used,13301696}, {system,19257880}, {atom,778121}, {atom_used,752794}, {binary,215152}, {code,7791241}, {ets,1190664}]}, {cluster_compatibility_version,1}, {version, [{kernel,"2.13.4"}, {mnesia,"4.4.12"}, {stdlib,"1.16.4"}, {os_mon,"2.2.4"}, {ns_server,"1.6.5.4r"}, {menelaus,"1.6.5.4r"}, {sasl,"2.1.8"}]}, {system_arch,"x86_64-unknown-linux-gnu"}, {wall_clock,657}, {memory_data,{2058731520,710541312,{<9352.12.0>,4114680}}}, {disk_data, [{"/",11204616,44},{"/boot",101086,14},{"/dev/shm",1005240,0}]}, {meminfo, <<"MemTotal: 2010480 kB\nMemFree: 1314856 kB\nBuffers: 80900 kB\nCached: 340188 kB\nSwapCached: 0 kB\nActive: 364976 kB\nInactive: 213940 kB\nHighTotal: 0 kB\nHighFree: 0 kB\nLowTotal: 2010480 kB\nLowFree: 1314856 kB\nSwapTotal: 4030456 kB\nSwapFree: 4030388 kB\nDirty: 128 kB\nWriteback: 0 kB\nAnonPages: 157816 kB\nMapped: 28104 kB\nSlab: 91900 kB\nPageTables: 5864 kB\nNFS_Unstable: 0 kB\nBounce: 0 kB\nCommitLimit: 5035696 kB\nCommitted_AS: 481764 kB\nVmallocTotal: 34359738367 kB\nVmallocUsed: 265484 kB\nVmallocChunk: 34359471011 kB\nHugePages_Total: 0\nHugePages_Free: 0\nHugePages_Rsvd: 0\nHugepagesize: 2048 kB\n">>}, {replication,[{"default",1.0}]}, {system_memory_data, [{system_total_memory,2058731520}, {free_swap,4127117312}, {total_swap,4127186944}, {cached_memory,348352512}, {buffered_memory,82841600}, {free_memory,1346412544}, {total_memory,2058731520}]}, {statistics, [{wall_clock,{649807,1}}, {context_switches,{60015,0}}, {garbage_collection,{12895,46916123,0}}, {io,{{input,7583720},{output,3622062}}}, {reductions,{15922961,232496}}, {run_queue,0}, {runtime,{2600,30}}]}]}] INFO REPORT <0.4098.0> 2011-05-10 14:01:46 =============================================================================== vbucketmigrator<0.4098.0>: Validate bucket states vbucketmigrator<0.4098.0>: 82 ok INFO REPORT <0.4102.0> 2011-05-10 14:01:47 =============================================================================== vbucketmigrator<0.4102.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.4102.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.4102.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.4102.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.4102.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.4102.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.4102.0>: Starting to move bucket 83 vbucketmigrator<0.4102.0>: Bucket 83 moved to the next server INFO REPORT <0.104.0> 2011-05-10 14:01:47 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.109.0> 2011-05-10 14:01:48 =============================================================================== ns_1@10.1.5.227:ns_orchestrator:164: Skipping janitor in state rebalancing: {rebalancing_state, <0.577.0>, {dict, 2, 16, 16, 8, 80, 48, {[], [], [], [], [], [], [], [], [], [], [], [], [], [], [], []}, {{[], [], [], [], [], [], [], [['ns_1@10.1.5.227'| 0.162109375]], [], [['ns_1@10.1.5.229'| 0.162109375]], [], [], [], [], [], []}}}} INFO REPORT <0.85.0> 2011-05-10 14:01:49 =============================================================================== Pulling config from: 'ns_1@10.1.5.229' INFO REPORT <0.4102.0> 2011-05-10 14:01:52 =============================================================================== vbucketmigrator<0.4102.0>: Validate bucket states vbucketmigrator<0.4102.0>: 83 ok INFO REPORT <0.4115.0> 2011-05-10 14:01:53 =============================================================================== vbucketmigrator<0.4115.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.4115.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.4115.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.4115.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.4115.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.4115.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.4115.0>: Starting to move bucket 84 vbucketmigrator<0.4115.0>: Bucket 84 moved to the next server INFO REPORT <0.104.0> 2011-05-10 14:01:53 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.85.0> 2011-05-10 14:01:55 =============================================================================== Pulling config from: 'ns_1@10.1.5.229' INFO REPORT <0.4115.0> 2011-05-10 14:01:58 =============================================================================== vbucketmigrator<0.4115.0>: Validate bucket states vbucketmigrator<0.4115.0>: 84 ok INFO REPORT <0.109.0> 2011-05-10 14:01:58 =============================================================================== ns_1@10.1.5.227:ns_orchestrator:164: Skipping janitor in state rebalancing: {rebalancing_state, <0.577.0>, {dict, 2, 16, 16, 8, 80, 48, {[], [], [], [], [], [], [], [], [], [], [], [], [], [], [], []}, {{[], [], [], [], [], [], [], [['ns_1@10.1.5.227'| 0.166015625]], [], [['ns_1@10.1.5.229'| 0.166015625]], [], [], [], [], [], []}}}} INFO REPORT <0.4119.0> 2011-05-10 14:01:59 =============================================================================== vbucketmigrator<0.4119.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.4119.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.4119.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.4119.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.4119.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.4119.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.4119.0>: Starting to move bucket 85 vbucketmigrator<0.4119.0>: Bucket 85 moved to the next server INFO REPORT <0.104.0> 2011-05-10 14:01:59 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.85.0> 2011-05-10 14:02:02 =============================================================================== Pulling config from: 'ns_1@10.1.5.229' INFO REPORT <0.469.0> 2011-05-10 14:02:04 =============================================================================== ns_1@10.1.5.227:stats_collector:71: Stats for bucket "default": auth_cmds 118 auth_errors 0 bucket_conns 34 bytes_read 1607437 bytes_written 5834843 cas_badval 0 cas_hits 0 cas_misses 0 cmd_flush 1 cmd_get 0 cmd_set 16004 conn_yields 0 connection_structures 44 curr_connections 44 curr_items 7374 curr_items_tot 8040 daemon_connections 10 decr_hits 0 decr_misses 0 delete_hits 0 delete_misses 0 ep_bg_fetched 0 ep_commit_num 53 ep_commit_time 0 ep_commit_time_total 4 ep_data_age 13 ep_data_age_highwat 13 ep_db_cleaner_status complete ep_db_strategy multiMTVBDB ep_dbinit 1 ep_dbname /var/opt/membase/1.6.5.4r/data/ns_1/default ep_dbshards 4 ep_expired 0 ep_flush_duration 1 ep_flush_duration_highwat 8 ep_flush_duration_total 12 ep_flush_preempts 0 ep_flusher_state running ep_flusher_todo 0 ep_io_num_read 666 ep_io_num_write 15741 ep_io_read_bytes 37695 ep_io_write_bytes 1011432 ep_item_begin_failed 0 ep_item_commit_failed 0 ep_item_flush_expired 0 ep_item_flush_failed 0 ep_kv_size 1181400 ep_max_data_size 1096810496 ep_max_txn_size 1000 ep_mem_high_wat 822607872 ep_mem_low_wat 658086297 ep_min_data_age 0 ep_num_active_non_resident 0 ep_num_eject_failures 0 ep_num_eject_replicas 0 ep_num_expiry_pager_runs 0 ep_num_non_resident 0 ep_num_not_my_vbuckets 0 ep_num_pager_runs 0 ep_num_value_ejects 0 ep_oom_errors 0 ep_overhead 25854448 ep_pending_ops 0 ep_pending_ops_max 0 ep_pending_ops_max_duration 0 ep_pending_ops_total 0 ep_queue_age_cap 900 ep_queue_size 0 ep_storage_age 0 ep_storage_age_highwat 10 ep_storage_type featured ep_store_max_concurrency 10 ep_store_max_readers 9 ep_store_max_readwrite 1 ep_tap_bg_fetch_requeued 0 ep_tap_bg_fetched 666 ep_tap_keepalive 300 ep_tmp_oom_errors 0 ep_too_old 0 ep_too_young 0 ep_total_cache_size 1181400 ep_total_del_items 0 ep_total_enqueued 16035 ep_total_new_items 15741 ep_total_persisted 15741 ep_vbucket_del 0 ep_vbucket_del_fail 0 ep_version 1.6.5.3.1_13_g81a6cf1 ep_warmed_up 0 ep_warmup true ep_warmup_dups 0 ep_warmup_oom 0 ep_warmup_thread complete ep_warmup_time 15662 get_hits 0 get_misses 0 incr_hits 0 incr_misses 0 libevent 1.4.13-stable limit_maxbytes 67108864 mem_used 27035848 pid 16183 pointer_size 64 rejected_conns 0 rusage_system 1.288804 rusage_user 3.235508 tap_connect_received 86 tap_mutation_sent 1332 tap_opaque_sent 86 tap_vbucket_set_sent 172 threads 4 time 1305061324 total_connections 129 uptime 718 version 1.4.4_382_g9df3289 INFO REPORT <0.4119.0> 2011-05-10 14:02:04 =============================================================================== vbucketmigrator<0.4119.0>: Validate bucket states vbucketmigrator<0.4119.0>: 85 ok INFO REPORT <0.4126.0> 2011-05-10 14:02:05 =============================================================================== vbucketmigrator<0.4126.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.4126.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.4126.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.4126.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.4126.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.4126.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.4126.0>: Starting to move bucket 86 vbucketmigrator<0.4126.0>: Bucket 86 moved to the next server INFO REPORT <0.104.0> 2011-05-10 14:02:05 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.109.0> 2011-05-10 14:02:08 =============================================================================== ns_1@10.1.5.227:ns_orchestrator:164: Skipping janitor in state rebalancing: {rebalancing_state, <0.577.0>, {dict, 2, 16, 16, 8, 80, 48, {[], [], [], [], [], [], [], [], [], [], [], [], [], [], [], []}, {{[], [], [], [], [], [], [], [['ns_1@10.1.5.227'| 0.16796875]], [], [['ns_1@10.1.5.229'| 0.16796875]], [], [], [], [], [], []}}}} INFO REPORT <0.4126.0> 2011-05-10 14:02:10 =============================================================================== vbucketmigrator<0.4126.0>: Validate bucket states vbucketmigrator<0.4126.0>: 86 ok INFO REPORT <0.4136.0> 2011-05-10 14:02:11 =============================================================================== vbucketmigrator<0.4136.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.4136.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.4136.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.4136.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.4136.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.4136.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.4136.0>: Starting to move bucket 87 vbucketmigrator<0.4136.0>: Bucket 87 moved to the next server INFO REPORT <0.104.0> 2011-05-10 14:02:11 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.4136.0> 2011-05-10 14:02:16 =============================================================================== vbucketmigrator<0.4136.0>: Validate bucket states vbucketmigrator<0.4136.0>: 87 ok INFO REPORT <0.4144.0> 2011-05-10 14:02:17 =============================================================================== vbucketmigrator<0.4144.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.4144.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.4144.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.4144.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.4144.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.4144.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.4144.0>: Starting to move bucket 88 vbucketmigrator<0.4144.0>: Bucket 88 moved to the next server INFO REPORT <0.104.0> 2011-05-10 14:02:17 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.109.0> 2011-05-10 14:02:18 =============================================================================== ns_1@10.1.5.227:ns_orchestrator:164: Skipping janitor in state rebalancing: {rebalancing_state, <0.577.0>, {dict, 2, 16, 16, 8, 80, 48, {[], [], [], [], [], [], [], [], [], [], [], [], [], [], [], []}, {{[], [], [], [], [], [], [], [['ns_1@10.1.5.227'| 0.171875]], [], [['ns_1@10.1.5.229'| 0.171875]], [], [], [], [], [], []}}}} INFO REPORT <0.4144.0> 2011-05-10 14:02:22 =============================================================================== vbucketmigrator<0.4144.0>: Validate bucket states vbucketmigrator<0.4144.0>: 88 ok INFO REPORT <0.4149.0> 2011-05-10 14:02:23 =============================================================================== vbucketmigrator<0.4149.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.4149.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.4149.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.4149.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.4149.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.4149.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.4149.0>: Starting to move bucket 89 vbucketmigrator<0.4149.0>: Bucket 89 moved to the next server INFO REPORT <0.104.0> 2011-05-10 14:02:23 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.4149.0> 2011-05-10 14:02:28 =============================================================================== vbucketmigrator<0.4149.0>: Validate bucket states vbucketmigrator<0.4149.0>: 89 ok INFO REPORT <0.109.0> 2011-05-10 14:02:28 =============================================================================== ns_1@10.1.5.227:ns_orchestrator:164: Skipping janitor in state rebalancing: {rebalancing_state, <0.577.0>, {dict, 2, 16, 16, 8, 80, 48, {[], [], [], [], [], [], [], [], [], [], [], [], [], [], [], []}, {{[], [], [], [], [], [], [], [['ns_1@10.1.5.227'| 0.17578125]], [], [['ns_1@10.1.5.229'| 0.17578125]], [], [], [], [], [], []}}}} PROGRESS REPORT <0.34.0> 2011-05-10 14:02:29 =============================================================================== supervisor {local,sasl_sup} started [{pid,<0.4164.0>}, {name,rb_server}, {mfa,{rb,start_link, [[{report_dir,"/var/opt/membase/1.6.5.4r/logs"}, {start_log,"/opt/membase/1.6.5.4r/tmp/nslogs911378_16139.log"}, {type,all}, {max,all}, {report_dir}]]}}, {restart_type,temporary}, {shutdown,brutal_kill}, {child_type,worker}] INFO REPORT <0.4153.0> 2011-05-10 14:02:29 =============================================================================== vbucketmigrator<0.4153.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.4153.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.4153.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.4153.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.4153.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.4153.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.4153.0>: Starting to move bucket 90 vbucketmigrator<0.4153.0>: Bucket 90 moved to the next server INFO REPORT <0.104.0> 2011-05-10 14:02:29 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.4153.0> 2011-05-10 14:02:34 =============================================================================== vbucketmigrator<0.4153.0>: Validate bucket states vbucketmigrator<0.4153.0>: 90 ok INFO REPORT <0.5884.0> 2011-05-10 14:02:35 =============================================================================== vbucketmigrator<0.5884.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.5884.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.5884.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.5884.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.5884.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.5884.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.5884.0>: Starting to move bucket 91 vbucketmigrator<0.5884.0>: Bucket 91 moved to the next server INFO REPORT <0.104.0> 2011-05-10 14:02:35 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.109.0> 2011-05-10 14:02:38 =============================================================================== ns_1@10.1.5.227:ns_orchestrator:164: Skipping janitor in state rebalancing: {rebalancing_state, <0.577.0>, {dict, 2, 16, 16, 8, 80, 48, {[], [], [], [], [], [], [], [], [], [], [], [], [], [], [], []}, {{[], [], [], [], [], [], [], [['ns_1@10.1.5.227'| 0.177734375]], [], [['ns_1@10.1.5.229'| 0.177734375]], [], [], [], [], [], []}}}} INFO REPORT <0.5884.0> 2011-05-10 14:02:40 =============================================================================== vbucketmigrator<0.5884.0>: Validate bucket states vbucketmigrator<0.5884.0>: 91 ok INFO REPORT <0.5889.0> 2011-05-10 14:02:41 =============================================================================== vbucketmigrator<0.5889.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.5889.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.5889.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.5889.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.5889.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.5889.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.5889.0>: Starting to move bucket 92 vbucketmigrator<0.5889.0>: Bucket 92 moved to the next server INFO REPORT <0.104.0> 2011-05-10 14:02:41 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.474.0> 2011-05-10 14:02:43 =============================================================================== ns_1@10.1.5.227:ns_doctor:82: Current node statuses: [{'ns_1@10.1.5.227', [{last_heard,{1305,61362,411205}}, {active_buckets,["default"]}, {memory, [{total,52047464}, {processes,31633512}, {processes_used,31621736}, {system,20413952}, {atom,819441}, {atom_used,806779}, {binary,102352}, {code,8552890}, {ets,1325928}]}, {cluster_compatibility_version,1}, {version, [{kernel,"2.13.4"}, {mnesia,"4.4.12"}, {stdlib,"1.16.4"}, {os_mon,"2.2.4"}, {inets,"5.2"}, {ns_server,"1.6.5.4r"}, {menelaus,"1.6.5.4r"}, {sasl,"2.1.8"}]}, {system_arch,"x86_64-unknown-linux-gnu"}, {wall_clock,754}, {memory_data,{2058731520,672104448,{<0.12.0>,6657168}}}, {disk_data, [{"/",11204616,49},{"/boot",101086,14},{"/dev/shm",1005240,0}]}, {meminfo, <<"MemTotal: 2010480 kB\nMemFree: 1346688 kB\nBuffers: 63692 kB\nCached: 342144 kB\nSwapCached: 20084 kB\nActive: 353356 kB\nInactive: 197740 kB\nHighTotal: 0 kB\nHighFree: 0 kB\nLowTotal: 2010480 kB\nLowFree: 1346688 kB\nSwapTotal: 4030456 kB\nSwapFree: 3952288 kB\nDirty: 208 kB\nWriteback: 0 kB\nAnonPages: 143064 kB\nMapped: 15732 kB\nSlab: 86548 kB\nPageTables: 6632 kB\nNFS_Unstable: 0 kB\nBounce: 0 kB\nCommitLimit: 5035696 kB\nCommitted_AS: 563604 kB\nVmallocTotal: 34359738367 kB\nVmallocUsed: 265484 kB\nVmallocChunk: 34359471011 kB\nHugePages_Total: 0\nHugePages_Free: 0\nHugePages_Rsvd: 0\nHugepagesize: 2048 kB\n">>}, {replication,[{"default",1.0}]}, {system_memory_data, [{system_total_memory,2058731520}, {free_swap,4047142912}, {total_swap,4127186944}, {cached_memory,350355456}, {buffered_memory,65220608}, {free_memory,1379008512}, {total_memory,2058731520}]}, {statistics, [{wall_clock,{750579,1}}, {context_switches,{307602,0}}, {garbage_collection,{59495,261793591,0}}, {io,{{input,85141131},{output,13239347}}}, {reductions,{81439747,13410913}}, {run_queue,0}, {runtime,{11270,1570}}]}]}, {'ns_1@10.1.5.229', [{last_heard,{1305,61362,242313}}, {active_buckets,["default"]}, {memory, [{total,32609432}, {processes,13280352}, {processes_used,13256832}, {system,19329080}, {atom,778121}, {atom_used,752794}, {binary,223408}, {code,7791241}, {ets,1253152}]}, {cluster_compatibility_version,1}, {version, [{kernel,"2.13.4"}, {mnesia,"4.4.12"}, {stdlib,"1.16.4"}, {os_mon,"2.2.4"}, {ns_server,"1.6.5.4r"}, {menelaus,"1.6.5.4r"}, {sasl,"2.1.8"}]}, {system_arch,"x86_64-unknown-linux-gnu"}, {wall_clock,717}, {memory_data,{2058731520,712445952,{<9352.12.0>,4114680}}}, {disk_data, [{"/",11204616,44},{"/boot",101086,14},{"/dev/shm",1005240,0}]}, {meminfo, <<"MemTotal: 2010480 kB\nMemFree: 1313616 kB\nBuffers: 81636 kB\nCached: 340440 kB\nSwapCached: 0 kB\nActive: 364740 kB\nInactive: 214736 kB\nHighTotal: 0 kB\nHighFree: 0 kB\nLowTotal: 2010480 kB\nLowFree: 1313616 kB\nSwapTotal: 4030456 kB\nSwapFree: 4030388 kB\nDirty: 220 kB\nWriteback: 0 kB\nAnonPages: 157400 kB\nMapped: 28104 kB\nSlab: 91900 kB\nPageTables: 5868 kB\nNFS_Unstable: 0 kB\nBounce: 0 kB\nCommitLimit: 5035696 kB\nCommitted_AS: 478828 kB\nVmallocTotal: 34359738367 kB\nVmallocUsed: 265484 kB\nVmallocChunk: 34359471011 kB\nHugePages_Total: 0\nHugePages_Free: 0\nHugePages_Rsvd: 0\nHugepagesize: 2048 kB\n">>}, {replication,[{"default",1.0}]}, {system_memory_data, [{system_total_memory,2058731520}, {free_swap,4127117312}, {total_swap,4127186944}, {cached_memory,348610560}, {buffered_memory,83595264}, {free_memory,1345142784}, {total_memory,2058731520}]}, {statistics, [{wall_clock,{709806,1}}, {context_switches,{65371,0}}, {garbage_collection,{13977,52803544,0}}, {io,{{input,7806996},{output,3886952}}}, {reductions,{18364284,1308822}}, {run_queue,0}, {runtime,{2860,80}}]}]}] INFO REPORT <0.5889.0> 2011-05-10 14:02:46 =============================================================================== vbucketmigrator<0.5889.0>: Validate bucket states vbucketmigrator<0.5889.0>: 92 ok INFO REPORT <0.5908.0> 2011-05-10 14:02:47 =============================================================================== vbucketmigrator<0.5908.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.5908.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.5908.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.5908.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.5908.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.5908.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.5908.0>: Starting to move bucket 93 vbucketmigrator<0.5908.0>: Bucket 93 moved to the next server INFO REPORT <0.104.0> 2011-05-10 14:02:47 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.109.0> 2011-05-10 14:02:48 =============================================================================== ns_1@10.1.5.227:ns_orchestrator:164: Skipping janitor in state rebalancing: {rebalancing_state, <0.577.0>, {dict, 2, 16, 16, 8, 80, 48, {[], [], [], [], [], [], [], [], [], [], [], [], [], [], [], []}, {{[], [], [], [], [], [], [], [['ns_1@10.1.5.227'| 0.181640625]], [], [['ns_1@10.1.5.229'| 0.181640625]], [], [], [], [], [], []}}}} INFO REPORT <0.5908.0> 2011-05-10 14:02:52 =============================================================================== vbucketmigrator<0.5908.0>: Validate bucket states vbucketmigrator<0.5908.0>: 93 ok INFO REPORT <0.5921.0> 2011-05-10 14:02:53 =============================================================================== vbucketmigrator<0.5921.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.5921.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.5921.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.5921.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.5921.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.5921.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.5921.0>: Starting to move bucket 94 vbucketmigrator<0.5921.0>: Bucket 94 moved to the next server INFO REPORT <0.104.0> 2011-05-10 14:02:53 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.5921.0> 2011-05-10 14:02:58 =============================================================================== vbucketmigrator<0.5921.0>: Validate bucket states vbucketmigrator<0.5921.0>: 94 ok INFO REPORT <0.109.0> 2011-05-10 14:02:58 =============================================================================== ns_1@10.1.5.227:ns_orchestrator:164: Skipping janitor in state rebalancing: {rebalancing_state, <0.577.0>, {dict, 2, 16, 16, 8, 80, 48, {[], [], [], [], [], [], [], [], [], [], [], [], [], [], [], []}, {{[], [], [], [], [], [], [], [['ns_1@10.1.5.227'| 0.185546875]], [], [['ns_1@10.1.5.229'| 0.185546875]], [], [], [], [], [], []}}}} INFO REPORT <0.5925.0> 2011-05-10 14:02:59 =============================================================================== vbucketmigrator<0.5925.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.5925.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.5925.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.5925.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.5925.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.5925.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.5925.0>: Starting to move bucket 95 vbucketmigrator<0.5925.0>: Bucket 95 moved to the next server INFO REPORT <0.104.0> 2011-05-10 14:02:59 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.85.0> 2011-05-10 14:03:02 =============================================================================== Pulling config from: 'ns_1@10.1.5.229' INFO REPORT <0.5925.0> 2011-05-10 14:03:04 =============================================================================== vbucketmigrator<0.5925.0>: Validate bucket states vbucketmigrator<0.5925.0>: 95 ok INFO REPORT <0.5932.0> 2011-05-10 14:03:05 =============================================================================== vbucketmigrator<0.5932.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.5932.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.5932.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.5932.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.5932.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.5932.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.5932.0>: Starting to move bucket 96 vbucketmigrator<0.5932.0>: Bucket 96 moved to the next server INFO REPORT <0.104.0> 2011-05-10 14:03:05 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.109.0> 2011-05-10 14:03:08 =============================================================================== ns_1@10.1.5.227:ns_orchestrator:164: Skipping janitor in state rebalancing: {rebalancing_state, <0.577.0>, {dict, 2, 16, 16, 8, 80, 48, {[], [], [], [], [], [], [], [], [], [], [], [], [], [], [], []}, {{[], [], [], [], [], [], [], [['ns_1@10.1.5.227'| 0.1875]], [], [['ns_1@10.1.5.229'| 0.1875]], [], [], [], [], [], []}}}} INFO REPORT <0.5932.0> 2011-05-10 14:03:10 =============================================================================== vbucketmigrator<0.5932.0>: Validate bucket states vbucketmigrator<0.5932.0>: 96 ok INFO REPORT <0.5942.0> 2011-05-10 14:03:11 =============================================================================== vbucketmigrator<0.5942.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.5942.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.5942.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.5942.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.5942.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.5942.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.5942.0>: Starting to move bucket 97 vbucketmigrator<0.5942.0>: Bucket 97 moved to the next server INFO REPORT <0.104.0> 2011-05-10 14:03:11 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.5942.0> 2011-05-10 14:03:16 =============================================================================== vbucketmigrator<0.5942.0>: Validate bucket states vbucketmigrator<0.5942.0>: 97 ok INFO REPORT <0.5950.0> 2011-05-10 14:03:17 =============================================================================== vbucketmigrator<0.5950.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.5950.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.5950.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.5950.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.5950.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.5950.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.5950.0>: Starting to move bucket 98 vbucketmigrator<0.5950.0>: Bucket 98 moved to the next server INFO REPORT <0.104.0> 2011-05-10 14:03:17 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.109.0> 2011-05-10 14:03:18 =============================================================================== ns_1@10.1.5.227:ns_orchestrator:164: Skipping janitor in state rebalancing: {rebalancing_state, <0.577.0>, {dict, 2, 16, 16, 8, 80, 48, {[], [], [], [], [], [], [], [], [], [], [], [], [], [], [], []}, {{[], [], [], [], [], [], [], [['ns_1@10.1.5.227'| 0.19140625]], [], [['ns_1@10.1.5.229'| 0.19140625]], [], [], [], [], [], []}}}} INFO REPORT <0.5950.0> 2011-05-10 14:03:22 =============================================================================== vbucketmigrator<0.5950.0>: Validate bucket states vbucketmigrator<0.5950.0>: 98 ok INFO REPORT <0.5955.0> 2011-05-10 14:03:23 =============================================================================== vbucketmigrator<0.5955.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.5955.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.5955.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.5955.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.5955.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.5955.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.5955.0>: Starting to move bucket 99 vbucketmigrator<0.5955.0>: Bucket 99 moved to the next server INFO REPORT <0.104.0> 2011-05-10 14:03:23 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.5955.0> 2011-05-10 14:03:28 =============================================================================== vbucketmigrator<0.5955.0>: Validate bucket states vbucketmigrator<0.5955.0>: 99 ok INFO REPORT <0.109.0> 2011-05-10 14:03:28 =============================================================================== ns_1@10.1.5.227:ns_orchestrator:164: Skipping janitor in state rebalancing: {rebalancing_state, <0.577.0>, {dict, 2, 16, 16, 8, 80, 48, {[], [], [], [], [], [], [], [], [], [], [], [], [], [], [], []}, {{[], [], [], [], [], [], [], [['ns_1@10.1.5.227'| 0.1953125]], [], [['ns_1@10.1.5.229'| 0.1953125]], [], [], [], [], [], []}}}} INFO REPORT <0.5959.0> 2011-05-10 14:03:29 =============================================================================== vbucketmigrator<0.5959.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.5959.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.5959.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.5959.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.5959.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.5959.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.5959.0>: Starting to move bucket 100 vbucketmigrator<0.5959.0>: Bucket 100 moved to the next server INFO REPORT <0.104.0> 2011-05-10 14:03:29 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.5959.0> 2011-05-10 14:03:34 =============================================================================== vbucketmigrator<0.5959.0>: Validate bucket states vbucketmigrator<0.5959.0>: 100 ok INFO REPORT <0.5974.0> 2011-05-10 14:03:35 =============================================================================== vbucketmigrator<0.5974.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.5974.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.5974.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.5974.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.5974.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.5974.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.5974.0>: Starting to move bucket 101 vbucketmigrator<0.5974.0>: Bucket 101 moved to the next server INFO REPORT <0.104.0> 2011-05-10 14:03:35 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.109.0> 2011-05-10 14:03:38 =============================================================================== ns_1@10.1.5.227:ns_orchestrator:164: Skipping janitor in state rebalancing: {rebalancing_state, <0.577.0>, {dict, 2, 16, 16, 8, 80, 48, {[], [], [], [], [], [], [], [], [], [], [], [], [], [], [], []}, {{[], [], [], [], [], [], [], [['ns_1@10.1.5.227'| 0.197265625]], [], [['ns_1@10.1.5.229'| 0.197265625]], [], [], [], [], [], []}}}} INFO REPORT <0.5974.0> 2011-05-10 14:03:40 =============================================================================== vbucketmigrator<0.5974.0>: Validate bucket states vbucketmigrator<0.5974.0>: 101 ok INFO REPORT <0.85.0> 2011-05-10 14:03:40 =============================================================================== Pulling config from: 'ns_1@10.1.5.229' INFO REPORT <0.5979.0> 2011-05-10 14:03:41 =============================================================================== vbucketmigrator<0.5979.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.5979.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.5979.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.5979.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.5979.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.5979.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.5979.0>: Starting to move bucket 102 vbucketmigrator<0.5979.0>: Bucket 102 moved to the next server INFO REPORT <0.104.0> 2011-05-10 14:03:41 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.474.0> 2011-05-10 14:03:43 =============================================================================== ns_1@10.1.5.227:ns_doctor:82: Current node statuses: [{'ns_1@10.1.5.227', [{last_heard,{1305,61422,411231}}, {active_buckets,["default"]}, {memory, [{total,39834936}, {processes,19352144}, {processes_used,19339504}, {system,20482792}, {atom,819441}, {atom_used,806779}, {binary,106544}, {code,8552890}, {ets,1389296}]}, {cluster_compatibility_version,1}, {version, [{kernel,"2.13.4"}, {mnesia,"4.4.12"}, {stdlib,"1.16.4"}, {os_mon,"2.2.4"}, {inets,"5.2"}, {ns_server,"1.6.5.4r"}, {menelaus,"1.6.5.4r"}, {sasl,"2.1.8"}]}, {system_arch,"x86_64-unknown-linux-gnu"}, {wall_clock,814}, {memory_data,{2058731520,678580224,{<0.3948.0>,12342384}}}, {disk_data, [{"/",11204616,49},{"/boot",101086,14},{"/dev/shm",1005240,0}]}, {meminfo, <<"MemTotal: 2010480 kB\nMemFree: 1351648 kB\nBuffers: 63984 kB\nCached: 342388 kB\nSwapCached: 20084 kB\nActive: 348060 kB\nInactive: 198304 kB\nHighTotal: 0 kB\nHighFree: 0 kB\nLowTotal: 2010480 kB\nLowFree: 1351648 kB\nSwapTotal: 4030456 kB\nSwapFree: 3952288 kB\nDirty: 128 kB\nWriteback: 0 kB\nAnonPages: 137796 kB\nMapped: 15732 kB\nSlab: 86548 kB\nPageTables: 6608 kB\nNFS_Unstable: 0 kB\nBounce: 0 kB\nCommitLimit: 5035696 kB\nCommitted_AS: 549040 kB\nVmallocTotal: 34359738367 kB\nVmallocUsed: 265484 kB\nVmallocChunk: 34359471011 kB\nHugePages_Total: 0\nHugePages_Free: 0\nHugePages_Rsvd: 0\nHugepagesize: 2048 kB\n">>}, {replication,[{"default",1.0}]}, {system_memory_data, [{system_total_memory,2058731520}, {free_swap,4047142912}, {total_swap,4127186944}, {cached_memory,350605312}, {buffered_memory,65519616}, {free_memory,1384087552}, {total_memory,2058731520}]}, {statistics, [{wall_clock,{810580,2}}, {context_switches,{317260,0}}, {garbage_collection,{61847,278592611,0}}, {io,{{input,86026040},{output,14340045}}}, {reductions,{85654992,674578}}, {run_queue,0}, {runtime,{11950,100}}]}]}, {'ns_1@10.1.5.229', [{last_heard,{1305,61422,242211}}, {active_buckets,["default"]}, {memory, [{total,32700000}, {processes,13312648}, {processes_used,13289128}, {system,19387352}, {atom,778121}, {atom_used,752794}, {binary,216984}, {code,7791241}, {ets,1316352}]}, {cluster_compatibility_version,1}, {version, [{kernel,"2.13.4"}, {mnesia,"4.4.12"}, {stdlib,"1.16.4"}, {os_mon,"2.2.4"}, {ns_server,"1.6.5.4r"}, {menelaus,"1.6.5.4r"}, {sasl,"2.1.8"}]}, {system_arch,"x86_64-unknown-linux-gnu"}, {wall_clock,777}, {memory_data,{2058731520,713969664,{<9352.12.0>,4114680}}}, {disk_data, [{"/",11204616,44},{"/boot",101086,14},{"/dev/shm",1005240,0}]}, {meminfo, <<"MemTotal: 2010480 kB\nMemFree: 1312748 kB\nBuffers: 82416 kB\nCached: 340692 kB\nSwapCached: 0 kB\nActive: 364860 kB\nInactive: 215736 kB\nHighTotal: 0 kB\nHighFree: 0 kB\nLowTotal: 2010480 kB\nLowFree: 1312748 kB\nSwapTotal: 4030456 kB\nSwapFree: 4030388 kB\nDirty: 216 kB\nWriteback: 0 kB\nAnonPages: 157500 kB\nMapped: 28104 kB\nSlab: 91908 kB\nPageTables: 5868 kB\nNFS_Unstable: 0 kB\nBounce: 0 kB\nCommitLimit: 5035696 kB\nCommitted_AS: 479852 kB\nVmallocTotal: 34359738367 kB\nVmallocUsed: 265484 kB\nVmallocChunk: 34359471011 kB\nHugePages_Total: 0\nHugePages_Free: 0\nHugePages_Rsvd: 0\nHugepagesize: 2048 kB\n">>}, {replication,[{"default",1.0}]}, {system_memory_data, [{system_total_memory,2058731520}, {free_swap,4127117312}, {total_swap,4127186944}, {cached_memory,348868608}, {buffered_memory,84393984}, {free_memory,1344253952}, {total_memory,2058731520}]}, {statistics, [{wall_clock,{769806,1}}, {context_switches,{70572,0}}, {garbage_collection,{15291,57461072,0}}, {io,{{input,8798961},{output,4865527}}}, {reductions,{19782306,237930}}, {run_queue,0}, {runtime,{3150,40}}]}]}] INFO REPORT <0.469.0> 2011-05-10 14:03:44 =============================================================================== ns_1@10.1.5.227:stats_collector:71: Stats for bucket "default": auth_cmds 135 auth_errors 0 bucket_conns 34 bytes_read 1614288 bytes_written 6878113 cas_badval 0 cas_hits 0 cas_misses 0 cmd_flush 1 cmd_get 0 cmd_set 16004 conn_yields 0 connection_structures 44 curr_connections 44 curr_items 7226 curr_items_tot 8040 daemon_connections 10 decr_hits 0 decr_misses 0 delete_hits 0 delete_misses 0 ep_bg_fetched 0 ep_commit_num 53 ep_commit_time 0 ep_commit_time_total 4 ep_data_age 13 ep_data_age_highwat 13 ep_db_cleaner_status complete ep_db_strategy multiMTVBDB ep_dbinit 1 ep_dbname /var/opt/membase/1.6.5.4r/data/ns_1/default ep_dbshards 4 ep_expired 0 ep_flush_duration 1 ep_flush_duration_highwat 8 ep_flush_duration_total 12 ep_flush_preempts 0 ep_flusher_state running ep_flusher_todo 0 ep_io_num_read 814 ep_io_num_write 15741 ep_io_read_bytes 46128 ep_io_write_bytes 1011432 ep_item_begin_failed 0 ep_item_commit_failed 0 ep_item_flush_expired 0 ep_item_flush_failed 0 ep_kv_size 1181400 ep_max_data_size 1096810496 ep_max_txn_size 1000 ep_mem_high_wat 822607872 ep_mem_low_wat 658086297 ep_min_data_age 0 ep_num_active_non_resident 0 ep_num_eject_failures 0 ep_num_eject_replicas 0 ep_num_expiry_pager_runs 0 ep_num_non_resident 0 ep_num_not_my_vbuckets 0 ep_num_pager_runs 0 ep_num_value_ejects 0 ep_oom_errors 0 ep_overhead 25854448 ep_pending_ops 0 ep_pending_ops_max 0 ep_pending_ops_max_duration 0 ep_pending_ops_total 0 ep_queue_age_cap 900 ep_queue_size 0 ep_storage_age 0 ep_storage_age_highwat 10 ep_storage_type featured ep_store_max_concurrency 10 ep_store_max_readers 9 ep_store_max_readwrite 1 ep_tap_bg_fetch_requeued 0 ep_tap_bg_fetched 814 ep_tap_keepalive 300 ep_tmp_oom_errors 0 ep_too_old 0 ep_too_young 0 ep_total_cache_size 1181400 ep_total_del_items 0 ep_total_enqueued 16035 ep_total_new_items 15741 ep_total_persisted 15741 ep_vbucket_del 0 ep_vbucket_del_fail 0 ep_version 1.6.5.3.1_13_g81a6cf1 ep_warmed_up 0 ep_warmup true ep_warmup_dups 0 ep_warmup_oom 0 ep_warmup_thread complete ep_warmup_time 15662 get_hits 0 get_misses 0 incr_hits 0 incr_misses 0 libevent 1.4.13-stable limit_maxbytes 67108864 mem_used 27035848 pid 16183 pointer_size 64 rejected_conns 0 rusage_system 1.328797 rusage_user 3.419480 tap_connect_received 103 tap_mutation_sent 1628 tap_opaque_sent 103 tap_vbucket_set_sent 206 threads 4 time 1305061423 total_connections 146 uptime 817 version 1.4.4_382_g9df3289 INFO REPORT <0.5979.0> 2011-05-10 14:03:46 =============================================================================== vbucketmigrator<0.5979.0>: Validate bucket states vbucketmigrator<0.5979.0>: 102 ok INFO REPORT <0.5983.0> 2011-05-10 14:03:47 =============================================================================== vbucketmigrator<0.5983.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.5983.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.5983.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.5983.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.5983.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.5983.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.5983.0>: Starting to move bucket 103 vbucketmigrator<0.5983.0>: Bucket 103 moved to the next server INFO REPORT <0.104.0> 2011-05-10 14:03:47 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.109.0> 2011-05-10 14:03:48 =============================================================================== ns_1@10.1.5.227:ns_orchestrator:164: Skipping janitor in state rebalancing: {rebalancing_state, <0.577.0>, {dict, 2, 16, 16, 8, 80, 48, {[], [], [], [], [], [], [], [], [], [], [], [], [], [], [], []}, {{[], [], [], [], [], [], [], [['ns_1@10.1.5.227'| 0.201171875]], [], [['ns_1@10.1.5.229'| 0.201171875]], [], [], [], [], [], []}}}} INFO REPORT <0.5983.0> 2011-05-10 14:03:52 =============================================================================== vbucketmigrator<0.5983.0>: Validate bucket states vbucketmigrator<0.5983.0>: 103 ok INFO REPORT <0.5996.0> 2011-05-10 14:03:53 =============================================================================== vbucketmigrator<0.5996.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.5996.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.5996.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.5996.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.5996.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.5996.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.5996.0>: Starting to move bucket 104 vbucketmigrator<0.5996.0>: Bucket 104 moved to the next server INFO REPORT <0.104.0> 2011-05-10 14:03:53 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.5996.0> 2011-05-10 14:03:58 =============================================================================== vbucketmigrator<0.5996.0>: Validate bucket states vbucketmigrator<0.5996.0>: 104 ok INFO REPORT <0.109.0> 2011-05-10 14:03:58 =============================================================================== ns_1@10.1.5.227:ns_orchestrator:164: Skipping janitor in state rebalancing: {rebalancing_state, <0.577.0>, {dict, 2, 16, 16, 8, 80, 48, {[], [], [], [], [], [], [], [], [], [], [], [], [], [], [], []}, {{[], [], [], [], [], [], [], [['ns_1@10.1.5.227'| 0.205078125]], [], [['ns_1@10.1.5.229'| 0.205078125]], [], [], [], [], [], []}}}} INFO REPORT <0.6000.0> 2011-05-10 14:03:59 =============================================================================== vbucketmigrator<0.6000.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.6000.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6000.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6000.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.6000.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6000.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6000.0>: Starting to move bucket 105 vbucketmigrator<0.6000.0>: Bucket 105 moved to the next server INFO REPORT <0.104.0> 2011-05-10 14:03:59 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.6000.0> 2011-05-10 14:04:04 =============================================================================== vbucketmigrator<0.6000.0>: Validate bucket states vbucketmigrator<0.6000.0>: 105 ok INFO REPORT <0.6007.0> 2011-05-10 14:04:05 =============================================================================== vbucketmigrator<0.6007.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.6007.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6007.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6007.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.6007.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6007.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6007.0>: Starting to move bucket 106 vbucketmigrator<0.6007.0>: Bucket 106 moved to the next server INFO REPORT <0.104.0> 2011-05-10 14:04:05 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.109.0> 2011-05-10 14:04:08 =============================================================================== ns_1@10.1.5.227:ns_orchestrator:164: Skipping janitor in state rebalancing: {rebalancing_state, <0.577.0>, {dict, 2, 16, 16, 8, 80, 48, {[], [], [], [], [], [], [], [], [], [], [], [], [], [], [], []}, {{[], [], [], [], [], [], [], [['ns_1@10.1.5.227'| 0.20703125]], [], [['ns_1@10.1.5.229'| 0.20703125]], [], [], [], [], [], []}}}} INFO REPORT <0.6007.0> 2011-05-10 14:04:10 =============================================================================== vbucketmigrator<0.6007.0>: Validate bucket states vbucketmigrator<0.6007.0>: 106 ok INFO REPORT <0.6018.0> 2011-05-10 14:04:11 =============================================================================== vbucketmigrator<0.6018.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.6018.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6018.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6018.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.6018.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6018.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6018.0>: Starting to move bucket 107 vbucketmigrator<0.6018.0>: Bucket 107 moved to the next server INFO REPORT <0.104.0> 2011-05-10 14:04:11 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.6018.0> 2011-05-10 14:04:16 =============================================================================== vbucketmigrator<0.6018.0>: Validate bucket states vbucketmigrator<0.6018.0>: 107 ok INFO REPORT <0.6026.0> 2011-05-10 14:04:17 =============================================================================== vbucketmigrator<0.6026.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.6026.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6026.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6026.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.6026.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6026.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6026.0>: Starting to move bucket 108 vbucketmigrator<0.6026.0>: Bucket 108 moved to the next server INFO REPORT <0.104.0> 2011-05-10 14:04:17 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.109.0> 2011-05-10 14:04:18 =============================================================================== ns_1@10.1.5.227:ns_orchestrator:164: Skipping janitor in state rebalancing: {rebalancing_state, <0.577.0>, {dict, 2, 16, 16, 8, 80, 48, {[], [], [], [], [], [], [], [], [], [], [], [], [], [], [], []}, {{[], [], [], [], [], [], [], [['ns_1@10.1.5.227'| 0.2109375]], [], [['ns_1@10.1.5.229'| 0.2109375]], [], [], [], [], [], []}}}} INFO REPORT <0.85.0> 2011-05-10 14:04:22 =============================================================================== Pulling config from: 'ns_1@10.1.5.229' INFO REPORT <0.6026.0> 2011-05-10 14:04:22 =============================================================================== vbucketmigrator<0.6026.0>: Validate bucket states vbucketmigrator<0.6026.0>: 108 ok INFO REPORT <0.6031.0> 2011-05-10 14:04:23 =============================================================================== vbucketmigrator<0.6031.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.6031.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6031.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6031.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.6031.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6031.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6031.0>: Starting to move bucket 109 vbucketmigrator<0.6031.0>: Bucket 109 moved to the next server INFO REPORT <0.104.0> 2011-05-10 14:04:23 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.6031.0> 2011-05-10 14:04:28 =============================================================================== vbucketmigrator<0.6031.0>: Validate bucket states vbucketmigrator<0.6031.0>: 109 ok INFO REPORT <0.109.0> 2011-05-10 14:04:28 =============================================================================== ns_1@10.1.5.227:ns_orchestrator:164: Skipping janitor in state rebalancing: {rebalancing_state, <0.577.0>, {dict, 2, 16, 16, 8, 80, 48, {[], [], [], [], [], [], [], [], [], [], [], [], [], [], [], []}, {{[], [], [], [], [], [], [], [['ns_1@10.1.5.227'| 0.21484375]], [], [['ns_1@10.1.5.229'| 0.21484375]], [], [], [], [], [], []}}}} INFO REPORT <0.6035.0> 2011-05-10 14:04:29 =============================================================================== vbucketmigrator<0.6035.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.6035.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6035.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6035.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.6035.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6035.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6035.0>: Starting to move bucket 110 vbucketmigrator<0.6035.0>: Bucket 110 moved to the next server INFO REPORT <0.104.0> 2011-05-10 14:04:29 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.6035.0> 2011-05-10 14:04:34 =============================================================================== vbucketmigrator<0.6035.0>: Validate bucket states vbucketmigrator<0.6035.0>: 110 ok INFO REPORT <0.6050.0> 2011-05-10 14:04:35 =============================================================================== vbucketmigrator<0.6050.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.6050.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6050.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6050.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.6050.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6050.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6050.0>: Starting to move bucket 111 vbucketmigrator<0.6050.0>: Bucket 111 moved to the next server INFO REPORT <0.104.0> 2011-05-10 14:04:35 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.109.0> 2011-05-10 14:04:38 =============================================================================== ns_1@10.1.5.227:ns_orchestrator:164: Skipping janitor in state rebalancing: {rebalancing_state, <0.577.0>, {dict, 2, 16, 16, 8, 80, 48, {[], [], [], [], [], [], [], [], [], [], [], [], [], [], [], []}, {{[], [], [], [], [], [], [], [['ns_1@10.1.5.227'| 0.216796875]], [], [['ns_1@10.1.5.229'| 0.216796875]], [], [], [], [], [], []}}}} INFO REPORT <0.6050.0> 2011-05-10 14:04:40 =============================================================================== vbucketmigrator<0.6050.0>: Validate bucket states vbucketmigrator<0.6050.0>: 111 ok INFO REPORT <0.6055.0> 2011-05-10 14:04:41 =============================================================================== vbucketmigrator<0.6055.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.6055.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6055.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6055.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.6055.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6055.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6055.0>: Starting to move bucket 112 vbucketmigrator<0.6055.0>: Bucket 112 moved to the next server INFO REPORT <0.104.0> 2011-05-10 14:04:41 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.474.0> 2011-05-10 14:04:43 =============================================================================== ns_1@10.1.5.227:ns_doctor:82: Current node statuses: [{'ns_1@10.1.5.227', [{last_heard,{1305,61482,410641}}, {active_buckets,["default"]}, {memory, [{total,39917528}, {processes,19455520}, {processes_used,19443744}, {system,20462008}, {atom,819441}, {atom_used,806779}, {binary,88640}, {code,8552890}, {ets,1384584}]}, {cluster_compatibility_version,1}, {version, [{kernel,"2.13.4"}, {mnesia,"4.4.12"}, {stdlib,"1.16.4"}, {os_mon,"2.2.4"}, {inets,"5.2"}, {ns_server,"1.6.5.4r"}, {menelaus,"1.6.5.4r"}, {sasl,"2.1.8"}]}, {system_arch,"x86_64-unknown-linux-gnu"}, {wall_clock,874}, {memory_data,{2058731520,674643968,{<0.12.0>,6657168}}}, {disk_data, [{"/",11204616,49},{"/boot",101086,14},{"/dev/shm",1005240,0}]}, {meminfo, <<"MemTotal: 2010480 kB\nMemFree: 1350904 kB\nBuffers: 64240 kB\nCached: 342648 kB\nSwapCached: 20084 kB\nActive: 348420 kB\nInactive: 198680 kB\nHighTotal: 0 kB\nHighFree: 0 kB\nLowTotal: 2010480 kB\nLowFree: 1350904 kB\nSwapTotal: 4030456 kB\nSwapFree: 3952288 kB\nDirty: 228 kB\nWriteback: 0 kB\nAnonPages: 138000 kB\nMapped: 15732 kB\nSlab: 86536 kB\nPageTables: 6608 kB\nNFS_Unstable: 0 kB\nBounce: 0 kB\nCommitLimit: 5035696 kB\nCommitted_AS: 549040 kB\nVmallocTotal: 34359738367 kB\nVmallocUsed: 265484 kB\nVmallocChunk: 34359471011 kB\nHugePages_Total: 0\nHugePages_Free: 0\nHugePages_Rsvd: 0\nHugepagesize: 2048 kB\n">>}, {replication,[{"default",1.0}]}, {system_memory_data, [{system_total_memory,2058731520}, {free_swap,4047142912}, {total_swap,4127186944}, {cached_memory,350871552}, {buffered_memory,65781760}, {free_memory,1383325696}, {total_memory,2058731520}]}, {statistics, [{wall_clock,{870579,1}}, {context_switches,{326062,0}}, {garbage_collection,{64083,293752528,0}}, {io,{{input,86170337},{output,14711006}}}, {reductions,{89662579,787599}}, {run_queue,0}, {runtime,{12530,140}}]}]}, {'ns_1@10.1.5.229', [{last_heard,{1305,61482,242073}}, {active_buckets,["default"]}, {memory, [{total,32695608}, {processes,13315456}, {processes_used,13291936}, {system,19380152}, {atom,778121}, {atom_used,752794}, {binary,213096}, {code,7791241}, {ets,1312520}]}, {cluster_compatibility_version,1}, {version, [{kernel,"2.13.4"}, {mnesia,"4.4.12"}, {stdlib,"1.16.4"}, {os_mon,"2.2.4"}, {ns_server,"1.6.5.4r"}, {menelaus,"1.6.5.4r"}, {sasl,"2.1.8"}]}, {system_arch,"x86_64-unknown-linux-gnu"}, {wall_clock,837}, {memory_data,{2058731520,715112448,{<9352.12.0>,4114680}}}, {disk_data, [{"/",11204616,44},{"/boot",101086,14},{"/dev/shm",1005240,0}]}, {meminfo, <<"MemTotal: 2010480 kB\nMemFree: 1311260 kB\nBuffers: 83084 kB\nCached: 340972 kB\nSwapCached: 0 kB\nActive: 366100 kB\nInactive: 216320 kB\nHighTotal: 0 kB\nHighFree: 0 kB\nLowTotal: 2010480 kB\nLowFree: 1311260 kB\nSwapTotal: 4030456 kB\nSwapFree: 4030388 kB\nDirty: 112 kB\nWriteback: 0 kB\nAnonPages: 158380 kB\nMapped: 28104 kB\nSlab: 91900 kB\nPageTables: 5872 kB\nNFS_Unstable: 0 kB\nBounce: 0 kB\nCommitLimit: 5035696 kB\nCommitted_AS: 484736 kB\nVmallocTotal: 34359738367 kB\nVmallocUsed: 265484 kB\nVmallocChunk: 34359471011 kB\nHugePages_Total: 0\nHugePages_Free: 0\nHugePages_Rsvd: 0\nHugepagesize: 2048 kB\n">>}, {replication,[{"default",1.0}]}, {system_memory_data, [{system_total_memory,2058731520}, {free_swap,4127117312}, {total_swap,4127186944}, {cached_memory,349155328}, {buffered_memory,85078016}, {free_memory,1342730240}, {total_memory,2058731520}]}, {statistics, [{wall_clock,{829806,1}}, {context_switches,{75348,0}}, {garbage_collection,{16346,61898569,0}}, {io,{{input,9005641},{output,5123573}}}, {reductions,{21145845,231489}}, {run_queue,0}, {runtime,{3340,30}}]}]}] INFO REPORT <0.6055.0> 2011-05-10 14:04:46 =============================================================================== vbucketmigrator<0.6055.0>: Validate bucket states vbucketmigrator<0.6055.0>: 112 ok INFO REPORT <0.6059.0> 2011-05-10 14:04:47 =============================================================================== vbucketmigrator<0.6059.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.6059.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6059.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6059.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.6059.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6059.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6059.0>: Starting to move bucket 113 vbucketmigrator<0.6059.0>: Bucket 113 moved to the next server INFO REPORT <0.104.0> 2011-05-10 14:04:47 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.109.0> 2011-05-10 14:04:48 =============================================================================== ns_1@10.1.5.227:ns_orchestrator:164: Skipping janitor in state rebalancing: {rebalancing_state, <0.577.0>, {dict, 2, 16, 16, 8, 80, 48, {[], [], [], [], [], [], [], [], [], [], [], [], [], [], [], []}, {{[], [], [], [], [], [], [], [['ns_1@10.1.5.227'| 0.220703125]], [], [['ns_1@10.1.5.229'| 0.220703125]], [], [], [], [], [], []}}}} INFO REPORT <0.6059.0> 2011-05-10 14:04:52 =============================================================================== vbucketmigrator<0.6059.0>: Validate bucket states vbucketmigrator<0.6059.0>: 113 ok INFO REPORT <0.6072.0> 2011-05-10 14:04:53 =============================================================================== vbucketmigrator<0.6072.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.6072.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6072.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6072.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.6072.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6072.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6072.0>: Starting to move bucket 114 vbucketmigrator<0.6072.0>: Bucket 114 moved to the next server INFO REPORT <0.104.0> 2011-05-10 14:04:53 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.6072.0> 2011-05-10 14:04:58 =============================================================================== vbucketmigrator<0.6072.0>: Validate bucket states vbucketmigrator<0.6072.0>: 114 ok INFO REPORT <0.109.0> 2011-05-10 14:04:58 =============================================================================== ns_1@10.1.5.227:ns_orchestrator:164: Skipping janitor in state rebalancing: {rebalancing_state, <0.577.0>, {dict, 2, 16, 16, 8, 80, 48, {[], [], [], [], [], [], [], [], [], [], [], [], [], [], [], []}, {{[], [], [], [], [], [], [], [['ns_1@10.1.5.227'| 0.224609375]], [], [['ns_1@10.1.5.229'| 0.224609375]], [], [], [], [], [], []}}}} INFO REPORT <0.6076.0> 2011-05-10 14:04:59 =============================================================================== vbucketmigrator<0.6076.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.6076.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6076.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6076.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.6076.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6076.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6076.0>: Starting to move bucket 115 vbucketmigrator<0.6076.0>: Bucket 115 moved to the next server INFO REPORT <0.104.0> 2011-05-10 14:04:59 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.6076.0> 2011-05-10 14:05:04 =============================================================================== vbucketmigrator<0.6076.0>: Validate bucket states vbucketmigrator<0.6076.0>: 115 ok INFO REPORT <0.6083.0> 2011-05-10 14:05:05 =============================================================================== vbucketmigrator<0.6083.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.6083.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6083.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6083.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.6083.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6083.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6083.0>: Starting to move bucket 116 vbucketmigrator<0.6083.0>: Bucket 116 moved to the next server INFO REPORT <0.104.0> 2011-05-10 14:05:05 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.109.0> 2011-05-10 14:05:08 =============================================================================== ns_1@10.1.5.227:ns_orchestrator:164: Skipping janitor in state rebalancing: {rebalancing_state, <0.577.0>, {dict, 2, 16, 16, 8, 80, 48, {[], [], [], [], [], [], [], [], [], [], [], [], [], [], [], []}, {{[], [], [], [], [], [], [], [['ns_1@10.1.5.227'| 0.2265625]], [], [['ns_1@10.1.5.229'| 0.2265625]], [], [], [], [], [], []}}}} INFO REPORT <0.6083.0> 2011-05-10 14:05:10 =============================================================================== vbucketmigrator<0.6083.0>: Validate bucket states vbucketmigrator<0.6083.0>: 116 ok INFO REPORT <0.6093.0> 2011-05-10 14:05:11 =============================================================================== vbucketmigrator<0.6093.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.6093.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6093.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6093.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.6093.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6093.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6093.0>: Starting to move bucket 117 vbucketmigrator<0.6093.0>: Bucket 117 moved to the next server INFO REPORT <0.104.0> 2011-05-10 14:05:11 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.85.0> 2011-05-10 14:05:16 =============================================================================== Pulling config from: 'ns_1@10.1.5.229' INFO REPORT <0.6093.0> 2011-05-10 14:05:16 =============================================================================== vbucketmigrator<0.6093.0>: Validate bucket states vbucketmigrator<0.6093.0>: 117 ok INFO REPORT <0.6102.0> 2011-05-10 14:05:17 =============================================================================== vbucketmigrator<0.6102.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.6102.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6102.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6102.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.6102.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6102.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6102.0>: Starting to move bucket 118 vbucketmigrator<0.6102.0>: Bucket 118 moved to the next server INFO REPORT <0.104.0> 2011-05-10 14:05:17 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.109.0> 2011-05-10 14:05:18 =============================================================================== ns_1@10.1.5.227:ns_orchestrator:164: Skipping janitor in state rebalancing: {rebalancing_state, <0.577.0>, {dict, 2, 16, 16, 8, 80, 48, {[], [], [], [], [], [], [], [], [], [], [], [], [], [], [], []}, {{[], [], [], [], [], [], [], [['ns_1@10.1.5.227'| 0.23046875]], [], [['ns_1@10.1.5.229'| 0.23046875]], [], [], [], [], [], []}}}} INFO REPORT <0.6102.0> 2011-05-10 14:05:22 =============================================================================== vbucketmigrator<0.6102.0>: Validate bucket states vbucketmigrator<0.6102.0>: 118 ok INFO REPORT <0.6107.0> 2011-05-10 14:05:23 =============================================================================== vbucketmigrator<0.6107.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.6107.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6107.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6107.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.6107.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6107.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6107.0>: Starting to move bucket 119 vbucketmigrator<0.6107.0>: Bucket 119 moved to the next server INFO REPORT <0.104.0> 2011-05-10 14:05:23 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.469.0> 2011-05-10 14:05:24 =============================================================================== ns_1@10.1.5.227:stats_collector:71: Stats for bucket "default": auth_cmds 152 auth_errors 0 bucket_conns 34 bytes_read 1621139 bytes_written 7921557 cas_badval 0 cas_hits 0 cas_misses 0 cmd_flush 1 cmd_get 0 cmd_set 16004 conn_yields 0 connection_structures 44 curr_connections 44 curr_items 7077 curr_items_tot 8040 daemon_connections 10 decr_hits 0 decr_misses 0 delete_hits 0 delete_misses 0 ep_bg_fetched 0 ep_commit_num 53 ep_commit_time 0 ep_commit_time_total 4 ep_data_age 13 ep_data_age_highwat 13 ep_db_cleaner_status complete ep_db_strategy multiMTVBDB ep_dbinit 1 ep_dbname /var/opt/membase/1.6.5.4r/data/ns_1/default ep_dbshards 4 ep_expired 0 ep_flush_duration 1 ep_flush_duration_highwat 8 ep_flush_duration_total 12 ep_flush_preempts 0 ep_flusher_state running ep_flusher_todo 0 ep_io_num_read 963 ep_io_num_write 15741 ep_io_read_bytes 54632 ep_io_write_bytes 1011432 ep_item_begin_failed 0 ep_item_commit_failed 0 ep_item_flush_expired 0 ep_item_flush_failed 0 ep_kv_size 1181400 ep_max_data_size 1096810496 ep_max_txn_size 1000 ep_mem_high_wat 822607872 ep_mem_low_wat 658086297 ep_min_data_age 0 ep_num_active_non_resident 0 ep_num_eject_failures 0 ep_num_eject_replicas 0 ep_num_expiry_pager_runs 0 ep_num_non_resident 0 ep_num_not_my_vbuckets 0 ep_num_pager_runs 0 ep_num_value_ejects 0 ep_oom_errors 0 ep_overhead 25854448 ep_pending_ops 0 ep_pending_ops_max 0 ep_pending_ops_max_duration 0 ep_pending_ops_total 0 ep_queue_age_cap 900 ep_queue_size 0 ep_storage_age 0 ep_storage_age_highwat 10 ep_storage_type featured ep_store_max_concurrency 10 ep_store_max_readers 9 ep_store_max_readwrite 1 ep_tap_bg_fetch_requeued 0 ep_tap_bg_fetched 963 ep_tap_keepalive 300 ep_tmp_oom_errors 0 ep_too_old 0 ep_too_young 0 ep_total_cache_size 1181400 ep_total_del_items 0 ep_total_enqueued 16035 ep_total_new_items 15741 ep_total_persisted 15741 ep_vbucket_del 0 ep_vbucket_del_fail 0 ep_version 1.6.5.3.1_13_g81a6cf1 ep_warmed_up 0 ep_warmup true ep_warmup_dups 0 ep_warmup_oom 0 ep_warmup_thread complete ep_warmup_time 15662 get_hits 0 get_misses 0 incr_hits 0 incr_misses 0 libevent 1.4.13-stable limit_maxbytes 67108864 mem_used 27035848 pid 16183 pointer_size 64 rejected_conns 0 rusage_system 1.364792 rusage_user 3.600452 tap_connect_received 120 tap_mutation_sent 1926 tap_opaque_sent 120 tap_vbucket_set_sent 240 threads 4 time 1305061523 total_connections 163 uptime 917 version 1.4.4_382_g9df3289 INFO REPORT <0.6107.0> 2011-05-10 14:05:28 =============================================================================== vbucketmigrator<0.6107.0>: Validate bucket states vbucketmigrator<0.6107.0>: 119 ok INFO REPORT <0.109.0> 2011-05-10 14:05:28 =============================================================================== ns_1@10.1.5.227:ns_orchestrator:164: Skipping janitor in state rebalancing: {rebalancing_state, <0.577.0>, {dict, 2, 16, 16, 8, 80, 48, {[], [], [], [], [], [], [], [], [], [], [], [], [], [], [], []}, {{[], [], [], [], [], [], [], [['ns_1@10.1.5.227'| 0.234375]], [], [['ns_1@10.1.5.229'| 0.234375]], [], [], [], [], [], []}}}} INFO REPORT <0.6111.0> 2011-05-10 14:05:29 =============================================================================== vbucketmigrator<0.6111.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.6111.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6111.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6111.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.6111.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6111.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6111.0>: Starting to move bucket 120 vbucketmigrator<0.6111.0>: Bucket 120 moved to the next server INFO REPORT <0.104.0> 2011-05-10 14:05:29 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.6111.0> 2011-05-10 14:05:34 =============================================================================== vbucketmigrator<0.6111.0>: Validate bucket states vbucketmigrator<0.6111.0>: 120 ok INFO REPORT <0.6126.0> 2011-05-10 14:05:35 =============================================================================== vbucketmigrator<0.6126.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.6126.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6126.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6126.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.6126.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6126.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6126.0>: Starting to move bucket 121 vbucketmigrator<0.6126.0>: Bucket 121 moved to the next server INFO REPORT <0.104.0> 2011-05-10 14:05:35 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.109.0> 2011-05-10 14:05:38 =============================================================================== ns_1@10.1.5.227:ns_orchestrator:164: Skipping janitor in state rebalancing: {rebalancing_state, <0.577.0>, {dict, 2, 16, 16, 8, 80, 48, {[], [], [], [], [], [], [], [], [], [], [], [], [], [], [], []}, {{[], [], [], [], [], [], [], [['ns_1@10.1.5.227'| 0.236328125]], [], [['ns_1@10.1.5.229'| 0.236328125]], [], [], [], [], [], []}}}} INFO REPORT <0.6126.0> 2011-05-10 14:05:40 =============================================================================== vbucketmigrator<0.6126.0>: Validate bucket states vbucketmigrator<0.6126.0>: 121 ok INFO REPORT <0.6131.0> 2011-05-10 14:05:41 =============================================================================== vbucketmigrator<0.6131.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.6131.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6131.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6131.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.6131.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6131.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6131.0>: Starting to move bucket 122 vbucketmigrator<0.6131.0>: Bucket 122 moved to the next server INFO REPORT <0.104.0> 2011-05-10 14:05:41 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.474.0> 2011-05-10 14:05:43 =============================================================================== ns_1@10.1.5.227:ns_doctor:82: Current node statuses: [{'ns_1@10.1.5.227', [{last_heard,{1305,61542,411202}}, {active_buckets,["default"]}, {memory, [{total,40051816}, {processes,19512672}, {processes_used,19500896}, {system,20539144}, {atom,819441}, {atom_used,806779}, {binary,104880}, {code,8552890}, {ets,1444464}]}, {cluster_compatibility_version,1}, {version, [{kernel,"2.13.4"}, {mnesia,"4.4.12"}, {stdlib,"1.16.4"}, {os_mon,"2.2.4"}, {inets,"5.2"}, {ns_server,"1.6.5.4r"}, {menelaus,"1.6.5.4r"}, {sasl,"2.1.8"}]}, {system_arch,"x86_64-unknown-linux-gnu"}, {wall_clock,934}, {memory_data,{2058731520,676438016,{<0.12.0>,6657168}}}, {disk_data, [{"/",11204616,49},{"/boot",101086,14},{"/dev/shm",1005240,0}]}, {meminfo, <<"MemTotal: 2010480 kB\nMemFree: 1348904 kB\nBuffers: 64532 kB\nCached: 342920 kB\nSwapCached: 20084 kB\nActive: 349636 kB\nInactive: 199044 kB\nHighTotal: 0 kB\nHighFree: 0 kB\nLowTotal: 2010480 kB\nLowFree: 1348904 kB\nSwapTotal: 4030456 kB\nSwapFree: 3952288 kB\nDirty: 228 kB\nWriteback: 0 kB\nAnonPages: 139024 kB\nMapped: 15736 kB\nSlab: 86624 kB\nPageTables: 6860 kB\nNFS_Unstable: 0 kB\nBounce: 0 kB\nCommitLimit: 5035696 kB\nCommitted_AS: 553456 kB\nVmallocTotal: 34359738367 kB\nVmallocUsed: 265484 kB\nVmallocChunk: 34359471011 kB\nHugePages_Total: 0\nHugePages_Free: 0\nHugePages_Rsvd: 0\nHugepagesize: 2048 kB\n">>}, {replication,[{"default",1.0}]}, {system_memory_data, [{system_total_memory,2058731520}, {free_swap,4047142912}, {total_swap,4127186944}, {cached_memory,351150080}, {buffered_memory,66080768}, {free_memory,1381277696}, {total_memory,2058731520}]}, {statistics, [{wall_clock,{930579,1}}, {context_switches,{336282,0}}, {garbage_collection,{66711,312995738,0}}, {io,{{input,86299298},{output,15130264}}}, {reductions,{94188512,806984}}, {run_queue,0}, {runtime,{13220,130}}]}]}, {'ns_1@10.1.5.229', [{last_heard,{1305,61542,242672}}, {active_buckets,["default"]}, {memory, [{total,32764304}, {processes,13298024}, {processes_used,13274504}, {system,19466280}, {atom,778121}, {atom_used,752794}, {binary,238808}, {code,7791241}, {ets,1372592}]}, {cluster_compatibility_version,1}, {version, [{kernel,"2.13.4"}, {mnesia,"4.4.12"}, {stdlib,"1.16.4"}, {os_mon,"2.2.4"}, {ns_server,"1.6.5.4r"}, {menelaus,"1.6.5.4r"}, {sasl,"2.1.8"}]}, {system_arch,"x86_64-unknown-linux-gnu"}, {wall_clock,897}, {memory_data,{2058731520,716255232,{<9352.12.0>,4114680}}}, {disk_data, [{"/",11204616,44},{"/boot",101086,14},{"/dev/shm",1005240,0}]}, {meminfo, <<"MemTotal: 2010480 kB\nMemFree: 1309772 kB\nBuffers: 83776 kB\nCached: 341252 kB\nSwapCached: 0 kB\nActive: 366916 kB\nInactive: 216972 kB\nHighTotal: 0 kB\nHighFree: 0 kB\nLowTotal: 2010480 kB\nLowFree: 1309772 kB\nSwapTotal: 4030456 kB\nSwapFree: 4030388 kB\nDirty: 108 kB\nWriteback: 0 kB\nAnonPages: 158860 kB\nMapped: 28104 kB\nSlab: 91908 kB\nPageTables: 5872 kB\nNFS_Unstable: 0 kB\nBounce: 0 kB\nCommitLimit: 5035696 kB\nCommitted_AS: 484736 kB\nVmallocTotal: 34359738367 kB\nVmallocUsed: 265484 kB\nVmallocChunk: 34359471011 kB\nHugePages_Total: 0\nHugePages_Free: 0\nHugePages_Rsvd: 0\nHugepagesize: 2048 kB\n">>}, {replication,[{"default",1.0}]}, {system_memory_data, [{system_total_memory,2058731520}, {free_swap,4127117312}, {total_swap,4127186944}, {cached_memory,349442048}, {buffered_memory,85786624}, {free_memory,1341206528}, {total_memory,2058731520}]}, {statistics, [{wall_clock,{889807,1}}, {context_switches,{80130,0}}, {garbage_collection,{17384,66309436,0}}, {io,{{input,9227290},{output,5392139}}}, {reductions,{22513809,240587}}, {run_queue,0}, {runtime,{3620,40}}]}]}] INFO REPORT <0.6131.0> 2011-05-10 14:05:46 =============================================================================== vbucketmigrator<0.6131.0>: Validate bucket states vbucketmigrator<0.6131.0>: 122 ok INFO REPORT <0.6148.0> 2011-05-10 14:05:47 =============================================================================== vbucketmigrator<0.6148.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.6148.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6148.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6148.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.6148.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6148.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6148.0>: Starting to move bucket 123 vbucketmigrator<0.6148.0>: Bucket 123 moved to the next server INFO REPORT <0.104.0> 2011-05-10 14:05:47 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.109.0> 2011-05-10 14:05:48 =============================================================================== ns_1@10.1.5.227:ns_orchestrator:164: Skipping janitor in state rebalancing: {rebalancing_state, <0.577.0>, {dict, 2, 16, 16, 8, 80, 48, {[], [], [], [], [], [], [], [], [], [], [], [], [], [], [], []}, {{[], [], [], [], [], [], [], [['ns_1@10.1.5.227'| 0.240234375]], [], [['ns_1@10.1.5.229'| 0.240234375]], [], [], [], [], [], []}}}} INFO REPORT <0.6148.0> 2011-05-10 14:05:52 =============================================================================== vbucketmigrator<0.6148.0>: Validate bucket states vbucketmigrator<0.6148.0>: 123 ok INFO REPORT <0.6161.0> 2011-05-10 14:05:53 =============================================================================== vbucketmigrator<0.6161.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.6161.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6161.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6161.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.6161.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6161.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6161.0>: Starting to move bucket 124 vbucketmigrator<0.6161.0>: Bucket 124 moved to the next server INFO REPORT <0.104.0> 2011-05-10 14:05:53 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.85.0> 2011-05-10 14:05:53 =============================================================================== Pulling config from: 'ns_1@10.1.5.229' INFO REPORT <0.6161.0> 2011-05-10 14:05:58 =============================================================================== vbucketmigrator<0.6161.0>: Validate bucket states vbucketmigrator<0.6161.0>: 124 ok INFO REPORT <0.109.0> 2011-05-10 14:05:58 =============================================================================== ns_1@10.1.5.227:ns_orchestrator:164: Skipping janitor in state rebalancing: {rebalancing_state, <0.577.0>, {dict, 2, 16, 16, 8, 80, 48, {[], [], [], [], [], [], [], [], [], [], [], [], [], [], [], []}, {{[], [], [], [], [], [], [], [['ns_1@10.1.5.227'| 0.244140625]], [], [['ns_1@10.1.5.229'| 0.244140625]], [], [], [], [], [], []}}}} INFO REPORT <0.6165.0> 2011-05-10 14:05:59 =============================================================================== vbucketmigrator<0.6165.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.6165.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6165.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6165.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.6165.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6165.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6165.0>: Starting to move bucket 125 vbucketmigrator<0.6165.0>: Bucket 125 moved to the next server INFO REPORT <0.104.0> 2011-05-10 14:05:59 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.6165.0> 2011-05-10 14:06:04 =============================================================================== vbucketmigrator<0.6165.0>: Validate bucket states vbucketmigrator<0.6165.0>: 125 ok INFO REPORT <0.6172.0> 2011-05-10 14:06:05 =============================================================================== vbucketmigrator<0.6172.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.6172.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6172.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6172.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.6172.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6172.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6172.0>: Starting to move bucket 126 vbucketmigrator<0.6172.0>: Bucket 126 moved to the next server INFO REPORT <0.104.0> 2011-05-10 14:06:05 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.109.0> 2011-05-10 14:06:08 =============================================================================== ns_1@10.1.5.227:ns_orchestrator:164: Skipping janitor in state rebalancing: {rebalancing_state, <0.577.0>, {dict, 2, 16, 16, 8, 80, 48, {[], [], [], [], [], [], [], [], [], [], [], [], [], [], [], []}, {{[], [], [], [], [], [], [], [['ns_1@10.1.5.227'| 0.24609375]], [], [['ns_1@10.1.5.229'| 0.24609375]], [], [], [], [], [], []}}}} INFO REPORT <0.6172.0> 2011-05-10 14:06:10 =============================================================================== vbucketmigrator<0.6172.0>: Validate bucket states vbucketmigrator<0.6172.0>: 126 ok INFO REPORT <0.6182.0> 2011-05-10 14:06:11 =============================================================================== vbucketmigrator<0.6182.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.6182.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6182.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6182.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.6182.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6182.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6182.0>: Starting to move bucket 127 vbucketmigrator<0.6182.0>: Bucket 127 moved to the next server INFO REPORT <0.104.0> 2011-05-10 14:06:11 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.6182.0> 2011-05-10 14:06:16 =============================================================================== vbucketmigrator<0.6182.0>: Validate bucket states vbucketmigrator<0.6182.0>: 127 ok INFO REPORT <0.6190.0> 2011-05-10 14:06:17 =============================================================================== vbucketmigrator<0.6190.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.6190.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6190.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6190.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.6190.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6190.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6190.0>: Starting to move bucket 128 vbucketmigrator<0.6190.0>: Bucket 128 moved to the next server INFO REPORT <0.104.0> 2011-05-10 14:06:17 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.109.0> 2011-05-10 14:06:18 =============================================================================== ns_1@10.1.5.227:ns_orchestrator:164: Skipping janitor in state rebalancing: {rebalancing_state, <0.577.0>, {dict, 2, 16, 16, 8, 80, 48, {[], [], [], [], [], [], [], [], [], [], [], [], [], [], [], []}, {{[], [], [], [], [], [], [], [['ns_1@10.1.5.227'| 0.25]], [], [['ns_1@10.1.5.229'| 0.25]], [], [], [], [], [], []}}}} INFO REPORT <0.6190.0> 2011-05-10 14:06:22 =============================================================================== vbucketmigrator<0.6190.0>: Validate bucket states vbucketmigrator<0.6190.0>: 128 ok INFO REPORT <0.6195.0> 2011-05-10 14:06:23 =============================================================================== vbucketmigrator<0.6195.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.6195.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6195.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6195.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.6195.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6195.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6195.0>: Starting to move bucket 129 vbucketmigrator<0.6195.0>: Bucket 129 moved to the next server INFO REPORT <0.104.0> 2011-05-10 14:06:23 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.6195.0> 2011-05-10 14:06:28 =============================================================================== vbucketmigrator<0.6195.0>: Validate bucket states vbucketmigrator<0.6195.0>: 129 ok INFO REPORT <0.109.0> 2011-05-10 14:06:28 =============================================================================== ns_1@10.1.5.227:ns_orchestrator:164: Skipping janitor in state rebalancing: {rebalancing_state, <0.577.0>, {dict, 2, 16, 16, 8, 80, 48, {[], [], [], [], [], [], [], [], [], [], [], [], [], [], [], []}, {{[], [], [], [], [], [], [], [['ns_1@10.1.5.227'| 0.25390625]], [], [['ns_1@10.1.5.229'| 0.25390625]], [], [], [], [], [], []}}}} INFO REPORT <0.6199.0> 2011-05-10 14:06:29 =============================================================================== vbucketmigrator<0.6199.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.6199.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6199.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6199.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.6199.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6199.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6199.0>: Starting to move bucket 130 vbucketmigrator<0.6199.0>: Bucket 130 moved to the next server INFO REPORT <0.104.0> 2011-05-10 14:06:29 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.6199.0> 2011-05-10 14:06:34 =============================================================================== vbucketmigrator<0.6199.0>: Validate bucket states vbucketmigrator<0.6199.0>: 130 ok INFO REPORT <0.6215.0> 2011-05-10 14:06:35 =============================================================================== vbucketmigrator<0.6215.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.6215.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6215.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6215.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.6215.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6215.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6215.0>: Starting to move bucket 131 vbucketmigrator<0.6215.0>: Bucket 131 moved to the next server INFO REPORT <0.104.0> 2011-05-10 14:06:35 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.85.0> 2011-05-10 14:06:37 =============================================================================== Pulling config from: 'ns_1@10.1.5.229' INFO REPORT <0.109.0> 2011-05-10 14:06:38 =============================================================================== ns_1@10.1.5.227:ns_orchestrator:164: Skipping janitor in state rebalancing: {rebalancing_state, <0.577.0>, {dict, 2, 16, 16, 8, 80, 48, {[], [], [], [], [], [], [], [], [], [], [], [], [], [], [], []}, {{[], [], [], [], [], [], [], [['ns_1@10.1.5.227'| 0.255859375]], [], [['ns_1@10.1.5.229'| 0.255859375]], [], [], [], [], [], []}}}} INFO REPORT <0.6215.0> 2011-05-10 14:06:40 =============================================================================== vbucketmigrator<0.6215.0>: Validate bucket states vbucketmigrator<0.6215.0>: 131 ok INFO REPORT <0.6220.0> 2011-05-10 14:06:41 =============================================================================== vbucketmigrator<0.6220.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.6220.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6220.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6220.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.6220.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6220.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6220.0>: Starting to move bucket 132 vbucketmigrator<0.6220.0>: Bucket 132 moved to the next server INFO REPORT <0.104.0> 2011-05-10 14:06:41 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.474.0> 2011-05-10 14:06:43 =============================================================================== ns_1@10.1.5.227:ns_doctor:82: Current node statuses: [{'ns_1@10.1.5.227', [{last_heard,{1305,61602,411223}}, {active_buckets,["default"]}, {memory, [{total,40060752}, {processes,19468304}, {processes_used,19456528}, {system,20592448}, {atom,819441}, {atom_used,806779}, {binary,90528}, {code,8552890}, {ets,1510384}]}, {cluster_compatibility_version,1}, {version, [{kernel,"2.13.4"}, {mnesia,"4.4.12"}, {stdlib,"1.16.4"}, {os_mon,"2.2.4"}, {inets,"5.2"}, {ns_server,"1.6.5.4r"}, {menelaus,"1.6.5.4r"}, {sasl,"2.1.8"}]}, {system_arch,"x86_64-unknown-linux-gnu"}, {wall_clock,994}, {memory_data,{2058731520,676311040,{<0.12.0>,6657168}}}, {disk_data, [{"/",11204616,49},{"/boot",101086,14},{"/dev/shm",1005240,0}]}, {meminfo, <<"MemTotal: 2010480 kB\nMemFree: 1348780 kB\nBuffers: 64876 kB\nCached: 341964 kB\nSwapCached: 20084 kB\nActive: 349760 kB\nInactive: 198628 kB\nHighTotal: 0 kB\nHighFree: 0 kB\nLowTotal: 2010480 kB\nLowFree: 1348780 kB\nSwapTotal: 4030456 kB\nSwapFree: 3952288 kB\nDirty: 120 kB\nWriteback: 0 kB\nAnonPages: 139320 kB\nMapped: 15736 kB\nSlab: 86656 kB\nPageTables: 6864 kB\nNFS_Unstable: 0 kB\nBounce: 0 kB\nCommitLimit: 5035696 kB\nCommitted_AS: 555500 kB\nVmallocTotal: 34359738367 kB\nVmallocUsed: 265484 kB\nVmallocChunk: 34359471011 kB\nHugePages_Total: 0\nHugePages_Free: 0\nHugePages_Rsvd: 0\nHugepagesize: 2048 kB\n">>}, {replication,[{"default",1.0}]}, {system_memory_data, [{system_total_memory,2058731520}, {free_swap,4047142912}, {total_swap,4127186944}, {cached_memory,350171136}, {buffered_memory,66433024}, {free_memory,1381150720}, {total_memory,2058731520}]}, {statistics, [{wall_clock,{990579,1}}, {context_switches,{346987,0}}, {garbage_collection,{69498,332860755,0}}, {io,{{input,87193580},{output,16045802}}}, {reductions,{98827983,797375}}, {run_queue,0}, {runtime,{14000,130}}]}]}, {'ns_1@10.1.5.229', [{last_heard,{1305,61602,242158}}, {active_buckets,["default"]}, {memory, [{total,32820600}, {processes,13294992}, {processes_used,13271472}, {system,19525608}, {atom,778121}, {atom_used,752794}, {binary,230928}, {code,7791241}, {ets,1438792}]}, {cluster_compatibility_version,1}, {version, [{kernel,"2.13.4"}, {mnesia,"4.4.12"}, {stdlib,"1.16.4"}, {os_mon,"2.2.4"}, {ns_server,"1.6.5.4r"}, {menelaus,"1.6.5.4r"}, {sasl,"2.1.8"}]}, {system_arch,"x86_64-unknown-linux-gnu"}, {wall_clock,957}, {memory_data,{2058731520,718794752,{<9352.12.0>,4114680}}}, {disk_data, [{"/",11204616,44},{"/boot",101086,14},{"/dev/shm",1005240,0}]}, {meminfo, <<"MemTotal: 2010480 kB\nMemFree: 1309152 kB\nBuffers: 84536 kB\nCached: 340304 kB\nSwapCached: 0 kB\nActive: 367388 kB\nInactive: 216800 kB\nHighTotal: 0 kB\nHighFree: 0 kB\nLowTotal: 2010480 kB\nLowFree: 1309152 kB\nSwapTotal: 4030456 kB\nSwapFree: 4030388 kB\nDirty: 228 kB\nWriteback: 0 kB\nAnonPages: 159384 kB\nMapped: 28104 kB\nSlab: 91908 kB\nPageTables: 5876 kB\nNFS_Unstable: 0 kB\nBounce: 0 kB\nCommitLimit: 5035696 kB\nCommitted_AS: 484668 kB\nVmallocTotal: 34359738367 kB\nVmallocUsed: 265484 kB\nVmallocChunk: 34359471011 kB\nHugePages_Total: 0\nHugePages_Free: 0\nHugePages_Rsvd: 0\nHugepagesize: 2048 kB\n">>}, {replication,[{"default",1.0}]}, {system_memory_data, [{system_total_memory,2058731520}, {free_swap,4127117312}, {total_swap,4127186944}, {cached_memory,348471296}, {buffered_memory,86564864}, {free_memory,1340571648}, {total_memory,2058731520}]}, {statistics, [{wall_clock,{949806,1}}, {context_switches,{85255,0}}, {garbage_collection,{18592,71000394,0}}, {io,{{input,10184194},{output,6117907}}}, {reductions,{23960826,231139}}, {run_queue,0}, {runtime,{3880,40}}]}]}] INFO REPORT <0.6220.0> 2011-05-10 14:06:46 =============================================================================== vbucketmigrator<0.6220.0>: Validate bucket states vbucketmigrator<0.6220.0>: 132 ok INFO REPORT <0.6224.0> 2011-05-10 14:06:47 =============================================================================== vbucketmigrator<0.6224.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.6224.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6224.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6224.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.6224.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6224.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6224.0>: Starting to move bucket 133 vbucketmigrator<0.6224.0>: Bucket 133 moved to the next server INFO REPORT <0.104.0> 2011-05-10 14:06:47 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.109.0> 2011-05-10 14:06:48 =============================================================================== ns_1@10.1.5.227:ns_orchestrator:164: Skipping janitor in state rebalancing: {rebalancing_state, <0.577.0>, {dict, 2, 16, 16, 8, 80, 48, {[], [], [], [], [], [], [], [], [], [], [], [], [], [], [], []}, {{[], [], [], [], [], [], [], [['ns_1@10.1.5.227'| 0.259765625]], [], [['ns_1@10.1.5.229'| 0.259765625]], [], [], [], [], [], []}}}} INFO REPORT <0.6224.0> 2011-05-10 14:06:52 =============================================================================== vbucketmigrator<0.6224.0>: Validate bucket states vbucketmigrator<0.6224.0>: 133 ok INFO REPORT <0.6237.0> 2011-05-10 14:06:53 =============================================================================== vbucketmigrator<0.6237.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.6237.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6237.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6237.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.6237.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6237.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6237.0>: Starting to move bucket 134 vbucketmigrator<0.6237.0>: Bucket 134 moved to the next server INFO REPORT <0.104.0> 2011-05-10 14:06:53 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.6237.0> 2011-05-10 14:06:58 =============================================================================== vbucketmigrator<0.6237.0>: Validate bucket states vbucketmigrator<0.6237.0>: 134 ok INFO REPORT <0.109.0> 2011-05-10 14:06:58 =============================================================================== ns_1@10.1.5.227:ns_orchestrator:164: Skipping janitor in state rebalancing: {rebalancing_state, <0.577.0>, {dict, 2, 16, 16, 8, 80, 48, {[], [], [], [], [], [], [], [], [], [], [], [], [], [], [], []}, {{[], [], [], [], [], [], [], [['ns_1@10.1.5.227'| 0.263671875]], [], [['ns_1@10.1.5.229'| 0.263671875]], [], [], [], [], [], []}}}} INFO REPORT <0.6241.0> 2011-05-10 14:06:59 =============================================================================== vbucketmigrator<0.6241.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.6241.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6241.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6241.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.6241.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6241.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6241.0>: Starting to move bucket 135 vbucketmigrator<0.6241.0>: Bucket 135 moved to the next server INFO REPORT <0.104.0> 2011-05-10 14:06:59 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.469.0> 2011-05-10 14:07:04 =============================================================================== ns_1@10.1.5.227:stats_collector:71: Stats for bucket "default": auth_cmds 168 auth_errors 0 bucket_conns 34 bytes_read 1627797 bytes_written 8965028 cas_badval 0 cas_hits 0 cas_misses 0 cmd_flush 1 cmd_get 0 cmd_set 16004 conn_yields 0 connection_structures 44 curr_connections 44 curr_items 6960 curr_items_tot 8040 daemon_connections 10 decr_hits 0 decr_misses 0 delete_hits 0 delete_misses 0 ep_bg_fetched 0 ep_commit_num 53 ep_commit_time 0 ep_commit_time_total 4 ep_data_age 13 ep_data_age_highwat 13 ep_db_cleaner_status complete ep_db_strategy multiMTVBDB ep_dbinit 1 ep_dbname /var/opt/membase/1.6.5.4r/data/ns_1/default ep_dbshards 4 ep_expired 0 ep_flush_duration 1 ep_flush_duration_highwat 8 ep_flush_duration_total 12 ep_flush_preempts 0 ep_flusher_state running ep_flusher_todo 0 ep_io_num_read 1080 ep_io_num_write 15741 ep_io_read_bytes 61249 ep_io_write_bytes 1011432 ep_item_begin_failed 0 ep_item_commit_failed 0 ep_item_flush_expired 0 ep_item_flush_failed 0 ep_kv_size 1181400 ep_max_data_size 1096810496 ep_max_txn_size 1000 ep_mem_high_wat 822607872 ep_mem_low_wat 658086297 ep_min_data_age 0 ep_num_active_non_resident 0 ep_num_eject_failures 0 ep_num_eject_replicas 0 ep_num_expiry_pager_runs 0 ep_num_non_resident 0 ep_num_not_my_vbuckets 0 ep_num_pager_runs 0 ep_num_value_ejects 0 ep_oom_errors 0 ep_overhead 25854448 ep_pending_ops 0 ep_pending_ops_max 0 ep_pending_ops_max_duration 0 ep_pending_ops_total 0 ep_queue_age_cap 900 ep_queue_size 0 ep_storage_age 0 ep_storage_age_highwat 10 ep_storage_type featured ep_store_max_concurrency 10 ep_store_max_readers 9 ep_store_max_readwrite 1 ep_tap_bg_fetch_requeued 0 ep_tap_bg_fetched 1080 ep_tap_keepalive 300 ep_tmp_oom_errors 0 ep_too_old 0 ep_too_young 0 ep_total_cache_size 1181400 ep_total_del_items 0 ep_total_enqueued 16035 ep_total_new_items 15741 ep_total_persisted 15741 ep_vbucket_del 0 ep_vbucket_del_fail 0 ep_version 1.6.5.3.1_13_g81a6cf1 ep_warmed_up 0 ep_warmup true ep_warmup_dups 0 ep_warmup_oom 0 ep_warmup_thread complete ep_warmup_time 15662 get_hits 0 get_misses 0 incr_hits 0 incr_misses 0 libevent 1.4.13-stable limit_maxbytes 67108864 mem_used 27035848 pid 16183 pointer_size 64 rejected_conns 0 rusage_system 1.396787 rusage_user 3.777425 tap_connect_received 136 tap_mutation_sent 2160 tap_opaque_sent 136 tap_vbucket_set_sent 272 threads 4 time 1305061623 total_connections 179 uptime 1017 version 1.4.4_382_g9df3289 INFO REPORT <0.6241.0> 2011-05-10 14:07:04 =============================================================================== vbucketmigrator<0.6241.0>: Validate bucket states vbucketmigrator<0.6241.0>: 135 ok INFO REPORT <0.6248.0> 2011-05-10 14:07:05 =============================================================================== vbucketmigrator<0.6248.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.6248.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6248.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6248.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.6248.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6248.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6248.0>: Starting to move bucket 136 vbucketmigrator<0.6248.0>: Bucket 136 moved to the next server INFO REPORT <0.104.0> 2011-05-10 14:07:05 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.109.0> 2011-05-10 14:07:08 =============================================================================== ns_1@10.1.5.227:ns_orchestrator:164: Skipping janitor in state rebalancing: {rebalancing_state, <0.577.0>, {dict, 2, 16, 16, 8, 80, 48, {[], [], [], [], [], [], [], [], [], [], [], [], [], [], [], []}, {{[], [], [], [], [], [], [], [['ns_1@10.1.5.227'| 0.265625]], [], [['ns_1@10.1.5.229'| 0.265625]], [], [], [], [], [], []}}}} INFO REPORT <0.6248.0> 2011-05-10 14:07:10 =============================================================================== vbucketmigrator<0.6248.0>: Validate bucket states vbucketmigrator<0.6248.0>: 136 ok INFO REPORT <0.6258.0> 2011-05-10 14:07:11 =============================================================================== vbucketmigrator<0.6258.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.6258.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6258.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6258.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.6258.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6258.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6258.0>: Starting to move bucket 137 vbucketmigrator<0.6258.0>: Bucket 137 moved to the next server INFO REPORT <0.104.0> 2011-05-10 14:07:11 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.6258.0> 2011-05-10 14:07:16 =============================================================================== vbucketmigrator<0.6258.0>: Validate bucket states vbucketmigrator<0.6258.0>: 137 ok INFO REPORT <0.6266.0> 2011-05-10 14:07:17 =============================================================================== vbucketmigrator<0.6266.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.6266.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6266.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6266.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.6266.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6266.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6266.0>: Starting to move bucket 138 vbucketmigrator<0.6266.0>: Bucket 138 moved to the next server INFO REPORT <0.104.0> 2011-05-10 14:07:17 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.109.0> 2011-05-10 14:07:18 =============================================================================== ns_1@10.1.5.227:ns_orchestrator:164: Skipping janitor in state rebalancing: {rebalancing_state, <0.577.0>, {dict, 2, 16, 16, 8, 80, 48, {[], [], [], [], [], [], [], [], [], [], [], [], [], [], [], []}, {{[], [], [], [], [], [], [], [['ns_1@10.1.5.227'| 0.26953125]], [], [['ns_1@10.1.5.229'| 0.26953125]], [], [], [], [], [], []}}}} INFO REPORT <0.6266.0> 2011-05-10 14:07:22 =============================================================================== vbucketmigrator<0.6266.0>: Validate bucket states vbucketmigrator<0.6266.0>: 138 ok INFO REPORT <0.6271.0> 2011-05-10 14:07:23 =============================================================================== vbucketmigrator<0.6271.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.6271.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6271.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6271.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.6271.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6271.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6271.0>: Starting to move bucket 139 vbucketmigrator<0.6271.0>: Bucket 139 moved to the next server INFO REPORT <0.104.0> 2011-05-10 14:07:23 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.85.0> 2011-05-10 14:07:26 =============================================================================== Pulling config from: 'ns_1@10.1.5.229' INFO REPORT <0.6271.0> 2011-05-10 14:07:28 =============================================================================== vbucketmigrator<0.6271.0>: Validate bucket states vbucketmigrator<0.6271.0>: 139 ok INFO REPORT <0.109.0> 2011-05-10 14:07:28 =============================================================================== ns_1@10.1.5.227:ns_orchestrator:164: Skipping janitor in state rebalancing: {rebalancing_state, <0.577.0>, {dict, 2, 16, 16, 8, 80, 48, {[], [], [], [], [], [], [], [], [], [], [], [], [], [], [], []}, {{[], [], [], [], [], [], [], [['ns_1@10.1.5.227'| 0.2734375]], [], [['ns_1@10.1.5.229'| 0.2734375]], [], [], [], [], [], []}}}} INFO REPORT <0.6275.0> 2011-05-10 14:07:29 =============================================================================== vbucketmigrator<0.6275.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.6275.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6275.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6275.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.6275.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6275.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6275.0>: Starting to move bucket 140 vbucketmigrator<0.6275.0>: Bucket 140 moved to the next server INFO REPORT <0.104.0> 2011-05-10 14:07:29 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.6275.0> 2011-05-10 14:07:34 =============================================================================== vbucketmigrator<0.6275.0>: Validate bucket states vbucketmigrator<0.6275.0>: 140 ok INFO REPORT <0.6290.0> 2011-05-10 14:07:35 =============================================================================== vbucketmigrator<0.6290.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.6290.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6290.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6290.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.6290.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6290.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6290.0>: Starting to move bucket 141 vbucketmigrator<0.6290.0>: Bucket 141 moved to the next server INFO REPORT <0.104.0> 2011-05-10 14:07:35 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.109.0> 2011-05-10 14:07:38 =============================================================================== ns_1@10.1.5.227:ns_orchestrator:164: Skipping janitor in state rebalancing: {rebalancing_state, <0.577.0>, {dict, 2, 16, 16, 8, 80, 48, {[], [], [], [], [], [], [], [], [], [], [], [], [], [], [], []}, {{[], [], [], [], [], [], [], [['ns_1@10.1.5.227'| 0.275390625]], [], [['ns_1@10.1.5.229'| 0.275390625]], [], [], [], [], [], []}}}} INFO REPORT <0.6290.0> 2011-05-10 14:07:40 =============================================================================== vbucketmigrator<0.6290.0>: Validate bucket states vbucketmigrator<0.6290.0>: 141 ok INFO REPORT <0.6295.0> 2011-05-10 14:07:41 =============================================================================== vbucketmigrator<0.6295.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.6295.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6295.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6295.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.6295.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6295.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6295.0>: Starting to move bucket 142 vbucketmigrator<0.6295.0>: Bucket 142 moved to the next server INFO REPORT <0.104.0> 2011-05-10 14:07:41 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.474.0> 2011-05-10 14:07:43 =============================================================================== ns_1@10.1.5.227:ns_doctor:82: Current node statuses: [{'ns_1@10.1.5.227', [{last_heard,{1305,61662,411213}}, {active_buckets,["default"]}, {memory, [{total,40010120}, {processes,19347280}, {processes_used,19334640}, {system,20662840}, {atom,819441}, {atom_used,806779}, {binary,100480}, {code,8552890}, {ets,1570488}]}, {cluster_compatibility_version,1}, {version, [{kernel,"2.13.4"}, {mnesia,"4.4.12"}, {stdlib,"1.16.4"}, {os_mon,"2.2.4"}, {inets,"5.2"}, {ns_server,"1.6.5.4r"}, {menelaus,"1.6.5.4r"}, {sasl,"2.1.8"}]}, {system_arch,"x86_64-unknown-linux-gnu"}, {wall_clock,1054}, {memory_data,{2058731520,677072896,{<0.12.0>,6657168}}}, {disk_data, [{"/",11204616,49},{"/boot",101086,14},{"/dev/shm",1005240,0}]}, {meminfo, <<"MemTotal: 2010480 kB\nMemFree: 1348780 kB\nBuffers: 65176 kB\nCached: 342224 kB\nSwapCached: 20084 kB\nActive: 349696 kB\nInactive: 199084 kB\nHighTotal: 0 kB\nHighFree: 0 kB\nLowTotal: 2010480 kB\nLowFree: 1348780 kB\nSwapTotal: 4030456 kB\nSwapFree: 3952288 kB\nDirty: 232 kB\nWriteback: 0 kB\nAnonPages: 139156 kB\nMapped: 15736 kB\nSlab: 86620 kB\nPageTables: 6860 kB\nNFS_Unstable: 0 kB\nBounce: 0 kB\nCommitLimit: 5035696 kB\nCommitted_AS: 555500 kB\nVmallocTotal: 34359738367 kB\nVmallocUsed: 265484 kB\nVmallocChunk: 34359471011 kB\nHugePages_Total: 0\nHugePages_Free: 0\nHugePages_Rsvd: 0\nHugepagesize: 2048 kB\n">>}, {replication,[{"default",1.0}]}, {system_memory_data, [{system_total_memory,2058731520}, {free_swap,4047142912}, {total_swap,4127186944}, {cached_memory,350437376}, {buffered_memory,66740224}, {free_memory,1381150720}, {total_memory,2058731520}]}, {statistics, [{wall_clock,{1050579,1}}, {context_switches,{357339,0}}, {garbage_collection,{72168,352432308,0}}, {io,{{input,87322886},{output,16470766}}}, {reductions,{103403369,802177}}, {run_queue,0}, {runtime,{14700,120}}]}]}, {'ns_1@10.1.5.229', [{last_heard,{1305,61662,242146}}, {active_buckets,["default"]}, {memory, [{total,32855704}, {processes,13282208}, {processes_used,13258688}, {system,19573496}, {atom,778121}, {atom_used,752794}, {binary,218656}, {code,7791241}, {ets,1498592}]}, {cluster_compatibility_version,1}, {version, [{kernel,"2.13.4"}, {mnesia,"4.4.12"}, {stdlib,"1.16.4"}, {os_mon,"2.2.4"}, {ns_server,"1.6.5.4r"}, {menelaus,"1.6.5.4r"}, {sasl,"2.1.8"}]}, {system_arch,"x86_64-unknown-linux-gnu"}, {wall_clock,1017}, {memory_data,{2058731520,718032896,{<9352.12.0>,4114680}}}, {disk_data, [{"/",11204616,44},{"/boot",101086,14},{"/dev/shm",1005240,0}]}, {meminfo, <<"MemTotal: 2010480 kB\nMemFree: 1307788 kB\nBuffers: 85244 kB\nCached: 340564 kB\nSwapCached: 0 kB\nActive: 367984 kB\nInactive: 217548 kB\nHighTotal: 0 kB\nHighFree: 0 kB\nLowTotal: 2010480 kB\nLowFree: 1307788 kB\nSwapTotal: 4030456 kB\nSwapFree: 4030388 kB\nDirty: 100 kB\nWriteback: 0 kB\nAnonPages: 159732 kB\nMapped: 28104 kB\nSlab: 91908 kB\nPageTables: 5868 kB\nNFS_Unstable: 0 kB\nBounce: 0 kB\nCommitLimit: 5035696 kB\nCommitted_AS: 485760 kB\nVmallocTotal: 34359738367 kB\nVmallocUsed: 265484 kB\nVmallocChunk: 34359471011 kB\nHugePages_Total: 0\nHugePages_Free: 0\nHugePages_Rsvd: 0\nHugepagesize: 2048 kB\n">>}, {replication,[{"default",1.0}]}, {system_memory_data, [{system_total_memory,2058731520}, {free_swap,4127117312}, {total_swap,4127186944}, {cached_memory,348737536}, {buffered_memory,87289856}, {free_memory,1339174912}, {total_memory,2058731520}]}, {statistics, [{wall_clock,{1009806,1}}, {context_switches,{90053,0}}, {garbage_collection,{19645,75492904,0}}, {io,{{input,10421138},{output,6380503}}}, {reductions,{25327974,232444}}, {run_queue,0}, {runtime,{4160,50}}]}]}] INFO REPORT <0.6295.0> 2011-05-10 14:07:46 =============================================================================== vbucketmigrator<0.6295.0>: Validate bucket states vbucketmigrator<0.6295.0>: 142 ok INFO REPORT <0.6299.0> 2011-05-10 14:07:47 =============================================================================== vbucketmigrator<0.6299.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.6299.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6299.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6299.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.6299.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6299.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6299.0>: Starting to move bucket 143 vbucketmigrator<0.6299.0>: Bucket 143 moved to the next server INFO REPORT <0.104.0> 2011-05-10 14:07:47 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.109.0> 2011-05-10 14:07:48 =============================================================================== ns_1@10.1.5.227:ns_orchestrator:164: Skipping janitor in state rebalancing: {rebalancing_state, <0.577.0>, {dict, 2, 16, 16, 8, 80, 48, {[], [], [], [], [], [], [], [], [], [], [], [], [], [], [], []}, {{[], [], [], [], [], [], [], [['ns_1@10.1.5.227'| 0.279296875]], [], [['ns_1@10.1.5.229'| 0.279296875]], [], [], [], [], [], []}}}} INFO REPORT <0.6299.0> 2011-05-10 14:07:52 =============================================================================== vbucketmigrator<0.6299.0>: Validate bucket states vbucketmigrator<0.6299.0>: 143 ok INFO REPORT <0.85.0> 2011-05-10 14:07:53 =============================================================================== Pulling config from: 'ns_1@10.1.5.229' INFO REPORT <0.6312.0> 2011-05-10 14:07:53 =============================================================================== vbucketmigrator<0.6312.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.6312.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6312.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6312.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.6312.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6312.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6312.0>: Starting to move bucket 144 vbucketmigrator<0.6312.0>: Bucket 144 moved to the next server INFO REPORT <0.104.0> 2011-05-10 14:07:53 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.6312.0> 2011-05-10 14:07:58 =============================================================================== vbucketmigrator<0.6312.0>: Validate bucket states vbucketmigrator<0.6312.0>: 144 ok INFO REPORT <0.109.0> 2011-05-10 14:07:58 =============================================================================== ns_1@10.1.5.227:ns_orchestrator:164: Skipping janitor in state rebalancing: {rebalancing_state, <0.577.0>, {dict, 2, 16, 16, 8, 80, 48, {[], [], [], [], [], [], [], [], [], [], [], [], [], [], [], []}, {{[], [], [], [], [], [], [], [['ns_1@10.1.5.227'| 0.283203125]], [], [['ns_1@10.1.5.229'| 0.283203125]], [], [], [], [], [], []}}}} INFO REPORT <0.6317.0> 2011-05-10 14:07:59 =============================================================================== vbucketmigrator<0.6317.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.6317.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6317.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6317.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.6317.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6317.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6317.0>: Starting to move bucket 145 vbucketmigrator<0.6317.0>: Bucket 145 moved to the next server INFO REPORT <0.104.0> 2011-05-10 14:07:59 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.6317.0> 2011-05-10 14:08:04 =============================================================================== vbucketmigrator<0.6317.0>: Validate bucket states vbucketmigrator<0.6317.0>: 145 ok INFO REPORT <0.6323.0> 2011-05-10 14:08:05 =============================================================================== vbucketmigrator<0.6323.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.6323.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6323.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6323.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.6323.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6323.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6323.0>: Starting to move bucket 146 vbucketmigrator<0.6323.0>: Bucket 146 moved to the next server INFO REPORT <0.104.0> 2011-05-10 14:08:05 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.109.0> 2011-05-10 14:08:08 =============================================================================== ns_1@10.1.5.227:ns_orchestrator:164: Skipping janitor in state rebalancing: {rebalancing_state, <0.577.0>, {dict, 2, 16, 16, 8, 80, 48, {[], [], [], [], [], [], [], [], [], [], [], [], [], [], [], []}, {{[], [], [], [], [], [], [], [['ns_1@10.1.5.227'| 0.28515625]], [], [['ns_1@10.1.5.229'| 0.28515625]], [], [], [], [], [], []}}}} INFO REPORT <0.6323.0> 2011-05-10 14:08:10 =============================================================================== vbucketmigrator<0.6323.0>: Validate bucket states vbucketmigrator<0.6323.0>: 146 ok INFO REPORT <0.6333.0> 2011-05-10 14:08:11 =============================================================================== vbucketmigrator<0.6333.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.6333.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6333.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6333.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.6333.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6333.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6333.0>: Starting to move bucket 147 vbucketmigrator<0.6333.0>: Bucket 147 moved to the next server INFO REPORT <0.104.0> 2011-05-10 14:08:11 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.6333.0> 2011-05-10 14:08:16 =============================================================================== vbucketmigrator<0.6333.0>: Validate bucket states vbucketmigrator<0.6333.0>: 147 ok INFO REPORT <0.6341.0> 2011-05-10 14:08:17 =============================================================================== vbucketmigrator<0.6341.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.6341.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6341.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6341.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.6341.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6341.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6341.0>: Starting to move bucket 148 vbucketmigrator<0.6341.0>: Bucket 148 moved to the next server INFO REPORT <0.104.0> 2011-05-10 14:08:17 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.109.0> 2011-05-10 14:08:18 =============================================================================== ns_1@10.1.5.227:ns_orchestrator:164: Skipping janitor in state rebalancing: {rebalancing_state, <0.577.0>, {dict, 2, 16, 16, 8, 80, 48, {[], [], [], [], [], [], [], [], [], [], [], [], [], [], [], []}, {{[], [], [], [], [], [], [], [['ns_1@10.1.5.227'| 0.2890625]], [], [['ns_1@10.1.5.229'| 0.2890625]], [], [], [], [], [], []}}}} INFO REPORT <0.6341.0> 2011-05-10 14:08:22 =============================================================================== vbucketmigrator<0.6341.0>: Validate bucket states vbucketmigrator<0.6341.0>: 148 ok INFO REPORT <0.6346.0> 2011-05-10 14:08:23 =============================================================================== vbucketmigrator<0.6346.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.6346.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6346.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6346.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.6346.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6346.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6346.0>: Starting to move bucket 149 vbucketmigrator<0.6346.0>: Bucket 149 moved to the next server INFO REPORT <0.104.0> 2011-05-10 14:08:23 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.6346.0> 2011-05-10 14:08:28 =============================================================================== vbucketmigrator<0.6346.0>: Validate bucket states vbucketmigrator<0.6346.0>: 149 ok INFO REPORT <0.109.0> 2011-05-10 14:08:28 =============================================================================== ns_1@10.1.5.227:ns_orchestrator:164: Skipping janitor in state rebalancing: {rebalancing_state, <0.577.0>, {dict, 2, 16, 16, 8, 80, 48, {[], [], [], [], [], [], [], [], [], [], [], [], [], [], [], []}, {{[], [], [], [], [], [], [], [['ns_1@10.1.5.227'| 0.29296875]], [], [['ns_1@10.1.5.229'| 0.29296875]], [], [], [], [], [], []}}}} INFO REPORT <0.6352.0> 2011-05-10 14:08:29 =============================================================================== vbucketmigrator<0.6352.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.6352.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6352.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6352.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.6352.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6352.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6352.0>: Starting to move bucket 150 vbucketmigrator<0.6352.0>: Bucket 150 moved to the next server INFO REPORT <0.104.0> 2011-05-10 14:08:29 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.6352.0> 2011-05-10 14:08:34 =============================================================================== vbucketmigrator<0.6352.0>: Validate bucket states vbucketmigrator<0.6352.0>: 150 ok INFO REPORT <0.85.0> 2011-05-10 14:08:35 =============================================================================== Pulling config from: 'ns_1@10.1.5.229' INFO REPORT <0.6366.0> 2011-05-10 14:08:35 =============================================================================== vbucketmigrator<0.6366.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.6366.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6366.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6366.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.6366.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6366.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6366.0>: Starting to move bucket 151 vbucketmigrator<0.6366.0>: Bucket 151 moved to the next server INFO REPORT <0.104.0> 2011-05-10 14:08:35 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.109.0> 2011-05-10 14:08:38 =============================================================================== ns_1@10.1.5.227:ns_orchestrator:164: Skipping janitor in state rebalancing: {rebalancing_state, <0.577.0>, {dict, 2, 16, 16, 8, 80, 48, {[], [], [], [], [], [], [], [], [], [], [], [], [], [], [], []}, {{[], [], [], [], [], [], [], [['ns_1@10.1.5.227'| 0.294921875]], [], [['ns_1@10.1.5.229'| 0.294921875]], [], [], [], [], [], []}}}} INFO REPORT <0.6366.0> 2011-05-10 14:08:40 =============================================================================== vbucketmigrator<0.6366.0>: Validate bucket states vbucketmigrator<0.6366.0>: 151 ok INFO REPORT <0.6371.0> 2011-05-10 14:08:41 =============================================================================== vbucketmigrator<0.6371.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.6371.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6371.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6371.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.6371.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6371.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6371.0>: Starting to move bucket 152 vbucketmigrator<0.6371.0>: Bucket 152 moved to the next server INFO REPORT <0.104.0> 2011-05-10 14:08:41 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.474.0> 2011-05-10 14:08:43 =============================================================================== ns_1@10.1.5.227:ns_doctor:82: Current node statuses: [{'ns_1@10.1.5.227', [{last_heard,{1305,61722,411299}}, {active_buckets,["default"]}, {memory, [{total,40251864}, {processes,19569816}, {processes_used,19558040}, {system,20682048}, {atom,819441}, {atom_used,806779}, {binary,122648}, {code,8552890}, {ets,1565712}]}, {cluster_compatibility_version,1}, {version, [{kernel,"2.13.4"}, {mnesia,"4.4.12"}, {stdlib,"1.16.4"}, {os_mon,"2.2.4"}, {inets,"5.2"}, {ns_server,"1.6.5.4r"}, {menelaus,"1.6.5.4r"}, {sasl,"2.1.8"}]}, {system_arch,"x86_64-unknown-linux-gnu"}, {wall_clock,1114}, {memory_data,{2058731520,677580800,{<0.12.0>,6657168}}}, {disk_data, [{"/",11204616,49},{"/boot",101086,14},{"/dev/shm",1005240,0}]}, {meminfo, <<"MemTotal: 2010480 kB\nMemFree: 1348036 kB\nBuffers: 65476 kB\nCached: 342488 kB\nSwapCached: 20084 kB\nActive: 350196 kB\nInactive: 199364 kB\nHighTotal: 0 kB\nHighFree: 0 kB\nLowTotal: 2010480 kB\nLowFree: 1348036 kB\nSwapTotal: 4030456 kB\nSwapFree: 3952288 kB\nDirty: 208 kB\nWriteback: 0 kB\nAnonPages: 139376 kB\nMapped: 15736 kB\nSlab: 86696 kB\nPageTables: 6864 kB\nNFS_Unstable: 0 kB\nBounce: 0 kB\nCommitLimit: 5035696 kB\nCommitted_AS: 554164 kB\nVmallocTotal: 34359738367 kB\nVmallocUsed: 265484 kB\nVmallocChunk: 34359471011 kB\nHugePages_Total: 0\nHugePages_Free: 0\nHugePages_Rsvd: 0\nHugepagesize: 2048 kB\n">>}, {replication,[{"default",1.0}]}, {system_memory_data, [{system_total_memory,2058731520}, {free_swap,4047142912}, {total_swap,4127186944}, {cached_memory,350707712}, {buffered_memory,67047424}, {free_memory,1380388864}, {total_memory,2058731520}]}, {statistics, [{wall_clock,{1110579,1}}, {context_switches,{367701,0}}, {garbage_collection,{74785,372017543,0}}, {io,{{input,87467139},{output,16897752}}}, {reductions,{107958761,823504}}, {run_queue,0}, {runtime,{15420,120}}]}]}, {'ns_1@10.1.5.229', [{last_heard,{1305,61722,242214}}, {active_buckets,["default"]}, {memory, [{total,32890664}, {processes,13310792}, {processes_used,13287272}, {system,19579872}, {atom,778121}, {atom_used,752794}, {binary,228672}, {code,7791241}, {ets,1494424}]}, {cluster_compatibility_version,1}, {version, [{kernel,"2.13.4"}, {mnesia,"4.4.12"}, {stdlib,"1.16.4"}, {os_mon,"2.2.4"}, {ns_server,"1.6.5.4r"}, {menelaus,"1.6.5.4r"}, {sasl,"2.1.8"}]}, {system_arch,"x86_64-unknown-linux-gnu"}, {wall_clock,1077}, {memory_data,{2058731520,720064512,{<9352.12.0>,4114680}}}, {disk_data, [{"/",11204616,44},{"/boot",101086,14},{"/dev/shm",1005240,0}]}, {meminfo, <<"MemTotal: 2010480 kB\nMemFree: 1306424 kB\nBuffers: 85932 kB\nCached: 340840 kB\nSwapCached: 0 kB\nActive: 368696 kB\nInactive: 218288 kB\nHighTotal: 0 kB\nHighFree: 0 kB\nLowTotal: 2010480 kB\nLowFree: 1306424 kB\nSwapTotal: 4030456 kB\nSwapFree: 4030388 kB\nDirty: 204 kB\nWriteback: 0 kB\nAnonPages: 160220 kB\nMapped: 28104 kB\nSlab: 91908 kB\nPageTables: 5872 kB\nNFS_Unstable: 0 kB\nBounce: 0 kB\nCommitLimit: 5035696 kB\nCommitted_AS: 485760 kB\nVmallocTotal: 34359738367 kB\nVmallocUsed: 265484 kB\nVmallocChunk: 34359471011 kB\nHugePages_Total: 0\nHugePages_Free: 0\nHugePages_Rsvd: 0\nHugepagesize: 2048 kB\n">>}, {replication,[{"default",1.0}]}, {system_memory_data, [{system_total_memory,2058731520}, {free_swap,4127117312}, {total_swap,4127186944}, {cached_memory,349020160}, {buffered_memory,87994368}, {free_memory,1337778176}, {total_memory,2058731520}]}, {statistics, [{wall_clock,{1069806,1}}, {context_switches,{94840,0}}, {garbage_collection,{20723,79907752,0}}, {io,{{input,10642744},{output,6634553}}}, {reductions,{26697749,240528}}, {run_queue,0}, {runtime,{4390,40}}]}]}] INFO REPORT <0.469.0> 2011-05-10 14:08:44 =============================================================================== ns_1@10.1.5.227:stats_collector:71: Stats for bucket "default": auth_cmds 185 auth_errors 0 bucket_conns 34 bytes_read 1634648 bytes_written 10008802 cas_badval 0 cas_hits 0 cas_misses 0 cmd_flush 1 cmd_get 0 cmd_set 16004 conn_yields 0 connection_structures 44 curr_connections 44 curr_items 6821 curr_items_tot 8040 daemon_connections 10 decr_hits 0 decr_misses 0 delete_hits 0 delete_misses 0 ep_bg_fetched 0 ep_commit_num 53 ep_commit_time 0 ep_commit_time_total 4 ep_data_age 13 ep_data_age_highwat 13 ep_db_cleaner_status complete ep_db_strategy multiMTVBDB ep_dbinit 1 ep_dbname /var/opt/membase/1.6.5.4r/data/ns_1/default ep_dbshards 4 ep_expired 0 ep_flush_duration 1 ep_flush_duration_highwat 8 ep_flush_duration_total 12 ep_flush_preempts 0 ep_flusher_state running ep_flusher_todo 0 ep_io_num_read 1219 ep_io_num_write 15741 ep_io_read_bytes 69167 ep_io_write_bytes 1011432 ep_item_begin_failed 0 ep_item_commit_failed 0 ep_item_flush_expired 0 ep_item_flush_failed 0 ep_kv_size 1181400 ep_max_data_size 1096810496 ep_max_txn_size 1000 ep_mem_high_wat 822607872 ep_mem_low_wat 658086297 ep_min_data_age 0 ep_num_active_non_resident 0 ep_num_eject_failures 0 ep_num_eject_replicas 0 ep_num_expiry_pager_runs 0 ep_num_non_resident 0 ep_num_not_my_vbuckets 0 ep_num_pager_runs 0 ep_num_value_ejects 0 ep_oom_errors 0 ep_overhead 25854448 ep_pending_ops 0 ep_pending_ops_max 0 ep_pending_ops_max_duration 0 ep_pending_ops_total 0 ep_queue_age_cap 900 ep_queue_size 0 ep_storage_age 0 ep_storage_age_highwat 10 ep_storage_type featured ep_store_max_concurrency 10 ep_store_max_readers 9 ep_store_max_readwrite 1 ep_tap_bg_fetch_requeued 0 ep_tap_bg_fetched 1219 ep_tap_keepalive 300 ep_tmp_oom_errors 0 ep_too_old 0 ep_too_young 0 ep_total_cache_size 1181400 ep_total_del_items 0 ep_total_enqueued 16035 ep_total_new_items 15741 ep_total_persisted 15741 ep_vbucket_del 0 ep_vbucket_del_fail 0 ep_version 1.6.5.3.1_13_g81a6cf1 ep_warmed_up 0 ep_warmup true ep_warmup_dups 0 ep_warmup_oom 0 ep_warmup_thread complete ep_warmup_time 15662 get_hits 0 get_misses 0 incr_hits 0 incr_misses 0 libevent 1.4.13-stable limit_maxbytes 67108864 mem_used 27035848 pid 16183 pointer_size 64 rejected_conns 0 rusage_system 1.440780 rusage_user 3.960397 tap_connect_received 153 tap_mutation_sent 2438 tap_opaque_sent 153 tap_vbucket_set_sent 306 threads 4 time 1305061723 total_connections 196 uptime 1117 version 1.4.4_382_g9df3289 INFO REPORT <0.85.0> 2011-05-10 14:08:44 =============================================================================== Pulling config from: 'ns_1@10.1.5.229' INFO REPORT <0.6371.0> 2011-05-10 14:08:46 =============================================================================== vbucketmigrator<0.6371.0>: Validate bucket states vbucketmigrator<0.6371.0>: 152 ok INFO REPORT <0.6390.0> 2011-05-10 14:08:47 =============================================================================== vbucketmigrator<0.6390.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.6390.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6390.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6390.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.6390.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6390.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6390.0>: Starting to move bucket 153 vbucketmigrator<0.6390.0>: Bucket 153 moved to the next server INFO REPORT <0.104.0> 2011-05-10 14:08:47 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.109.0> 2011-05-10 14:08:48 =============================================================================== ns_1@10.1.5.227:ns_orchestrator:164: Skipping janitor in state rebalancing: {rebalancing_state, <0.577.0>, {dict, 2, 16, 16, 8, 80, 48, {[], [], [], [], [], [], [], [], [], [], [], [], [], [], [], []}, {{[], [], [], [], [], [], [], [['ns_1@10.1.5.227'| 0.298828125]], [], [['ns_1@10.1.5.229'| 0.298828125]], [], [], [], [], [], []}}}} INFO REPORT <0.6390.0> 2011-05-10 14:08:52 =============================================================================== vbucketmigrator<0.6390.0>: Validate bucket states vbucketmigrator<0.6390.0>: 153 ok INFO REPORT <0.6403.0> 2011-05-10 14:08:53 =============================================================================== vbucketmigrator<0.6403.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.6403.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6403.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6403.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.6403.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6403.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6403.0>: Starting to move bucket 154 vbucketmigrator<0.6403.0>: Bucket 154 moved to the next server INFO REPORT <0.104.0> 2011-05-10 14:08:53 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.6403.0> 2011-05-10 14:08:58 =============================================================================== vbucketmigrator<0.6403.0>: Validate bucket states vbucketmigrator<0.6403.0>: 154 ok INFO REPORT <0.109.0> 2011-05-10 14:08:58 =============================================================================== ns_1@10.1.5.227:ns_orchestrator:164: Skipping janitor in state rebalancing: {rebalancing_state, <0.577.0>, {dict, 2, 16, 16, 8, 80, 48, {[], [], [], [], [], [], [], [], [], [], [], [], [], [], [], []}, {{[], [], [], [], [], [], [], [['ns_1@10.1.5.227'| 0.302734375]], [], [['ns_1@10.1.5.229'| 0.302734375]], [], [], [], [], [], []}}}} INFO REPORT <0.6408.0> 2011-05-10 14:08:59 =============================================================================== vbucketmigrator<0.6408.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.6408.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6408.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6408.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.6408.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6408.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6408.0>: Starting to move bucket 155 vbucketmigrator<0.6408.0>: Bucket 155 moved to the next server INFO REPORT <0.104.0> 2011-05-10 14:08:59 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.6408.0> 2011-05-10 14:09:04 =============================================================================== vbucketmigrator<0.6408.0>: Validate bucket states vbucketmigrator<0.6408.0>: 155 ok INFO REPORT <0.6415.0> 2011-05-10 14:09:05 =============================================================================== vbucketmigrator<0.6415.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.6415.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6415.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6415.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.6415.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6415.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6415.0>: Starting to move bucket 156 vbucketmigrator<0.6415.0>: Bucket 156 moved to the next server INFO REPORT <0.104.0> 2011-05-10 14:09:05 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.109.0> 2011-05-10 14:09:08 =============================================================================== ns_1@10.1.5.227:ns_orchestrator:164: Skipping janitor in state rebalancing: {rebalancing_state, <0.577.0>, {dict, 2, 16, 16, 8, 80, 48, {[], [], [], [], [], [], [], [], [], [], [], [], [], [], [], []}, {{[], [], [], [], [], [], [], [['ns_1@10.1.5.227'| 0.3046875]], [], [['ns_1@10.1.5.229'| 0.3046875]], [], [], [], [], [], []}}}} INFO REPORT <0.6415.0> 2011-05-10 14:09:10 =============================================================================== vbucketmigrator<0.6415.0>: Validate bucket states vbucketmigrator<0.6415.0>: 156 ok INFO REPORT <0.6425.0> 2011-05-10 14:09:11 =============================================================================== vbucketmigrator<0.6425.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.6425.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6425.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6425.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.6425.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6425.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6425.0>: Starting to move bucket 157 vbucketmigrator<0.6425.0>: Bucket 157 moved to the next server INFO REPORT <0.104.0> 2011-05-10 14:09:11 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.6425.0> 2011-05-10 14:09:16 =============================================================================== vbucketmigrator<0.6425.0>: Validate bucket states vbucketmigrator<0.6425.0>: 157 ok INFO REPORT <0.6433.0> 2011-05-10 14:09:17 =============================================================================== vbucketmigrator<0.6433.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.6433.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6433.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6433.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.6433.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6433.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6433.0>: Starting to move bucket 158 vbucketmigrator<0.6433.0>: Bucket 158 moved to the next server INFO REPORT <0.104.0> 2011-05-10 14:09:17 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.109.0> 2011-05-10 14:09:18 =============================================================================== ns_1@10.1.5.227:ns_orchestrator:164: Skipping janitor in state rebalancing: {rebalancing_state, <0.577.0>, {dict, 2, 16, 16, 8, 80, 48, {[], [], [], [], [], [], [], [], [], [], [], [], [], [], [], []}, {{[], [], [], [], [], [], [], [['ns_1@10.1.5.227'| 0.30859375]], [], [['ns_1@10.1.5.229'| 0.30859375]], [], [], [], [], [], []}}}} INFO REPORT <0.6433.0> 2011-05-10 14:09:22 =============================================================================== vbucketmigrator<0.6433.0>: Validate bucket states vbucketmigrator<0.6433.0>: 158 ok INFO REPORT <0.85.0> 2011-05-10 14:09:22 =============================================================================== Pulling config from: 'ns_1@10.1.5.229' INFO REPORT <0.6438.0> 2011-05-10 14:09:23 =============================================================================== vbucketmigrator<0.6438.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.6438.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6438.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6438.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.6438.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6438.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6438.0>: Starting to move bucket 159 vbucketmigrator<0.6438.0>: Bucket 159 moved to the next server INFO REPORT <0.104.0> 2011-05-10 14:09:23 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.6438.0> 2011-05-10 14:09:28 =============================================================================== vbucketmigrator<0.6438.0>: Validate bucket states vbucketmigrator<0.6438.0>: 159 ok INFO REPORT <0.109.0> 2011-05-10 14:09:28 =============================================================================== ns_1@10.1.5.227:ns_orchestrator:164: Skipping janitor in state rebalancing: {rebalancing_state, <0.577.0>, {dict, 2, 16, 16, 8, 80, 48, {[], [], [], [], [], [], [], [], [], [], [], [], [], [], [], []}, {{[], [], [], [], [], [], [], [['ns_1@10.1.5.227'| 0.3125]], [], [['ns_1@10.1.5.229'| 0.3125]], [], [], [], [], [], []}}}} INFO REPORT <0.6443.0> 2011-05-10 14:09:29 =============================================================================== vbucketmigrator<0.6443.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.6443.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6443.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6443.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.6443.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6443.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6443.0>: Starting to move bucket 160 vbucketmigrator<0.6443.0>: Bucket 160 moved to the next server INFO REPORT <0.104.0> 2011-05-10 14:09:29 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.6443.0> 2011-05-10 14:09:34 =============================================================================== vbucketmigrator<0.6443.0>: Validate bucket states vbucketmigrator<0.6443.0>: 160 ok INFO REPORT <0.6457.0> 2011-05-10 14:09:35 =============================================================================== vbucketmigrator<0.6457.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.6457.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6457.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6457.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.6457.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6457.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6457.0>: Starting to move bucket 161 vbucketmigrator<0.6457.0>: Bucket 161 moved to the next server INFO REPORT <0.104.0> 2011-05-10 14:09:35 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.109.0> 2011-05-10 14:09:38 =============================================================================== ns_1@10.1.5.227:ns_orchestrator:164: Skipping janitor in state rebalancing: {rebalancing_state, <0.577.0>, {dict, 2, 16, 16, 8, 80, 48, {[], [], [], [], [], [], [], [], [], [], [], [], [], [], [], []}, {{[], [], [], [], [], [], [], [['ns_1@10.1.5.227'| 0.314453125]], [], [['ns_1@10.1.5.229'| 0.314453125]], [], [], [], [], [], []}}}} INFO REPORT <0.6457.0> 2011-05-10 14:09:40 =============================================================================== vbucketmigrator<0.6457.0>: Validate bucket states vbucketmigrator<0.6457.0>: 161 ok INFO REPORT <0.6462.0> 2011-05-10 14:09:41 =============================================================================== vbucketmigrator<0.6462.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.6462.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6462.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6462.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.6462.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6462.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6462.0>: Starting to move bucket 162 vbucketmigrator<0.6462.0>: Bucket 162 moved to the next server INFO REPORT <0.104.0> 2011-05-10 14:09:41 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.474.0> 2011-05-10 14:09:43 =============================================================================== ns_1@10.1.5.227:ns_doctor:82: Current node statuses: [{'ns_1@10.1.5.227', [{last_heard,{1305,61782,411211}}, {active_buckets,["default"]}, {memory, [{total,39994384}, {processes,19269448}, {processes_used,19256808}, {system,20724936}, {atom,819441}, {atom_used,806779}, {binary,100720}, {code,8552890}, {ets,1629048}]}, {cluster_compatibility_version,1}, {version, [{kernel,"2.13.4"}, {mnesia,"4.4.12"}, {stdlib,"1.16.4"}, {os_mon,"2.2.4"}, {inets,"5.2"}, {ns_server,"1.6.5.4r"}, {menelaus,"1.6.5.4r"}, {sasl,"2.1.8"}]}, {system_arch,"x86_64-unknown-linux-gnu"}, {wall_clock,1174}, {memory_data,{2058731520,678088704,{<0.12.0>,6657168}}}, {disk_data, [{"/",11204616,49},{"/boot",101086,14},{"/dev/shm",1005240,0}]}, {meminfo, <<"MemTotal: 2010480 kB\nMemFree: 1347416 kB\nBuffers: 65776 kB\nCached: 342768 kB\nSwapCached: 20084 kB\nActive: 350212 kB\nInactive: 199908 kB\nHighTotal: 0 kB\nHighFree: 0 kB\nLowTotal: 2010480 kB\nLowFree: 1347416 kB\nSwapTotal: 4030456 kB\nSwapFree: 3952288 kB\nDirty: 148 kB\nWriteback: 0 kB\nAnonPages: 139364 kB\nMapped: 15740 kB\nSlab: 86716 kB\nPageTables: 6868 kB\nNFS_Unstable: 0 kB\nBounce: 0 kB\nCommitLimit: 5035696 kB\nCommitted_AS: 553456 kB\nVmallocTotal: 34359738367 kB\nVmallocUsed: 265484 kB\nVmallocChunk: 34359471011 kB\nHugePages_Total: 0\nHugePages_Free: 0\nHugePages_Rsvd: 0\nHugepagesize: 2048 kB\n">>}, {replication,[{"default",1.0}]}, {system_memory_data, [{system_total_memory,2058731520}, {free_swap,4047142912}, {total_swap,4127186944}, {cached_memory,350994432}, {buffered_memory,67354624}, {free_memory,1379753984}, {total_memory,2058731520}]}, {statistics, [{wall_clock,{1170579,1}}, {context_switches,{378439,0}}, {garbage_collection,{77676,391718314,0}}, {io,{{input,88366848},{output,18049901}}}, {reductions,{112527363,794660}}, {run_queue,0}, {runtime,{16200,140}}]}]}, {'ns_1@10.1.5.229', [{last_heard,{1305,61782,242178}}, {active_buckets,["default"]}, {memory, [{total,32924504}, {processes,13285232}, {processes_used,13261712}, {system,19639272}, {atom,778121}, {atom_used,752794}, {binary,223256}, {code,7791241}, {ets,1557800}]}, {cluster_compatibility_version,1}, {version, [{kernel,"2.13.4"}, {mnesia,"4.4.12"}, {stdlib,"1.16.4"}, {os_mon,"2.2.4"}, {ns_server,"1.6.5.4r"}, {menelaus,"1.6.5.4r"}, {sasl,"2.1.8"}]}, {system_arch,"x86_64-unknown-linux-gnu"}, {wall_clock,1137}, {memory_data,{2058731520,720953344,{<9352.12.0>,4114680}}}, {disk_data, [{"/",11204616,44},{"/boot",101086,14},{"/dev/shm",1005240,0}]}, {meminfo, <<"MemTotal: 2010480 kB\nMemFree: 1305408 kB\nBuffers: 86724 kB\nCached: 341088 kB\nSwapCached: 0 kB\nActive: 368340 kB\nInactive: 219292 kB\nHighTotal: 0 kB\nHighFree: 0 kB\nLowTotal: 2010480 kB\nLowFree: 1305408 kB\nSwapTotal: 4030456 kB\nSwapFree: 4030388 kB\nDirty: 1488 kB\nWriteback: 0 kB\nAnonPages: 159820 kB\nMapped: 28104 kB\nSlab: 92092 kB\nPageTables: 5876 kB\nNFS_Unstable: 0 kB\nBounce: 0 kB\nCommitLimit: 5035696 kB\nCommitted_AS: 484668 kB\nVmallocTotal: 34359738367 kB\nVmallocUsed: 265484 kB\nVmallocChunk: 34359471011 kB\nHugePages_Total: 0\nHugePages_Free: 0\nHugePages_Rsvd: 0\nHugepagesize: 2048 kB\n">>}, {replication,[{"default",1.0}]}, {system_memory_data, [{system_total_memory,2058731520}, {free_swap,4127117312}, {total_swap,4127186944}, {cached_memory,349274112}, {buffered_memory,88805376}, {free_memory,1336737792}, {total_memory,2058731520}]}, {statistics, [{wall_clock,{1129806,1}}, {context_switches,{100045,0}}, {garbage_collection,{22038,84573413,0}}, {io,{{input,11620136},{output,7622712}}}, {reductions,{28123831,232495}}, {run_queue,0}, {runtime,{4680,40}}]}]}] INFO REPORT <0.6462.0> 2011-05-10 14:09:46 =============================================================================== vbucketmigrator<0.6462.0>: Validate bucket states vbucketmigrator<0.6462.0>: 162 ok INFO REPORT <0.6466.0> 2011-05-10 14:09:47 =============================================================================== vbucketmigrator<0.6466.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.6466.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6466.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6466.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.6466.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6466.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6466.0>: Starting to move bucket 163 vbucketmigrator<0.6466.0>: Bucket 163 moved to the next server INFO REPORT <0.104.0> 2011-05-10 14:09:47 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.109.0> 2011-05-10 14:09:48 =============================================================================== ns_1@10.1.5.227:ns_orchestrator:164: Skipping janitor in state rebalancing: {rebalancing_state, <0.577.0>, {dict, 2, 16, 16, 8, 80, 48, {[], [], [], [], [], [], [], [], [], [], [], [], [], [], [], []}, {{[], [], [], [], [], [], [], [['ns_1@10.1.5.227'| 0.318359375]], [], [['ns_1@10.1.5.229'| 0.318359375]], [], [], [], [], [], []}}}} INFO REPORT <0.6466.0> 2011-05-10 14:09:52 =============================================================================== vbucketmigrator<0.6466.0>: Validate bucket states vbucketmigrator<0.6466.0>: 163 ok INFO REPORT <0.6479.0> 2011-05-10 14:09:53 =============================================================================== vbucketmigrator<0.6479.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.6479.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6479.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6479.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.6479.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6479.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6479.0>: Starting to move bucket 164 vbucketmigrator<0.6479.0>: Bucket 164 moved to the next server INFO REPORT <0.104.0> 2011-05-10 14:09:53 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.6479.0> 2011-05-10 14:09:58 =============================================================================== vbucketmigrator<0.6479.0>: Validate bucket states vbucketmigrator<0.6479.0>: 164 ok INFO REPORT <0.109.0> 2011-05-10 14:09:58 =============================================================================== ns_1@10.1.5.227:ns_orchestrator:164: Skipping janitor in state rebalancing: {rebalancing_state, <0.577.0>, {dict, 2, 16, 16, 8, 80, 48, {[], [], [], [], [], [], [], [], [], [], [], [], [], [], [], []}, {{[], [], [], [], [], [], [], [['ns_1@10.1.5.227'| 0.322265625]], [], [['ns_1@10.1.5.229'| 0.322265625]], [], [], [], [], [], []}}}} INFO REPORT <0.6485.0> 2011-05-10 14:09:59 =============================================================================== vbucketmigrator<0.6485.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.6485.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6485.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6485.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.6485.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6485.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6485.0>: Starting to move bucket 165 vbucketmigrator<0.6485.0>: Bucket 165 moved to the next server INFO REPORT <0.104.0> 2011-05-10 14:09:59 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.6485.0> 2011-05-10 14:10:04 =============================================================================== vbucketmigrator<0.6485.0>: Validate bucket states vbucketmigrator<0.6485.0>: 165 ok INFO REPORT <0.6491.0> 2011-05-10 14:10:05 =============================================================================== vbucketmigrator<0.6491.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.6491.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6491.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6491.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.6491.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6491.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6491.0>: Starting to move bucket 166 vbucketmigrator<0.6491.0>: Bucket 166 moved to the next server INFO REPORT <0.104.0> 2011-05-10 14:10:05 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.109.0> 2011-05-10 14:10:08 =============================================================================== ns_1@10.1.5.227:ns_orchestrator:164: Skipping janitor in state rebalancing: {rebalancing_state, <0.577.0>, {dict, 2, 16, 16, 8, 80, 48, {[], [], [], [], [], [], [], [], [], [], [], [], [], [], [], []}, {{[], [], [], [], [], [], [], [['ns_1@10.1.5.227'| 0.32421875]], [], [['ns_1@10.1.5.229'| 0.32421875]], [], [], [], [], [], []}}}} INFO REPORT <0.6491.0> 2011-05-10 14:10:10 =============================================================================== vbucketmigrator<0.6491.0>: Validate bucket states vbucketmigrator<0.6491.0>: 166 ok INFO REPORT <0.6501.0> 2011-05-10 14:10:11 =============================================================================== vbucketmigrator<0.6501.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.6501.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6501.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6501.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.6501.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6501.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6501.0>: Starting to move bucket 167 vbucketmigrator<0.6501.0>: Bucket 167 moved to the next server INFO REPORT <0.104.0> 2011-05-10 14:10:11 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.6501.0> 2011-05-10 14:10:16 =============================================================================== vbucketmigrator<0.6501.0>: Validate bucket states vbucketmigrator<0.6501.0>: 167 ok INFO REPORT <0.6509.0> 2011-05-10 14:10:17 =============================================================================== vbucketmigrator<0.6509.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.6509.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6509.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6509.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.6509.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6509.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6509.0>: Starting to move bucket 168 vbucketmigrator<0.6509.0>: Bucket 168 moved to the next server INFO REPORT <0.104.0> 2011-05-10 14:10:17 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.109.0> 2011-05-10 14:10:18 =============================================================================== ns_1@10.1.5.227:ns_orchestrator:164: Skipping janitor in state rebalancing: {rebalancing_state, <0.577.0>, {dict, 2, 16, 16, 8, 80, 48, {[], [], [], [], [], [], [], [], [], [], [], [], [], [], [], []}, {{[], [], [], [], [], [], [], [['ns_1@10.1.5.227'| 0.328125]], [], [['ns_1@10.1.5.229'| 0.328125]], [], [], [], [], [], []}}}} INFO REPORT <0.85.0> 2011-05-10 14:10:19 =============================================================================== Pulling config from: 'ns_1@10.1.5.229' INFO REPORT <0.6509.0> 2011-05-10 14:10:22 =============================================================================== vbucketmigrator<0.6509.0>: Validate bucket states vbucketmigrator<0.6509.0>: 168 ok INFO REPORT <0.6514.0> 2011-05-10 14:10:23 =============================================================================== vbucketmigrator<0.6514.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.6514.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6514.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6514.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.6514.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6514.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6514.0>: Starting to move bucket 169 vbucketmigrator<0.6514.0>: Bucket 169 moved to the next server INFO REPORT <0.104.0> 2011-05-10 14:10:23 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.469.0> 2011-05-10 14:10:24 =============================================================================== ns_1@10.1.5.227:stats_collector:71: Stats for bucket "default": auth_cmds 202 auth_errors 0 bucket_conns 35 bytes_read 1641742 bytes_written 12651869 cas_badval 0 cas_hits 0 cas_misses 0 cmd_flush 1 cmd_get 0 cmd_set 16004 conn_yields 0 connection_structures 45 curr_connections 45 curr_items 6696 curr_items_tot 8040 daemon_connections 10 decr_hits 0 decr_misses 0 delete_hits 0 delete_misses 0 ep_bg_fetched 0 ep_commit_num 53 ep_commit_time 0 ep_commit_time_total 4 ep_data_age 13 ep_data_age_highwat 13 ep_db_cleaner_status complete ep_db_strategy multiMTVBDB ep_dbinit 1 ep_dbname /var/opt/membase/1.6.5.4r/data/ns_1/default ep_dbshards 4 ep_expired 0 ep_flush_duration 1 ep_flush_duration_highwat 8 ep_flush_duration_total 12 ep_flush_preempts 0 ep_flusher_state running ep_flusher_todo 0 ep_io_num_read 1344 ep_io_num_write 15741 ep_io_read_bytes 76274 ep_io_write_bytes 1011432 ep_item_begin_failed 0 ep_item_commit_failed 0 ep_item_flush_expired 0 ep_item_flush_failed 0 ep_kv_size 1181400 ep_max_data_size 1096810496 ep_max_txn_size 1000 ep_mem_high_wat 822607872 ep_mem_low_wat 658086297 ep_min_data_age 0 ep_num_active_non_resident 0 ep_num_eject_failures 0 ep_num_eject_replicas 0 ep_num_expiry_pager_runs 0 ep_num_non_resident 0 ep_num_not_my_vbuckets 0 ep_num_pager_runs 0 ep_num_value_ejects 0 ep_oom_errors 0 ep_overhead 25854448 ep_pending_ops 0 ep_pending_ops_max 0 ep_pending_ops_max_duration 0 ep_pending_ops_total 0 ep_queue_age_cap 900 ep_queue_size 0 ep_storage_age 0 ep_storage_age_highwat 10 ep_storage_type featured ep_store_max_concurrency 10 ep_store_max_readers 9 ep_store_max_readwrite 1 ep_tap_bg_fetch_requeued 0 ep_tap_bg_fetched 1344 ep_tap_keepalive 300 ep_tmp_oom_errors 0 ep_too_old 0 ep_too_young 0 ep_total_cache_size 1181400 ep_total_del_items 0 ep_total_enqueued 16035 ep_total_new_items 15741 ep_total_persisted 15741 ep_vbucket_del 0 ep_vbucket_del_fail 0 ep_version 1.6.5.3.1_13_g81a6cf1 ep_warmed_up 0 ep_warmup true ep_warmup_dups 0 ep_warmup_oom 0 ep_warmup_thread complete ep_warmup_time 15662 get_hits 0 get_misses 0 incr_hits 0 incr_misses 0 libevent 1.4.13-stable limit_maxbytes 67108864 mem_used 27035848 pid 16183 pointer_size 64 rejected_conns 0 rusage_system 1.469776 rusage_user 4.169366 tap_connect_received 170 tap_mutation_sent 2688 tap_opaque_sent 170 tap_vbucket_set_sent 340 threads 4 time 1305061823 total_connections 222 uptime 1217 version 1.4.4_382_g9df3289 INFO REPORT <0.6514.0> 2011-05-10 14:10:28 =============================================================================== vbucketmigrator<0.6514.0>: Validate bucket states vbucketmigrator<0.6514.0>: 169 ok INFO REPORT <0.109.0> 2011-05-10 14:10:28 =============================================================================== ns_1@10.1.5.227:ns_orchestrator:164: Skipping janitor in state rebalancing: {rebalancing_state, <0.577.0>, {dict, 2, 16, 16, 8, 80, 48, {[], [], [], [], [], [], [], [], [], [], [], [], [], [], [], []}, {{[], [], [], [], [], [], [], [['ns_1@10.1.5.227'| 0.33203125]], [], [['ns_1@10.1.5.229'| 0.33203125]], [], [], [], [], [], []}}}} INFO REPORT <0.6519.0> 2011-05-10 14:10:29 =============================================================================== vbucketmigrator<0.6519.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.6519.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6519.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6519.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.6519.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6519.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6519.0>: Starting to move bucket 170 vbucketmigrator<0.6519.0>: Bucket 170 moved to the next server INFO REPORT <0.104.0> 2011-05-10 14:10:29 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.6519.0> 2011-05-10 14:10:34 =============================================================================== vbucketmigrator<0.6519.0>: Validate bucket states vbucketmigrator<0.6519.0>: 170 ok INFO REPORT <0.6533.0> 2011-05-10 14:10:35 =============================================================================== vbucketmigrator<0.6533.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.6533.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6533.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6533.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.6533.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6533.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6533.0>: Starting to move bucket 171 vbucketmigrator<0.6533.0>: Bucket 171 moved to the next server INFO REPORT <0.104.0> 2011-05-10 14:10:35 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.109.0> 2011-05-10 14:10:38 =============================================================================== ns_1@10.1.5.227:ns_orchestrator:164: Skipping janitor in state rebalancing: {rebalancing_state, <0.577.0>, {dict, 2, 16, 16, 8, 80, 48, {[], [], [], [], [], [], [], [], [], [], [], [], [], [], [], []}, {{[], [], [], [], [], [], [], [['ns_1@10.1.5.227'| 0.333984375]], [], [['ns_1@10.1.5.229'| 0.333984375]], [], [], [], [], [], []}}}} INFO REPORT <0.6533.0> 2011-05-10 14:10:40 =============================================================================== vbucketmigrator<0.6533.0>: Validate bucket states vbucketmigrator<0.6533.0>: 171 ok INFO REPORT <0.6538.0> 2011-05-10 14:10:41 =============================================================================== vbucketmigrator<0.6538.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.6538.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6538.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6538.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.6538.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6538.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6538.0>: Starting to move bucket 172 vbucketmigrator<0.6538.0>: Bucket 172 moved to the next server INFO REPORT <0.104.0> 2011-05-10 14:10:41 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.474.0> 2011-05-10 14:10:43 =============================================================================== ns_1@10.1.5.227:ns_doctor:82: Current node statuses: [{'ns_1@10.1.5.227', [{last_heard,{1305,61842,411250}}, {active_buckets,["default"]}, {memory, [{total,40274256}, {processes,19524304}, {processes_used,19512528}, {system,20749952}, {atom,819441}, {atom_used,806779}, {binary,128176}, {code,8552890}, {ets,1624792}]}, {cluster_compatibility_version,1}, {version, [{kernel,"2.13.4"}, {mnesia,"4.4.12"}, {stdlib,"1.16.4"}, {os_mon,"2.2.4"}, {inets,"5.2"}, {ns_server,"1.6.5.4r"}, {menelaus,"1.6.5.4r"}, {sasl,"2.1.8"}]}, {system_arch,"x86_64-unknown-linux-gnu"}, {wall_clock,1234}, {memory_data,{2058731520,678977536,{<0.12.0>,6657168}}}, {disk_data, [{"/",11204616,49},{"/boot",101086,14},{"/dev/shm",1005240,0}]}, {meminfo, <<"MemTotal: 2010480 kB\nMemFree: 1346796 kB\nBuffers: 66000 kB\nCached: 343032 kB\nSwapCached: 20084 kB\nActive: 350556 kB\nInactive: 200152 kB\nHighTotal: 0 kB\nHighFree: 0 kB\nLowTotal: 2010480 kB\nLowFree: 1346796 kB\nSwapTotal: 4030456 kB\nSwapFree: 3952288 kB\nDirty: 400 kB\nWriteback: 0 kB\nAnonPages: 139480 kB\nMapped: 15740 kB\nSlab: 86760 kB\nPageTables: 6868 kB\nNFS_Unstable: 0 kB\nBounce: 0 kB\nCommitLimit: 5035696 kB\nCommitted_AS: 553456 kB\nVmallocTotal: 34359738367 kB\nVmallocUsed: 265484 kB\nVmallocChunk: 34359471011 kB\nHugePages_Total: 0\nHugePages_Free: 0\nHugePages_Rsvd: 0\nHugepagesize: 2048 kB\n">>}, {replication,[{"default",1.0}]}, {system_memory_data, [{system_total_memory,2058731520}, {free_swap,4047142912}, {total_swap,4127186944}, {cached_memory,351264768}, {buffered_memory,67584000}, {free_memory,1379119104}, {total_memory,2058731520}]}, {statistics, [{wall_clock,{1230579,1}}, {context_switches,{388800,0}}, {garbage_collection,{80349,411272551,0}}, {io,{{input,88496128},{output,18484332}}}, {reductions,{117095737,802343}}, {run_queue,0}, {runtime,{16950,130}}]}]}, {'ns_1@10.1.5.229', [{last_heard,{1305,61842,242379}}, {active_buckets,["default"]}, {memory, [{total,32888824}, {processes,13275472}, {processes_used,13251952}, {system,19613352}, {atom,778121}, {atom_used,752794}, {binary,201520}, {code,7791241}, {ets,1553088}]}, {cluster_compatibility_version,1}, {version, [{kernel,"2.13.4"}, {mnesia,"4.4.12"}, {stdlib,"1.16.4"}, {os_mon,"2.2.4"}, {ns_server,"1.6.5.4r"}, {menelaus,"1.6.5.4r"}, {sasl,"2.1.8"}]}, {system_arch,"x86_64-unknown-linux-gnu"}, {wall_clock,1197}, {memory_data,{2058731520,722358272,{<9352.12.0>,4114680}}}, {disk_data, [{"/",11204616,44},{"/boot",101086,14},{"/dev/shm",1005240,0}]}, {meminfo, <<"MemTotal: 2010480 kB\nMemFree: 1303696 kB\nBuffers: 87400 kB\nCached: 341368 kB\nSwapCached: 0 kB\nActive: 369920 kB\nInactive: 219928 kB\nHighTotal: 0 kB\nHighFree: 0 kB\nLowTotal: 2010480 kB\nLowFree: 1303696 kB\nSwapTotal: 4030456 kB\nSwapFree: 4030388 kB\nDirty: 120 kB\nWriteback: 0 kB\nAnonPages: 161072 kB\nMapped: 28104 kB\nSlab: 92004 kB\nPageTables: 5872 kB\nNFS_Unstable: 0 kB\nBounce: 0 kB\nCommitLimit: 5035696 kB\nCommitted_AS: 486784 kB\nVmallocTotal: 34359738367 kB\nVmallocUsed: 265484 kB\nVmallocChunk: 34359471011 kB\nHugePages_Total: 0\nHugePages_Free: 0\nHugePages_Rsvd: 0\nHugepagesize: 2048 kB\n">>}, {replication,[{"default",1.0}]}, {system_memory_data, [{system_total_memory,2058731520}, {free_swap,4127117312}, {total_swap,4127186944}, {cached_memory,349560832}, {buffered_memory,89497600}, {free_memory,1334984704}, {total_memory,2058731520}]}, {statistics, [{wall_clock,{1189806,1}}, {context_switches,{104849,0}}, {garbage_collection,{23072,89046438,0}}, {io,{{input,11856907},{output,7884306}}}, {reductions,{29494564,235668}}, {run_queue,0}, {runtime,{4890,40}}]}]}] INFO REPORT <0.6538.0> 2011-05-10 14:10:46 =============================================================================== vbucketmigrator<0.6538.0>: Validate bucket states vbucketmigrator<0.6538.0>: 172 ok INFO REPORT <0.6542.0> 2011-05-10 14:10:47 =============================================================================== vbucketmigrator<0.6542.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.6542.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6542.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6542.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.6542.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6542.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6542.0>: Starting to move bucket 173 vbucketmigrator<0.6542.0>: Bucket 173 moved to the next server INFO REPORT <0.104.0> 2011-05-10 14:10:47 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.109.0> 2011-05-10 14:10:48 =============================================================================== ns_1@10.1.5.227:ns_orchestrator:164: Skipping janitor in state rebalancing: {rebalancing_state, <0.577.0>, {dict, 2, 16, 16, 8, 80, 48, {[], [], [], [], [], [], [], [], [], [], [], [], [], [], [], []}, {{[], [], [], [], [], [], [], [['ns_1@10.1.5.227'| 0.337890625]], [], [['ns_1@10.1.5.229'| 0.337890625]], [], [], [], [], [], []}}}} INFO REPORT <0.6542.0> 2011-05-10 14:10:52 =============================================================================== vbucketmigrator<0.6542.0>: Validate bucket states vbucketmigrator<0.6542.0>: 173 ok INFO REPORT <0.6555.0> 2011-05-10 14:10:53 =============================================================================== vbucketmigrator<0.6555.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.6555.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6555.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6555.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.6555.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6555.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6555.0>: Starting to move bucket 174 vbucketmigrator<0.6555.0>: Bucket 174 moved to the next server INFO REPORT <0.104.0> 2011-05-10 14:10:53 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.6555.0> 2011-05-10 14:10:58 =============================================================================== vbucketmigrator<0.6555.0>: Validate bucket states vbucketmigrator<0.6555.0>: 174 ok INFO REPORT <0.109.0> 2011-05-10 14:10:58 =============================================================================== ns_1@10.1.5.227:ns_orchestrator:164: Skipping janitor in state rebalancing: {rebalancing_state, <0.577.0>, {dict, 2, 16, 16, 8, 80, 48, {[], [], [], [], [], [], [], [], [], [], [], [], [], [], [], []}, {{[], [], [], [], [], [], [], [['ns_1@10.1.5.227'| 0.341796875]], [], [['ns_1@10.1.5.229'| 0.341796875]], [], [], [], [], [], []}}}} INFO REPORT <0.6560.0> 2011-05-10 14:10:59 =============================================================================== vbucketmigrator<0.6560.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.6560.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6560.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6560.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.6560.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6560.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6560.0>: Starting to move bucket 175 vbucketmigrator<0.6560.0>: Bucket 175 moved to the next server INFO REPORT <0.104.0> 2011-05-10 14:10:59 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.85.0> 2011-05-10 14:10:59 =============================================================================== Pulling config from: 'ns_1@10.1.5.229' INFO REPORT <0.6560.0> 2011-05-10 14:11:04 =============================================================================== vbucketmigrator<0.6560.0>: Validate bucket states vbucketmigrator<0.6560.0>: 175 ok INFO REPORT <0.6566.0> 2011-05-10 14:11:05 =============================================================================== vbucketmigrator<0.6566.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.6566.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6566.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6566.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.6566.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6566.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6566.0>: Starting to move bucket 176 vbucketmigrator<0.6566.0>: Bucket 176 moved to the next server INFO REPORT <0.104.0> 2011-05-10 14:11:05 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.109.0> 2011-05-10 14:11:08 =============================================================================== ns_1@10.1.5.227:ns_orchestrator:164: Skipping janitor in state rebalancing: {rebalancing_state, <0.577.0>, {dict, 2, 16, 16, 8, 80, 48, {[], [], [], [], [], [], [], [], [], [], [], [], [], [], [], []}, {{[], [], [], [], [], [], [], [['ns_1@10.1.5.227'| 0.34375]], [], [['ns_1@10.1.5.229'| 0.34375]], [], [], [], [], [], []}}}} INFO REPORT <0.6566.0> 2011-05-10 14:11:10 =============================================================================== vbucketmigrator<0.6566.0>: Validate bucket states vbucketmigrator<0.6566.0>: 176 ok INFO REPORT <0.6576.0> 2011-05-10 14:11:11 =============================================================================== vbucketmigrator<0.6576.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.6576.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6576.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6576.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.6576.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6576.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6576.0>: Starting to move bucket 177 vbucketmigrator<0.6576.0>: Bucket 177 moved to the next server INFO REPORT <0.104.0> 2011-05-10 14:11:11 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.6576.0> 2011-05-10 14:11:16 =============================================================================== vbucketmigrator<0.6576.0>: Validate bucket states vbucketmigrator<0.6576.0>: 177 ok INFO REPORT <0.6584.0> 2011-05-10 14:11:17 =============================================================================== vbucketmigrator<0.6584.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.6584.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6584.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6584.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.6584.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6584.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6584.0>: Starting to move bucket 178 vbucketmigrator<0.6584.0>: Bucket 178 moved to the next server INFO REPORT <0.104.0> 2011-05-10 14:11:17 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.109.0> 2011-05-10 14:11:18 =============================================================================== ns_1@10.1.5.227:ns_orchestrator:164: Skipping janitor in state rebalancing: {rebalancing_state, <0.577.0>, {dict, 2, 16, 16, 8, 80, 48, {[], [], [], [], [], [], [], [], [], [], [], [], [], [], [], []}, {{[], [], [], [], [], [], [], [['ns_1@10.1.5.227'| 0.34765625]], [], [['ns_1@10.1.5.229'| 0.34765625]], [], [], [], [], [], []}}}} INFO REPORT <0.6584.0> 2011-05-10 14:11:22 =============================================================================== vbucketmigrator<0.6584.0>: Validate bucket states vbucketmigrator<0.6584.0>: 178 ok INFO REPORT <0.6589.0> 2011-05-10 14:11:23 =============================================================================== vbucketmigrator<0.6589.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.6589.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6589.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6589.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.6589.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6589.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6589.0>: Starting to move bucket 179 vbucketmigrator<0.6589.0>: Bucket 179 moved to the next server INFO REPORT <0.104.0> 2011-05-10 14:11:23 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.6589.0> 2011-05-10 14:11:28 =============================================================================== vbucketmigrator<0.6589.0>: Validate bucket states vbucketmigrator<0.6589.0>: 179 ok INFO REPORT <0.109.0> 2011-05-10 14:11:28 =============================================================================== ns_1@10.1.5.227:ns_orchestrator:164: Skipping janitor in state rebalancing: {rebalancing_state, <0.577.0>, {dict, 2, 16, 16, 8, 80, 48, {[], [], [], [], [], [], [], [], [], [], [], [], [], [], [], []}, {{[], [], [], [], [], [], [], [['ns_1@10.1.5.227'| 0.3515625]], [], [['ns_1@10.1.5.229'| 0.3515625]], [], [], [], [], [], []}}}} INFO REPORT <0.6594.0> 2011-05-10 14:11:29 =============================================================================== vbucketmigrator<0.6594.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.6594.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6594.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6594.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.6594.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6594.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6594.0>: Starting to move bucket 180 vbucketmigrator<0.6594.0>: Bucket 180 moved to the next server INFO REPORT <0.104.0> 2011-05-10 14:11:29 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.85.0> 2011-05-10 14:11:29 =============================================================================== Pulling config from: 'ns_1@10.1.5.229' INFO REPORT <0.6594.0> 2011-05-10 14:11:34 =============================================================================== vbucketmigrator<0.6594.0>: Validate bucket states vbucketmigrator<0.6594.0>: 180 ok INFO REPORT <0.6608.0> 2011-05-10 14:11:35 =============================================================================== vbucketmigrator<0.6608.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.6608.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6608.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6608.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.6608.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6608.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6608.0>: Starting to move bucket 181 vbucketmigrator<0.6608.0>: Bucket 181 moved to the next server INFO REPORT <0.104.0> 2011-05-10 14:11:35 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.109.0> 2011-05-10 14:11:38 =============================================================================== ns_1@10.1.5.227:ns_orchestrator:164: Skipping janitor in state rebalancing: {rebalancing_state, <0.577.0>, {dict, 2, 16, 16, 8, 80, 48, {[], [], [], [], [], [], [], [], [], [], [], [], [], [], [], []}, {{[], [], [], [], [], [], [], [['ns_1@10.1.5.227'| 0.353515625]], [], [['ns_1@10.1.5.229'| 0.353515625]], [], [], [], [], [], []}}}} INFO REPORT <0.6608.0> 2011-05-10 14:11:40 =============================================================================== vbucketmigrator<0.6608.0>: Validate bucket states vbucketmigrator<0.6608.0>: 181 ok INFO REPORT <0.6613.0> 2011-05-10 14:11:41 =============================================================================== vbucketmigrator<0.6613.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.6613.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6613.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6613.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.6613.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6613.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6613.0>: Starting to move bucket 182 vbucketmigrator<0.6613.0>: Bucket 182 moved to the next server INFO REPORT <0.104.0> 2011-05-10 14:11:41 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.474.0> 2011-05-10 14:11:43 =============================================================================== ns_1@10.1.5.227:ns_doctor:82: Current node statuses: [{'ns_1@10.1.5.227', [{last_heard,{1305,61902,411240}}, {active_buckets,["default"]}, {memory, [{total,40023240}, {processes,19254824}, {processes_used,19242184}, {system,20768416}, {atom,819441}, {atom_used,806779}, {binary,86184}, {code,8552890}, {ets,1684896}]}, {cluster_compatibility_version,1}, {version, [{kernel,"2.13.4"}, {mnesia,"4.4.12"}, {stdlib,"1.16.4"}, {os_mon,"2.2.4"}, {inets,"5.2"}, {ns_server,"1.6.5.4r"}, {menelaus,"1.6.5.4r"}, {sasl,"2.1.8"}]}, {system_arch,"x86_64-unknown-linux-gnu"}, {wall_clock,1294}, {memory_data,{2058731520,679358464,{<0.12.0>,6657168}}}, {disk_data, [{"/",11204616,49},{"/boot",101086,14},{"/dev/shm",1005240,0}]}, {meminfo, <<"MemTotal: 2010480 kB\nMemFree: 1346176 kB\nBuffers: 66200 kB\nCached: 343288 kB\nSwapCached: 20084 kB\nActive: 350868 kB\nInactive: 200400 kB\nHighTotal: 0 kB\nHighFree: 0 kB\nLowTotal: 2010480 kB\nLowFree: 1346176 kB\nSwapTotal: 4030456 kB\nSwapFree: 3952288 kB\nDirty: 340 kB\nWriteback: 0 kB\nAnonPages: 139580 kB\nMapped: 15740 kB\nSlab: 86748 kB\nPageTables: 6872 kB\nNFS_Unstable: 0 kB\nBounce: 0 kB\nCommitLimit: 5035696 kB\nCommitted_AS: 554044 kB\nVmallocTotal: 34359738367 kB\nVmallocUsed: 265484 kB\nVmallocChunk: 34359471011 kB\nHugePages_Total: 0\nHugePages_Free: 0\nHugePages_Rsvd: 0\nHugepagesize: 2048 kB\n">>}, {replication,[{"default",1.0}]}, {system_memory_data, [{system_total_memory,2058731520}, {free_swap,4047142912}, {total_swap,4127186944}, {cached_memory,351526912}, {buffered_memory,67788800}, {free_memory,1378484224}, {total_memory,2058731520}]}, {statistics, [{wall_clock,{1290579,1}}, {context_switches,{399149,0}}, {garbage_collection,{83040,430876220,0}}, {io,{{input,88640252},{output,18905742}}}, {reductions,{121646552,804342}}, {run_queue,0}, {runtime,{17680,140}}]}]}, {'ns_1@10.1.5.229', [{last_heard,{1305,61902,242169}}, {active_buckets,["default"]}, {memory, [{total,33013912}, {processes,13310784}, {processes_used,13287264}, {system,19703128}, {atom,778121}, {atom_used,752794}, {binary,231032}, {code,7791241}, {ets,1612992}]}, {cluster_compatibility_version,1}, {version, [{kernel,"2.13.4"}, {mnesia,"4.4.12"}, {stdlib,"1.16.4"}, {os_mon,"2.2.4"}, {ns_server,"1.6.5.4r"}, {menelaus,"1.6.5.4r"}, {sasl,"2.1.8"}]}, {system_arch,"x86_64-unknown-linux-gnu"}, {wall_clock,1257}, {memory_data,{2058731520,723873792,{<9352.12.0>,4114680}}}, {disk_data, [{"/",11204616,44},{"/boot",101086,14},{"/dev/shm",1005240,0}]}, {meminfo, <<"MemTotal: 2010480 kB\nMemFree: 1302828 kB\nBuffers: 88024 kB\nCached: 341640 kB\nSwapCached: 0 kB\nActive: 369988 kB\nInactive: 220420 kB\nHighTotal: 0 kB\nHighFree: 0 kB\nLowTotal: 2010480 kB\nLowFree: 1302828 kB\nSwapTotal: 4030456 kB\nSwapFree: 4030388 kB\nDirty: 188 kB\nWriteback: 0 kB\nAnonPages: 160744 kB\nMapped: 28104 kB\nSlab: 91988 kB\nPageTables: 5872 kB\nNFS_Unstable: 0 kB\nBounce: 0 kB\nCommitLimit: 5035696 kB\nCommitted_AS: 481900 kB\nVmallocTotal: 34359738367 kB\nVmallocUsed: 265484 kB\nVmallocChunk: 34359471011 kB\nHugePages_Total: 0\nHugePages_Free: 0\nHugePages_Rsvd: 0\nHugepagesize: 2048 kB\n">>}, {replication,[{"default",1.0}]}, {system_memory_data, [{system_total_memory,2058731520}, {free_swap,4127117312}, {total_swap,4127186944}, {cached_memory,349839360}, {buffered_memory,90136576}, {free_memory,1334095872}, {total_memory,2058731520}]}, {statistics, [{wall_clock,{1249806,1}}, {context_switches,{109612,0}}, {garbage_collection,{24122,93429891,0}}, {io,{{input,12063562},{output,8145335}}}, {reductions,{30854536,233337}}, {run_queue,0}, {runtime,{5130,30}}]}]}] INFO REPORT <0.6613.0> 2011-05-10 14:11:46 =============================================================================== vbucketmigrator<0.6613.0>: Validate bucket states vbucketmigrator<0.6613.0>: 182 ok INFO REPORT <0.6630.0> 2011-05-10 14:11:47 =============================================================================== vbucketmigrator<0.6630.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.6630.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6630.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6630.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.6630.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6630.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6630.0>: Starting to move bucket 183 vbucketmigrator<0.6630.0>: Bucket 183 moved to the next server INFO REPORT <0.104.0> 2011-05-10 14:11:47 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.109.0> 2011-05-10 14:11:48 =============================================================================== ns_1@10.1.5.227:ns_orchestrator:164: Skipping janitor in state rebalancing: {rebalancing_state, <0.577.0>, {dict, 2, 16, 16, 8, 80, 48, {[], [], [], [], [], [], [], [], [], [], [], [], [], [], [], []}, {{[], [], [], [], [], [], [], [['ns_1@10.1.5.227'| 0.357421875]], [], [['ns_1@10.1.5.229'| 0.357421875]], [], [], [], [], [], []}}}} INFO REPORT <0.6630.0> 2011-05-10 14:11:52 =============================================================================== vbucketmigrator<0.6630.0>: Validate bucket states vbucketmigrator<0.6630.0>: 183 ok INFO REPORT <0.6643.0> 2011-05-10 14:11:53 =============================================================================== vbucketmigrator<0.6643.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.6643.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6643.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6643.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.6643.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6643.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6643.0>: Starting to move bucket 184 vbucketmigrator<0.6643.0>: Bucket 184 moved to the next server INFO REPORT <0.104.0> 2011-05-10 14:11:53 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.109.0> 2011-05-10 14:11:58 =============================================================================== ns_1@10.1.5.227:ns_orchestrator:164: Skipping janitor in state rebalancing: {rebalancing_state, <0.577.0>, {dict, 2, 16, 16, 8, 80, 48, {[], [], [], [], [], [], [], [], [], [], [], [], [], [], [], []}, {{[], [], [], [], [], [], [], [['ns_1@10.1.5.227'| 0.359375]], [], [['ns_1@10.1.5.229'| 0.359375]], [], [], [], [], [], []}}}} INFO REPORT <0.6643.0> 2011-05-10 14:11:58 =============================================================================== vbucketmigrator<0.6643.0>: Validate bucket states vbucketmigrator<0.6643.0>: 184 ok INFO REPORT <0.6650.0> 2011-05-10 14:11:59 =============================================================================== vbucketmigrator<0.6650.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.6650.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6650.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6650.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.6650.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6650.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6650.0>: Starting to move bucket 185 vbucketmigrator<0.6650.0>: Bucket 185 moved to the next server INFO REPORT <0.104.0> 2011-05-10 14:11:59 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.469.0> 2011-05-10 14:12:04 =============================================================================== ns_1@10.1.5.227:stats_collector:71: Stats for bucket "default": auth_cmds 218 auth_errors 0 bucket_conns 34 bytes_read 1648810 bytes_written 15347784 cas_badval 0 cas_hits 0 cas_misses 0 cmd_flush 1 cmd_get 0 cmd_set 16004 conn_yields 0 connection_structures 45 curr_connections 44 curr_items 6571 curr_items_tot 8040 daemon_connections 10 decr_hits 0 decr_misses 0 delete_hits 0 delete_misses 0 ep_bg_fetched 0 ep_commit_num 53 ep_commit_time 0 ep_commit_time_total 4 ep_data_age 13 ep_data_age_highwat 13 ep_db_cleaner_status complete ep_db_strategy multiMTVBDB ep_dbinit 1 ep_dbname /var/opt/membase/1.6.5.4r/data/ns_1/default ep_dbshards 4 ep_expired 0 ep_flush_duration 1 ep_flush_duration_highwat 8 ep_flush_duration_total 12 ep_flush_preempts 0 ep_flusher_state running ep_flusher_todo 0 ep_io_num_read 1469 ep_io_num_write 15741 ep_io_read_bytes 83404 ep_io_write_bytes 1011432 ep_item_begin_failed 0 ep_item_commit_failed 0 ep_item_flush_expired 0 ep_item_flush_failed 0 ep_kv_size 1181400 ep_max_data_size 1096810496 ep_max_txn_size 1000 ep_mem_high_wat 822607872 ep_mem_low_wat 658086297 ep_min_data_age 0 ep_num_active_non_resident 0 ep_num_eject_failures 0 ep_num_eject_replicas 0 ep_num_expiry_pager_runs 0 ep_num_non_resident 0 ep_num_not_my_vbuckets 0 ep_num_pager_runs 0 ep_num_value_ejects 0 ep_oom_errors 0 ep_overhead 25854448 ep_pending_ops 0 ep_pending_ops_max 0 ep_pending_ops_max_duration 0 ep_pending_ops_total 0 ep_queue_age_cap 900 ep_queue_size 0 ep_storage_age 0 ep_storage_age_highwat 10 ep_storage_type featured ep_store_max_concurrency 10 ep_store_max_readers 9 ep_store_max_readwrite 1 ep_tap_bg_fetch_requeued 0 ep_tap_bg_fetched 1469 ep_tap_keepalive 300 ep_tmp_oom_errors 0 ep_too_old 0 ep_too_young 0 ep_total_cache_size 1181400 ep_total_del_items 0 ep_total_enqueued 16035 ep_total_new_items 15741 ep_total_persisted 15741 ep_vbucket_del 0 ep_vbucket_del_fail 0 ep_version 1.6.5.3.1_13_g81a6cf1 ep_warmed_up 0 ep_warmup true ep_warmup_dups 0 ep_warmup_oom 0 ep_warmup_thread complete ep_warmup_time 15662 get_hits 0 get_misses 0 incr_hits 0 incr_misses 0 libevent 1.4.13-stable limit_maxbytes 67108864 mem_used 27035848 pid 16183 pointer_size 64 rejected_conns 0 rusage_system 1.501771 rusage_user 4.389332 tap_connect_received 186 tap_mutation_sent 2938 tap_opaque_sent 186 tap_vbucket_set_sent 372 threads 4 time 1305061923 total_connections 252 uptime 1317 version 1.4.4_382_g9df3289 INFO REPORT <0.6650.0> 2011-05-10 14:12:04 =============================================================================== vbucketmigrator<0.6650.0>: Validate bucket states vbucketmigrator<0.6650.0>: 185 ok INFO REPORT <0.6654.0> 2011-05-10 14:12:05 =============================================================================== vbucketmigrator<0.6654.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.6654.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6654.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6654.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.6654.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6654.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6654.0>: Starting to move bucket 186 vbucketmigrator<0.6654.0>: Bucket 186 moved to the next server INFO REPORT <0.104.0> 2011-05-10 14:12:05 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.109.0> 2011-05-10 14:12:08 =============================================================================== ns_1@10.1.5.227:ns_orchestrator:164: Skipping janitor in state rebalancing: {rebalancing_state, <0.577.0>, {dict, 2, 16, 16, 8, 80, 48, {[], [], [], [], [], [], [], [], [], [], [], [], [], [], [], []}, {{[], [], [], [], [], [], [], [['ns_1@10.1.5.227'| 0.36328125]], [], [['ns_1@10.1.5.229'| 0.36328125]], [], [], [], [], [], []}}}} INFO REPORT <0.85.0> 2011-05-10 14:12:08 =============================================================================== Pulling config from: 'ns_1@10.1.5.229' INFO REPORT <0.6654.0> 2011-05-10 14:12:10 =============================================================================== vbucketmigrator<0.6654.0>: Validate bucket states vbucketmigrator<0.6654.0>: 186 ok INFO REPORT <0.6664.0> 2011-05-10 14:12:11 =============================================================================== vbucketmigrator<0.6664.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.6664.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6664.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6664.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.6664.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6664.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6664.0>: Starting to move bucket 187 vbucketmigrator<0.6664.0>: Bucket 187 moved to the next server INFO REPORT <0.104.0> 2011-05-10 14:12:11 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.6664.0> 2011-05-10 14:12:16 =============================================================================== vbucketmigrator<0.6664.0>: Validate bucket states vbucketmigrator<0.6664.0>: 187 ok INFO REPORT <0.6672.0> 2011-05-10 14:12:17 =============================================================================== vbucketmigrator<0.6672.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.6672.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6672.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6672.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.6672.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6672.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6672.0>: Starting to move bucket 188 vbucketmigrator<0.6672.0>: Bucket 188 moved to the next server INFO REPORT <0.104.0> 2011-05-10 14:12:17 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.109.0> 2011-05-10 14:12:18 =============================================================================== ns_1@10.1.5.227:ns_orchestrator:164: Skipping janitor in state rebalancing: {rebalancing_state, <0.577.0>, {dict, 2, 16, 16, 8, 80, 48, {[], [], [], [], [], [], [], [], [], [], [], [], [], [], [], []}, {{[], [], [], [], [], [], [], [['ns_1@10.1.5.227'| 0.3671875]], [], [['ns_1@10.1.5.229'| 0.3671875]], [], [], [], [], [], []}}}} INFO REPORT <0.6672.0> 2011-05-10 14:12:22 =============================================================================== vbucketmigrator<0.6672.0>: Validate bucket states vbucketmigrator<0.6672.0>: 188 ok INFO REPORT <0.6677.0> 2011-05-10 14:12:23 =============================================================================== vbucketmigrator<0.6677.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.6677.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6677.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6677.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.6677.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6677.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6677.0>: Starting to move bucket 189 vbucketmigrator<0.6677.0>: Bucket 189 moved to the next server INFO REPORT <0.104.0> 2011-05-10 14:12:23 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.85.0> 2011-05-10 14:12:26 =============================================================================== Pulling config from: 'ns_1@10.1.5.229' INFO REPORT <0.109.0> 2011-05-10 14:12:28 =============================================================================== ns_1@10.1.5.227:ns_orchestrator:164: Skipping janitor in state rebalancing: {rebalancing_state, <0.577.0>, {dict, 2, 16, 16, 8, 80, 48, {[], [], [], [], [], [], [], [], [], [], [], [], [], [], [], []}, {{[], [], [], [], [], [], [], [['ns_1@10.1.5.227'| 0.369140625]], [], [['ns_1@10.1.5.229'| 0.369140625]], [], [], [], [], [], []}}}} INFO REPORT <0.6677.0> 2011-05-10 14:12:28 =============================================================================== vbucketmigrator<0.6677.0>: Validate bucket states vbucketmigrator<0.6677.0>: 189 ok INFO REPORT <0.6685.0> 2011-05-10 14:12:29 =============================================================================== vbucketmigrator<0.6685.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.6685.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6685.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6685.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.6685.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6685.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6685.0>: Starting to move bucket 190 vbucketmigrator<0.6685.0>: Bucket 190 moved to the next server INFO REPORT <0.104.0> 2011-05-10 14:12:29 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.6685.0> 2011-05-10 14:12:34 =============================================================================== vbucketmigrator<0.6685.0>: Validate bucket states vbucketmigrator<0.6685.0>: 190 ok INFO REPORT <0.6697.0> 2011-05-10 14:12:35 =============================================================================== vbucketmigrator<0.6697.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.6697.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6697.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6697.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.6697.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6697.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6697.0>: Starting to move bucket 191 vbucketmigrator<0.6697.0>: Bucket 191 moved to the next server INFO REPORT <0.104.0> 2011-05-10 14:12:35 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.109.0> 2011-05-10 14:12:38 =============================================================================== ns_1@10.1.5.227:ns_orchestrator:164: Skipping janitor in state rebalancing: {rebalancing_state, <0.577.0>, {dict, 2, 16, 16, 8, 80, 48, {[], [], [], [], [], [], [], [], [], [], [], [], [], [], [], []}, {{[], [], [], [], [], [], [], [['ns_1@10.1.5.227'| 0.373046875]], [], [['ns_1@10.1.5.229'| 0.373046875]], [], [], [], [], [], []}}}} INFO REPORT <0.6697.0> 2011-05-10 14:12:40 =============================================================================== vbucketmigrator<0.6697.0>: Validate bucket states vbucketmigrator<0.6697.0>: 191 ok INFO REPORT <0.6702.0> 2011-05-10 14:12:41 =============================================================================== vbucketmigrator<0.6702.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.6702.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6702.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6702.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.6702.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6702.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6702.0>: Starting to move bucket 192 vbucketmigrator<0.6702.0>: Bucket 192 moved to the next server INFO REPORT <0.104.0> 2011-05-10 14:12:41 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.474.0> 2011-05-10 14:12:43 =============================================================================== ns_1@10.1.5.227:ns_doctor:82: Current node statuses: [{'ns_1@10.1.5.227', [{last_heard,{1305,61962,411474}}, {active_buckets,["default"]}, {memory, [{total,40301904}, {processes,19460392}, {processes_used,19448616}, {system,20841512}, {atom,819441}, {atom_used,806779}, {binary,90528}, {code,8552890}, {ets,1751272}]}, {cluster_compatibility_version,1}, {version, [{kernel,"2.13.4"}, {mnesia,"4.4.12"}, {stdlib,"1.16.4"}, {os_mon,"2.2.4"}, {inets,"5.2"}, {ns_server,"1.6.5.4r"}, {menelaus,"1.6.5.4r"}, {sasl,"2.1.8"}]}, {system_arch,"x86_64-unknown-linux-gnu"}, {wall_clock,1354}, {memory_data,{2058731520,679358464,{<0.12.0>,6657168}}}, {disk_data, [{"/",11204616,49},{"/boot",101086,14},{"/dev/shm",1005240,0}]}, {meminfo, <<"MemTotal: 2010480 kB\nMemFree: 1347044 kB\nBuffers: 66520 kB\nCached: 342340 kB\nSwapCached: 20084 kB\nActive: 350588 kB\nInactive: 199916 kB\nHighTotal: 0 kB\nHighFree: 0 kB\nLowTotal: 2010480 kB\nLowFree: 1347044 kB\nSwapTotal: 4030456 kB\nSwapFree: 3952288 kB\nDirty: 144 kB\nWriteback: 0 kB\nAnonPages: 139444 kB\nMapped: 15740 kB\nSlab: 86744 kB\nPageTables: 6872 kB\nNFS_Unstable: 0 kB\nBounce: 0 kB\nCommitLimit: 5035696 kB\nCommitted_AS: 553456 kB\nVmallocTotal: 34359738367 kB\nVmallocUsed: 265484 kB\nVmallocChunk: 34359471011 kB\nHugePages_Total: 0\nHugePages_Free: 0\nHugePages_Rsvd: 0\nHugepagesize: 2048 kB\n">>}, {replication,[{"default",1.0}]}, {system_memory_data, [{system_total_memory,2058731520}, {free_swap,4047142912}, {total_swap,4127186944}, {cached_memory,350556160}, {buffered_memory,68116480}, {free_memory,1379373056}, {total_memory,2058731520}]}, {statistics, [{wall_clock,{1350579,1}}, {context_switches,{409850,0}}, {garbage_collection,{85727,450732659,0}}, {io,{{input,89534578},{output,20027708}}}, {reductions,{126260795,799482}}, {run_queue,0}, {runtime,{18460,130}}]}]}, {'ns_1@10.1.5.229', [{last_heard,{1305,61962,242063}}, {active_buckets,["default"]}, {memory, [{total,33043072}, {processes,13281056}, {processes_used,13257536}, {system,19762016}, {atom,778121}, {atom_used,752794}, {binary,222840}, {code,7791241}, {ets,1679064}]}, {cluster_compatibility_version,1}, {version, [{kernel,"2.13.4"}, {mnesia,"4.4.12"}, {stdlib,"1.16.4"}, {os_mon,"2.2.4"}, {ns_server,"1.6.5.4r"}, {menelaus,"1.6.5.4r"}, {sasl,"2.1.8"}]}, {system_arch,"x86_64-unknown-linux-gnu"}, {wall_clock,1317}, {memory_data,{2058731520,724635648,{<9352.12.0>,4114680}}}, {disk_data, [{"/",11204616,44},{"/boot",101086,14},{"/dev/shm",1005240,0}]}, {meminfo, <<"MemTotal: 2010480 kB\nMemFree: 1302704 kB\nBuffers: 88732 kB\nCached: 340680 kB\nSwapCached: 0 kB\nActive: 370344 kB\nInactive: 220164 kB\nHighTotal: 0 kB\nHighFree: 0 kB\nLowTotal: 2010480 kB\nLowFree: 1302704 kB\nSwapTotal: 4030456 kB\nSwapFree: 4030388 kB\nDirty: 116 kB\nWriteback: 0 kB\nAnonPages: 161100 kB\nMapped: 28104 kB\nSlab: 91996 kB\nPageTables: 5868 kB\nNFS_Unstable: 0 kB\nBounce: 0 kB\nCommitLimit: 5035696 kB\nCommitted_AS: 481900 kB\nVmallocTotal: 34359738367 kB\nVmallocUsed: 265484 kB\nVmallocChunk: 34359471011 kB\nHugePages_Total: 0\nHugePages_Free: 0\nHugePages_Rsvd: 0\nHugepagesize: 2048 kB\n">>}, {replication,[{"default",1.0}]}, {system_memory_data, [{system_total_memory,2058731520}, {free_swap,4127117312}, {total_swap,4127186944}, {cached_memory,348856320}, {buffered_memory,90861568}, {free_memory,1333968896}, {total_memory,2058731520}]}, {statistics, [{wall_clock,{1309806,1}}, {context_switches,{114748,0}}, {garbage_collection,{25281,98152329,0}}, {io,{{input,13020630},{output,9083169}}}, {reductions,{32314594,232521}}, {run_queue,0}, {runtime,{5390,40}}]}]}] INFO REPORT <0.6702.0> 2011-05-10 14:12:46 =============================================================================== vbucketmigrator<0.6702.0>: Validate bucket states vbucketmigrator<0.6702.0>: 192 ok INFO REPORT <0.6706.0> 2011-05-10 14:12:47 =============================================================================== vbucketmigrator<0.6706.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.6706.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6706.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6706.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.6706.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6706.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6706.0>: Starting to move bucket 193 vbucketmigrator<0.6706.0>: Bucket 193 moved to the next server INFO REPORT <0.104.0> 2011-05-10 14:12:47 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.109.0> 2011-05-10 14:12:48 =============================================================================== ns_1@10.1.5.227:ns_orchestrator:164: Skipping janitor in state rebalancing: {rebalancing_state, <0.577.0>, {dict, 2, 16, 16, 8, 80, 48, {[], [], [], [], [], [], [], [], [], [], [], [], [], [], [], []}, {{[], [], [], [], [], [], [], [['ns_1@10.1.5.227'| 0.376953125]], [], [['ns_1@10.1.5.229'| 0.376953125]], [], [], [], [], [], []}}}} INFO REPORT <0.6706.0> 2011-05-10 14:12:52 =============================================================================== vbucketmigrator<0.6706.0>: Validate bucket states vbucketmigrator<0.6706.0>: 193 ok INFO REPORT <0.6719.0> 2011-05-10 14:12:53 =============================================================================== vbucketmigrator<0.6719.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.6719.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6719.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6719.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.6719.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6719.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6719.0>: Starting to move bucket 194 vbucketmigrator<0.6719.0>: Bucket 194 moved to the next server INFO REPORT <0.104.0> 2011-05-10 14:12:53 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.109.0> 2011-05-10 14:12:58 =============================================================================== ns_1@10.1.5.227:ns_orchestrator:164: Skipping janitor in state rebalancing: {rebalancing_state, <0.577.0>, {dict, 2, 16, 16, 8, 80, 48, {[], [], [], [], [], [], [], [], [], [], [], [], [], [], [], []}, {{[], [], [], [], [], [], [], [['ns_1@10.1.5.227'| 0.37890625]], [], [['ns_1@10.1.5.229'| 0.37890625]], [], [], [], [], [], []}}}} INFO REPORT <0.6719.0> 2011-05-10 14:12:58 =============================================================================== vbucketmigrator<0.6719.0>: Validate bucket states vbucketmigrator<0.6719.0>: 194 ok INFO REPORT <0.6726.0> 2011-05-10 14:12:59 =============================================================================== vbucketmigrator<0.6726.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.6726.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6726.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6726.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.6726.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6726.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6726.0>: Starting to move bucket 195 vbucketmigrator<0.6726.0>: Bucket 195 moved to the next server INFO REPORT <0.104.0> 2011-05-10 14:12:59 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.6726.0> 2011-05-10 14:13:04 =============================================================================== vbucketmigrator<0.6726.0>: Validate bucket states vbucketmigrator<0.6726.0>: 195 ok INFO REPORT <0.6730.0> 2011-05-10 14:13:05 =============================================================================== vbucketmigrator<0.6730.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.6730.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6730.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6730.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.6730.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6730.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6730.0>: Starting to move bucket 196 vbucketmigrator<0.6730.0>: Bucket 196 moved to the next server INFO REPORT <0.104.0> 2011-05-10 14:13:05 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.109.0> 2011-05-10 14:13:08 =============================================================================== ns_1@10.1.5.227:ns_orchestrator:164: Skipping janitor in state rebalancing: {rebalancing_state, <0.577.0>, {dict, 2, 16, 16, 8, 80, 48, {[], [], [], [], [], [], [], [], [], [], [], [], [], [], [], []}, {{[], [], [], [], [], [], [], [['ns_1@10.1.5.227'| 0.3828125]], [], [['ns_1@10.1.5.229'| 0.3828125]], [], [], [], [], [], []}}}} INFO REPORT <0.6730.0> 2011-05-10 14:13:10 =============================================================================== vbucketmigrator<0.6730.0>: Validate bucket states vbucketmigrator<0.6730.0>: 196 ok INFO REPORT <0.6740.0> 2011-05-10 14:13:11 =============================================================================== vbucketmigrator<0.6740.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.6740.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6740.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6740.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.6740.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6740.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6740.0>: Starting to move bucket 197 vbucketmigrator<0.6740.0>: Bucket 197 moved to the next server INFO REPORT <0.104.0> 2011-05-10 14:13:11 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.6740.0> 2011-05-10 14:13:16 =============================================================================== vbucketmigrator<0.6740.0>: Validate bucket states vbucketmigrator<0.6740.0>: 197 ok INFO REPORT <0.6748.0> 2011-05-10 14:13:17 =============================================================================== vbucketmigrator<0.6748.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.6748.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6748.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6748.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.6748.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6748.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6748.0>: Starting to move bucket 198 vbucketmigrator<0.6748.0>: Bucket 198 moved to the next server INFO REPORT <0.104.0> 2011-05-10 14:13:17 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.109.0> 2011-05-10 14:13:18 =============================================================================== ns_1@10.1.5.227:ns_orchestrator:164: Skipping janitor in state rebalancing: {rebalancing_state, <0.577.0>, {dict, 2, 16, 16, 8, 80, 48, {[], [], [], [], [], [], [], [], [], [], [], [], [], [], [], []}, {{[], [], [], [], [], [], [], [['ns_1@10.1.5.227'| 0.38671875]], [], [['ns_1@10.1.5.229'| 0.38671875]], [], [], [], [], [], []}}}} INFO REPORT <0.6748.0> 2011-05-10 14:13:22 =============================================================================== vbucketmigrator<0.6748.0>: Validate bucket states vbucketmigrator<0.6748.0>: 198 ok INFO REPORT <0.85.0> 2011-05-10 14:13:23 =============================================================================== Pulling config from: 'ns_1@10.1.5.229' INFO REPORT <0.6753.0> 2011-05-10 14:13:23 =============================================================================== vbucketmigrator<0.6753.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.6753.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6753.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6753.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.6753.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6753.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6753.0>: Starting to move bucket 199 vbucketmigrator<0.6753.0>: Bucket 199 moved to the next server INFO REPORT <0.104.0> 2011-05-10 14:13:23 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.109.0> 2011-05-10 14:13:28 =============================================================================== ns_1@10.1.5.227:ns_orchestrator:164: Skipping janitor in state rebalancing: {rebalancing_state, <0.577.0>, {dict, 2, 16, 16, 8, 80, 48, {[], [], [], [], [], [], [], [], [], [], [], [], [], [], [], []}, {{[], [], [], [], [], [], [], [['ns_1@10.1.5.227'| 0.388671875]], [], [['ns_1@10.1.5.229'| 0.388671875]], [], [], [], [], [], []}}}} INFO REPORT <0.6753.0> 2011-05-10 14:13:28 =============================================================================== vbucketmigrator<0.6753.0>: Validate bucket states vbucketmigrator<0.6753.0>: 199 ok INFO REPORT <0.6760.0> 2011-05-10 14:13:29 =============================================================================== vbucketmigrator<0.6760.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.6760.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6760.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6760.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.6760.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6760.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6760.0>: Starting to move bucket 200 vbucketmigrator<0.6760.0>: Bucket 200 moved to the next server INFO REPORT <0.104.0> 2011-05-10 14:13:29 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.6760.0> 2011-05-10 14:13:34 =============================================================================== vbucketmigrator<0.6760.0>: Validate bucket states vbucketmigrator<0.6760.0>: 200 ok INFO REPORT <0.6772.0> 2011-05-10 14:13:35 =============================================================================== vbucketmigrator<0.6772.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.6772.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6772.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6772.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.6772.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6772.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6772.0>: Starting to move bucket 201 vbucketmigrator<0.6772.0>: Bucket 201 moved to the next server INFO REPORT <0.104.0> 2011-05-10 14:13:35 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.109.0> 2011-05-10 14:13:38 =============================================================================== ns_1@10.1.5.227:ns_orchestrator:164: Skipping janitor in state rebalancing: {rebalancing_state, <0.577.0>, {dict, 2, 16, 16, 8, 80, 48, {[], [], [], [], [], [], [], [], [], [], [], [], [], [], [], []}, {{[], [], [], [], [], [], [], [['ns_1@10.1.5.227'| 0.392578125]], [], [['ns_1@10.1.5.229'| 0.392578125]], [], [], [], [], [], []}}}} INFO REPORT <0.6772.0> 2011-05-10 14:13:40 =============================================================================== vbucketmigrator<0.6772.0>: Validate bucket states vbucketmigrator<0.6772.0>: 201 ok INFO REPORT <0.6777.0> 2011-05-10 14:13:41 =============================================================================== vbucketmigrator<0.6777.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.6777.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6777.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6777.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.6777.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6777.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6777.0>: Starting to move bucket 202 vbucketmigrator<0.6777.0>: Bucket 202 moved to the next server INFO REPORT <0.104.0> 2011-05-10 14:13:41 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.85.0> 2011-05-10 14:13:42 =============================================================================== Pulling config from: 'ns_1@10.1.5.229' INFO REPORT <0.474.0> 2011-05-10 14:13:43 =============================================================================== ns_1@10.1.5.227:ns_doctor:82: Current node statuses: [{'ns_1@10.1.5.227', [{last_heard,{1305,62022,411284}}, {active_buckets,["default"]}, {memory, [{total,40429976}, {processes,19517544}, {processes_used,19505768}, {system,20912432}, {atom,819441}, {atom_used,806779}, {binary,100680}, {code,8552890}, {ets,1811080}]}, {cluster_compatibility_version,1}, {version, [{kernel,"2.13.4"}, {mnesia,"4.4.12"}, {stdlib,"1.16.4"}, {os_mon,"2.2.4"}, {inets,"5.2"}, {ns_server,"1.6.5.4r"}, {menelaus,"1.6.5.4r"}, {sasl,"2.1.8"}]}, {system_arch,"x86_64-unknown-linux-gnu"}, {wall_clock,1414}, {memory_data,{2058731520,679358464,{<0.12.0>,6657168}}}, {disk_data, [{"/",11204616,49},{"/boot",101086,14},{"/dev/shm",1005240,0}]}, {meminfo, <<"MemTotal: 2010480 kB\nMemFree: 1346300 kB\nBuffers: 66804 kB\nCached: 342588 kB\nSwapCached: 20084 kB\nActive: 350852 kB\nInactive: 200360 kB\nHighTotal: 0 kB\nHighFree: 0 kB\nLowTotal: 2010480 kB\nLowFree: 1346300 kB\nSwapTotal: 4030456 kB\nSwapFree: 3952288 kB\nDirty: 396 kB\nWriteback: 0 kB\nAnonPages: 139600 kB\nMapped: 15740 kB\nSlab: 86740 kB\nPageTables: 6868 kB\nNFS_Unstable: 0 kB\nBounce: 0 kB\nCommitLimit: 5035696 kB\nCommitted_AS: 553456 kB\nVmallocTotal: 34359738367 kB\nVmallocUsed: 265484 kB\nVmallocChunk: 34359471011 kB\nHugePages_Total: 0\nHugePages_Free: 0\nHugePages_Rsvd: 0\nHugepagesize: 2048 kB\n">>}, {replication,[{"default",1.0}]}, {system_memory_data, [{system_total_memory,2058731520}, {free_swap,4047142912}, {total_swap,4127186944}, {cached_memory,350810112}, {buffered_memory,68407296}, {free_memory,1378611200}, {total_memory,2058731520}]}, {statistics, [{wall_clock,{1410579,1}}, {context_switches,{420187,0}}, {garbage_collection,{88382,470249856,0}}, {io,{{input,89663685},{output,20448687}}}, {reductions,{130812094,807717}}, {run_queue,0}, {runtime,{19170,130}}]}]}, {'ns_1@10.1.5.229', [{last_heard,{1305,62022,242817}}, {active_buckets,["default"]}, {memory, [{total,33113064}, {processes,13296856}, {processes_used,13273336}, {system,19816208}, {atom,778121}, {atom_used,752794}, {binary,216592}, {code,7791241}, {ets,1739144}]}, {cluster_compatibility_version,1}, {version, [{kernel,"2.13.4"}, {mnesia,"4.4.12"}, {stdlib,"1.16.4"}, {os_mon,"2.2.4"}, {ns_server,"1.6.5.4r"}, {menelaus,"1.6.5.4r"}, {sasl,"2.1.8"}]}, {system_arch,"x86_64-unknown-linux-gnu"}, {wall_clock,1377}, {memory_data,{2058731520,725143552,{<9352.12.0>,4114680}}}, {disk_data, [{"/",11204616,44},{"/boot",101086,14},{"/dev/shm",1005240,0}]}, {meminfo, <<"MemTotal: 2010480 kB\nMemFree: 1301464 kB\nBuffers: 89264 kB\nCached: 340964 kB\nSwapCached: 0 kB\nActive: 371192 kB\nInactive: 220500 kB\nHighTotal: 0 kB\nHighFree: 0 kB\nLowTotal: 2010480 kB\nLowFree: 1301464 kB\nSwapTotal: 4030456 kB\nSwapFree: 4030388 kB\nDirty: 136 kB\nWriteback: 0 kB\nAnonPages: 161468 kB\nMapped: 28104 kB\nSlab: 91980 kB\nPageTables: 5872 kB\nNFS_Unstable: 0 kB\nBounce: 0 kB\nCommitLimit: 5035696 kB\nCommitted_AS: 482924 kB\nVmallocTotal: 34359738367 kB\nVmallocUsed: 265484 kB\nVmallocChunk: 34359471011 kB\nHugePages_Total: 0\nHugePages_Free: 0\nHugePages_Rsvd: 0\nHugepagesize: 2048 kB\n">>}, {replication,[{"default",1.0}]}, {system_memory_data, [{system_total_memory,2058731520}, {free_swap,4127117312}, {total_swap,4127186944}, {cached_memory,349147136}, {buffered_memory,91406336}, {free_memory,1332699136}, {total_memory,2058731520}]}, {statistics, [{wall_clock,{1369806,1}}, {context_switches,{119525,0}}, {garbage_collection,{26336,102555745,0}}, {io,{{input,13242261},{output,9338221}}}, {reductions,{33675379,234484}}, {run_queue,0}, {runtime,{5660,50}}]}]}] INFO REPORT <0.469.0> 2011-05-10 14:13:44 =============================================================================== ns_1@10.1.5.227:stats_collector:71: Stats for bucket "default": auth_cmds 235 auth_errors 0 bucket_conns 34 bytes_read 1655749 bytes_written 16762731 cas_badval 0 cas_hits 0 cas_misses 0 cmd_flush 1 cmd_get 0 cmd_set 16004 conn_yields 0 connection_structures 45 curr_connections 44 curr_items 6412 curr_items_tot 8040 daemon_connections 10 decr_hits 0 decr_misses 0 delete_hits 0 delete_misses 0 ep_bg_fetched 0 ep_commit_num 53 ep_commit_time 0 ep_commit_time_total 4 ep_data_age 13 ep_data_age_highwat 13 ep_db_cleaner_status complete ep_db_strategy multiMTVBDB ep_dbinit 1 ep_dbname /var/opt/membase/1.6.5.4r/data/ns_1/default ep_dbshards 4 ep_expired 0 ep_flush_duration 1 ep_flush_duration_highwat 8 ep_flush_duration_total 12 ep_flush_preempts 0 ep_flusher_state running ep_flusher_todo 0 ep_io_num_read 1628 ep_io_num_write 15741 ep_io_read_bytes 92388 ep_io_write_bytes 1011432 ep_item_begin_failed 0 ep_item_commit_failed 0 ep_item_flush_expired 0 ep_item_flush_failed 0 ep_kv_size 1181400 ep_max_data_size 1096810496 ep_max_txn_size 1000 ep_mem_high_wat 822607872 ep_mem_low_wat 658086297 ep_min_data_age 0 ep_num_active_non_resident 0 ep_num_eject_failures 0 ep_num_eject_replicas 0 ep_num_expiry_pager_runs 0 ep_num_non_resident 0 ep_num_not_my_vbuckets 0 ep_num_pager_runs 0 ep_num_value_ejects 0 ep_oom_errors 0 ep_overhead 25854448 ep_pending_ops 0 ep_pending_ops_max 0 ep_pending_ops_max_duration 0 ep_pending_ops_total 0 ep_queue_age_cap 900 ep_queue_size 0 ep_storage_age 0 ep_storage_age_highwat 10 ep_storage_type featured ep_store_max_concurrency 10 ep_store_max_readers 9 ep_store_max_readwrite 1 ep_tap_bg_fetch_requeued 0 ep_tap_bg_fetched 1628 ep_tap_keepalive 300 ep_tmp_oom_errors 0 ep_too_old 0 ep_too_young 0 ep_total_cache_size 1181400 ep_total_del_items 0 ep_total_enqueued 16035 ep_total_new_items 15741 ep_total_persisted 15741 ep_vbucket_del 0 ep_vbucket_del_fail 0 ep_version 1.6.5.3.1_13_g81a6cf1 ep_warmed_up 0 ep_warmup true ep_warmup_dups 0 ep_warmup_oom 0 ep_warmup_thread complete ep_warmup_time 15662 get_hits 0 get_misses 0 incr_hits 0 incr_misses 0 libevent 1.4.13-stable limit_maxbytes 67108864 mem_used 27035848 pid 16183 pointer_size 64 rejected_conns 0 rusage_system 1.538766 rusage_user 4.592301 tap_connect_received 203 tap_mutation_sent 3256 tap_opaque_sent 203 tap_vbucket_set_sent 406 threads 4 time 1305062023 total_connections 272 uptime 1417 version 1.4.4_382_g9df3289 INFO REPORT <0.6777.0> 2011-05-10 14:13:46 =============================================================================== vbucketmigrator<0.6777.0>: Validate bucket states vbucketmigrator<0.6777.0>: 202 ok INFO REPORT <0.6781.0> 2011-05-10 14:13:47 =============================================================================== vbucketmigrator<0.6781.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.6781.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6781.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6781.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.6781.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6781.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6781.0>: Starting to move bucket 203 vbucketmigrator<0.6781.0>: Bucket 203 moved to the next server INFO REPORT <0.104.0> 2011-05-10 14:13:47 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.109.0> 2011-05-10 14:13:48 =============================================================================== ns_1@10.1.5.227:ns_orchestrator:164: Skipping janitor in state rebalancing: {rebalancing_state, <0.577.0>, {dict, 2, 16, 16, 8, 80, 48, {[], [], [], [], [], [], [], [], [], [], [], [], [], [], [], []}, {{[], [], [], [], [], [], [], [['ns_1@10.1.5.227'| 0.396484375]], [], [['ns_1@10.1.5.229'| 0.396484375]], [], [], [], [], [], []}}}} INFO REPORT <0.6781.0> 2011-05-10 14:13:52 =============================================================================== vbucketmigrator<0.6781.0>: Validate bucket states vbucketmigrator<0.6781.0>: 203 ok INFO REPORT <0.6794.0> 2011-05-10 14:13:53 =============================================================================== vbucketmigrator<0.6794.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.6794.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6794.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6794.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.6794.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6794.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6794.0>: Starting to move bucket 204 vbucketmigrator<0.6794.0>: Bucket 204 moved to the next server INFO REPORT <0.104.0> 2011-05-10 14:13:53 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.109.0> 2011-05-10 14:13:58 =============================================================================== ns_1@10.1.5.227:ns_orchestrator:164: Skipping janitor in state rebalancing: {rebalancing_state, <0.577.0>, {dict, 2, 16, 16, 8, 80, 48, {[], [], [], [], [], [], [], [], [], [], [], [], [], [], [], []}, {{[], [], [], [], [], [], [], [['ns_1@10.1.5.227'| 0.3984375]], [], [['ns_1@10.1.5.229'| 0.3984375]], [], [], [], [], [], []}}}} INFO REPORT <0.6794.0> 2011-05-10 14:13:58 =============================================================================== vbucketmigrator<0.6794.0>: Validate bucket states vbucketmigrator<0.6794.0>: 204 ok INFO REPORT <0.6801.0> 2011-05-10 14:13:59 =============================================================================== vbucketmigrator<0.6801.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.6801.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6801.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6801.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.6801.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6801.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6801.0>: Starting to move bucket 205 vbucketmigrator<0.6801.0>: Bucket 205 moved to the next server INFO REPORT <0.104.0> 2011-05-10 14:13:59 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.6801.0> 2011-05-10 14:14:04 =============================================================================== vbucketmigrator<0.6801.0>: Validate bucket states vbucketmigrator<0.6801.0>: 205 ok INFO REPORT <0.6805.0> 2011-05-10 14:14:05 =============================================================================== vbucketmigrator<0.6805.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.6805.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6805.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6805.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.6805.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6805.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6805.0>: Starting to move bucket 206 vbucketmigrator<0.6805.0>: Bucket 206 moved to the next server INFO REPORT <0.104.0> 2011-05-10 14:14:05 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.109.0> 2011-05-10 14:14:08 =============================================================================== ns_1@10.1.5.227:ns_orchestrator:164: Skipping janitor in state rebalancing: {rebalancing_state, <0.577.0>, {dict, 2, 16, 16, 8, 80, 48, {[], [], [], [], [], [], [], [], [], [], [], [], [], [], [], []}, {{[], [], [], [], [], [], [], [['ns_1@10.1.5.227'| 0.40234375]], [], [['ns_1@10.1.5.229'| 0.40234375]], [], [], [], [], [], []}}}} INFO REPORT <0.6805.0> 2011-05-10 14:14:10 =============================================================================== vbucketmigrator<0.6805.0>: Validate bucket states vbucketmigrator<0.6805.0>: 206 ok INFO REPORT <0.6816.0> 2011-05-10 14:14:11 =============================================================================== vbucketmigrator<0.6816.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.6816.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6816.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6816.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.6816.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6816.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6816.0>: Starting to move bucket 207 vbucketmigrator<0.6816.0>: Bucket 207 moved to the next server INFO REPORT <0.104.0> 2011-05-10 14:14:11 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.6816.0> 2011-05-10 14:14:16 =============================================================================== vbucketmigrator<0.6816.0>: Validate bucket states vbucketmigrator<0.6816.0>: 207 ok INFO REPORT <0.6824.0> 2011-05-10 14:14:17 =============================================================================== vbucketmigrator<0.6824.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.6824.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6824.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6824.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.6824.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6824.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6824.0>: Starting to move bucket 208 vbucketmigrator<0.6824.0>: Bucket 208 moved to the next server INFO REPORT <0.104.0> 2011-05-10 14:14:17 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.109.0> 2011-05-10 14:14:18 =============================================================================== ns_1@10.1.5.227:ns_orchestrator:164: Skipping janitor in state rebalancing: {rebalancing_state, <0.577.0>, {dict, 2, 16, 16, 8, 80, 48, {[], [], [], [], [], [], [], [], [], [], [], [], [], [], [], []}, {{[], [], [], [], [], [], [], [['ns_1@10.1.5.227'| 0.40625]], [], [['ns_1@10.1.5.229'| 0.40625]], [], [], [], [], [], []}}}} INFO REPORT <0.6824.0> 2011-05-10 14:14:22 =============================================================================== vbucketmigrator<0.6824.0>: Validate bucket states vbucketmigrator<0.6824.0>: 208 ok INFO REPORT <0.6829.0> 2011-05-10 14:14:23 =============================================================================== vbucketmigrator<0.6829.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.6829.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6829.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6829.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.6829.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6829.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6829.0>: Starting to move bucket 209 vbucketmigrator<0.6829.0>: Bucket 209 moved to the next server INFO REPORT <0.104.0> 2011-05-10 14:14:23 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.109.0> 2011-05-10 14:14:28 =============================================================================== ns_1@10.1.5.227:ns_orchestrator:164: Skipping janitor in state rebalancing: {rebalancing_state, <0.577.0>, {dict, 2, 16, 16, 8, 80, 48, {[], [], [], [], [], [], [], [], [], [], [], [], [], [], [], []}, {{[], [], [], [], [], [], [], [['ns_1@10.1.5.227'| 0.408203125]], [], [['ns_1@10.1.5.229'| 0.408203125]], [], [], [], [], [], []}}}} INFO REPORT <0.6829.0> 2011-05-10 14:14:28 =============================================================================== vbucketmigrator<0.6829.0>: Validate bucket states vbucketmigrator<0.6829.0>: 209 ok INFO REPORT <0.6836.0> 2011-05-10 14:14:29 =============================================================================== vbucketmigrator<0.6836.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.6836.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6836.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6836.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.6836.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6836.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6836.0>: Starting to move bucket 210 vbucketmigrator<0.6836.0>: Bucket 210 moved to the next server INFO REPORT <0.104.0> 2011-05-10 14:14:29 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.85.0> 2011-05-10 14:14:29 =============================================================================== Pulling config from: 'ns_1@10.1.5.229' INFO REPORT <0.6836.0> 2011-05-10 14:14:34 =============================================================================== vbucketmigrator<0.6836.0>: Validate bucket states vbucketmigrator<0.6836.0>: 210 ok INFO REPORT <0.6848.0> 2011-05-10 14:14:35 =============================================================================== vbucketmigrator<0.6848.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.6848.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6848.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6848.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.6848.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6848.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6848.0>: Starting to move bucket 211 vbucketmigrator<0.6848.0>: Bucket 211 moved to the next server INFO REPORT <0.104.0> 2011-05-10 14:14:35 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.109.0> 2011-05-10 14:14:38 =============================================================================== ns_1@10.1.5.227:ns_orchestrator:164: Skipping janitor in state rebalancing: {rebalancing_state, <0.577.0>, {dict, 2, 16, 16, 8, 80, 48, {[], [], [], [], [], [], [], [], [], [], [], [], [], [], [], []}, {{[], [], [], [], [], [], [], [['ns_1@10.1.5.227'| 0.412109375]], [], [['ns_1@10.1.5.229'| 0.412109375]], [], [], [], [], [], []}}}} INFO REPORT <0.6848.0> 2011-05-10 14:14:40 =============================================================================== vbucketmigrator<0.6848.0>: Validate bucket states vbucketmigrator<0.6848.0>: 211 ok INFO REPORT <0.6853.0> 2011-05-10 14:14:41 =============================================================================== vbucketmigrator<0.6853.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.6853.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6853.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6853.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.6853.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6853.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6853.0>: Starting to move bucket 212 vbucketmigrator<0.6853.0>: Bucket 212 moved to the next server INFO REPORT <0.104.0> 2011-05-10 14:14:41 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.474.0> 2011-05-10 14:14:43 =============================================================================== ns_1@10.1.5.227:ns_doctor:82: Current node statuses: [{'ns_1@10.1.5.227', [{last_heard,{1305,62082,411290}}, {active_buckets,["default"]}, {memory, [{total,40834256}, {processes,19921936}, {processes_used,19910160}, {system,20912320}, {atom,819441}, {atom_used,806779}, {binary,103992}, {code,8552890}, {ets,1806448}]}, {cluster_compatibility_version,1}, {version, [{kernel,"2.13.4"}, {mnesia,"4.4.12"}, {stdlib,"1.16.4"}, {os_mon,"2.2.4"}, {inets,"5.2"}, {ns_server,"1.6.5.4r"}, {menelaus,"1.6.5.4r"}, {sasl,"2.1.8"}]}, {system_arch,"x86_64-unknown-linux-gnu"}, {wall_clock,1474}, {memory_data,{2058731520,680501248,{<0.12.0>,6657168}}}, {disk_data, [{"/",11204616,49},{"/boot",101086,14},{"/dev/shm",1005240,0}]}, {meminfo, <<"MemTotal: 2010480 kB\nMemFree: 1345804 kB\nBuffers: 67088 kB\nCached: 342868 kB\nSwapCached: 20084 kB\nActive: 350876 kB\nInactive: 200816 kB\nHighTotal: 0 kB\nHighFree: 0 kB\nLowTotal: 2010480 kB\nLowFree: 1345804 kB\nSwapTotal: 4030456 kB\nSwapFree: 3952288 kB\nDirty: 196 kB\nWriteback: 0 kB\nAnonPages: 139528 kB\nMapped: 15740 kB\nSlab: 86748 kB\nPageTables: 6864 kB\nNFS_Unstable: 0 kB\nBounce: 0 kB\nCommitLimit: 5035696 kB\nCommitted_AS: 553456 kB\nVmallocTotal: 34359738367 kB\nVmallocUsed: 265484 kB\nVmallocChunk: 34359471011 kB\nHugePages_Total: 0\nHugePages_Free: 0\nHugePages_Rsvd: 0\nHugepagesize: 2048 kB\n">>}, {replication,[{"default",1.0}]}, {system_memory_data, [{system_total_memory,2058731520}, {free_swap,4047142912}, {total_swap,4127186944}, {cached_memory,351096832}, {buffered_memory,68698112}, {free_memory,1378103296}, {total_memory,2058731520}]}, {statistics, [{wall_clock,{1470579,1}}, {context_switches,{430543,0}}, {garbage_collection,{90844,489788845,0}}, {io,{{input,89808128},{output,20876693}}}, {reductions,{135352143,801257}}, {run_queue,0}, {runtime,{19890,140}}]}]}, {'ns_1@10.1.5.229', [{last_heard,{1305,62082,242130}}, {active_buckets,["default"]}, {memory, [{total,33103000}, {processes,13287088}, {processes_used,13263568}, {system,19815912}, {atom,778121}, {atom_used,752794}, {binary,219456}, {code,7791241}, {ets,1735456}]}, {cluster_compatibility_version,1}, {version, [{kernel,"2.13.4"}, {mnesia,"4.4.12"}, {stdlib,"1.16.4"}, {os_mon,"2.2.4"}, {ns_server,"1.6.5.4r"}, {menelaus,"1.6.5.4r"}, {sasl,"2.1.8"}]}, {system_arch,"x86_64-unknown-linux-gnu"}, {wall_clock,1437}, {memory_data,{2058731520,726413312,{<9352.12.0>,4114680}}}, {disk_data, [{"/",11204616,44},{"/boot",101086,14},{"/dev/shm",1005240,0}]}, {meminfo, <<"MemTotal: 2010480 kB\nMemFree: 1300720 kB\nBuffers: 89828 kB\nCached: 341244 kB\nSwapCached: 0 kB\nActive: 371832 kB\nInactive: 221000 kB\nHighTotal: 0 kB\nHighFree: 0 kB\nLowTotal: 2010480 kB\nLowFree: 1300720 kB\nSwapTotal: 4030456 kB\nSwapFree: 4030388 kB\nDirty: 200 kB\nWriteback: 0 kB\nAnonPages: 161752 kB\nMapped: 28104 kB\nSlab: 91972 kB\nPageTables: 5876 kB\nNFS_Unstable: 0 kB\nBounce: 0 kB\nCommitLimit: 5035696 kB\nCommitted_AS: 486716 kB\nVmallocTotal: 34359738367 kB\nVmallocUsed: 265484 kB\nVmallocChunk: 34359471011 kB\nHugePages_Total: 0\nHugePages_Free: 0\nHugePages_Rsvd: 0\nHugepagesize: 2048 kB\n">>}, {replication,[{"default",1.0}]}, {system_memory_data, [{system_total_memory,2058731520}, {free_swap,4127117312}, {total_swap,4127186944}, {cached_memory,349433856}, {buffered_memory,91983872}, {free_memory,1331937280}, {total_memory,2058731520}]}, {statistics, [{wall_clock,{1429806,1}}, {context_switches,{124338,0}}, {garbage_collection,{27409,107069519,0}}, {io,{{input,13479181},{output,9599815}}}, {reductions,{35057505,233435}}, {run_queue,0}, {runtime,{5850,20}}]}]}] INFO REPORT <0.85.0> 2011-05-10 14:14:46 =============================================================================== Pulling config from: 'ns_1@10.1.5.229' INFO REPORT <0.6853.0> 2011-05-10 14:14:46 =============================================================================== vbucketmigrator<0.6853.0>: Validate bucket states vbucketmigrator<0.6853.0>: 212 ok INFO REPORT <0.6872.0> 2011-05-10 14:14:47 =============================================================================== vbucketmigrator<0.6872.0>: Connecting to {Sock 10.1.5.229:11210} vbucketmigrator<0.6872.0>: Authenticating towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6872.0>: Authenticated towards: {Sock 10.1.5.229:11210} vbucketmigrator<0.6872.0>: Connecting to {Sock 10.1.5.227:11210} vbucketmigrator<0.6872.0>: Authenticating towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6872.0>: Authenticated towards: {Sock 10.1.5.227:11210} vbucketmigrator<0.6872.0>: Starting to move bucket 213 vbucketmigrator<0.6872.0>: Bucket 213 moved to the next server INFO REPORT <0.104.0> 2011-05-10 14:14:47 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.109.0> 2011-05-10 14:14:48 =============================================================================== ns_1@10.1.5.227:ns_orchestrator:164: Skipping janitor in state rebalancing: {rebalancing_state, <0.577.0>, {dict, 2, 16, 16, 8, 80, 48, {[], [], [], [], [], [], [], [], [], [], [], [], [], [], [], []}, {{[], [], [], [], [], [], [], [['ns_1@10.1.5.227'| 0.416015625]], [], [['ns_1@10.1.5.229'| 0.416015625]], [], [], [], [], [], []}}}}