per_node_diag = [{'ns_1@10.1.4.243', [{version, [{kernel,"2.13.4"}, {mnesia,"4.4.12"}, {stdlib,"1.16.4"}, {os_mon,"2.2.4"}, {inets,"5.2"}, {ns_server,"1.6.5.4r"}, {menelaus,"1.6.5.4r"}, {sasl,"2.1.8"}]}, {manifest, ["bucket_engine_1.6.5.3_1_g5ba5371-Linux.i686.tar.gz", "curl-7.21.1-w64_patched.tar.gz", "ep-engine_1.6.5.3.1_8_gce31baa-Linux.i686.tar.gz", "google-perftools-1.6","google-perftools-1.6.tar.gz", "libconflate_1.6.5-Linux.i686.tar.gz", "libevent-2.0.7-rc.tar.gz", "libmemcached-0.41_trond-norbye_mingw32-revno895.tar.gz", "libvbucket_1.6.4-Linux.i686.tar.gz", "membase-cli_1.6.5-Linux.i686.tar.gz", "memcached_1.4.4_382_g9df3289-Linux.i686.tar.gz", "moxi_1.6.5.2-Linux.i686.tar.gz", "ns_server_1.6.5.4r.tar.gz", "vbucketmigrator_1.6.5-Linux.i686.tar.gz", "wallace_1.6.5.4r-4-g8da20bb-Linux.i686"]}, {config, [{{node,'ns_1@127.0.0.1',ns_log}, [{filename, "/var/opt/membase/1.6.5.4r/data/ns_1/ns_log"}]}, {otp, [{'_vclock',[{'ns_1@10.1.4.244',{1,63472033663}}]}, {cookie,elyrdmyctqykhepq}]}, {memory_quota,801}, {{node,'ns_1@10.1.4.243',memcached}, [{'_vclock', [{'ns_1@10.1.4.243',{1,63472033663}}, {'ns_1@127.0.0.1',{1,63472032415}}]}, {dbdir,"/var/opt/membase/1.6.5.4r/data/ns_1"}, {port,11210}, {admin_user,"_admin"}, {admin_pass,"_admin"}, {bucket_engine,"./bin/bucket_engine/bucket_engine.so"}, {engines, [{membase, [{engine,"bin/ep_engine/ep.so"}, {initfile,"/etc/opt/membase/1.6.5.4r/init.sql"}]}, {memcached, [{engine,"bin/memcached/default_engine.so"}]}]}, {verbosity,[]}]}, {{node,'ns_1@10.1.4.243',ns_log}, [{'_vclock',[{'ns_1@10.1.4.243',{1,63472033663}}]}, {filename, "/var/opt/membase/1.6.5.4r/data/ns_1/ns_log"}]}, {{node,'ns_1@10.1.4.244',memcached}, [{'_vclock', [{'ns_1@10.1.4.244',{1,63472033663}}, {'ns_1@127.0.0.1',{1,63472033663}}]}, {dbdir,"/var/opt/membase/1.6.5.4r/data/ns_1"}, {port,11210}, {admin_user,"_admin"}, {admin_pass,"_admin"}, {bucket_engine,"./bin/bucket_engine/bucket_engine.so"}, {engines, [{membase, [{engine,"bin/ep_engine/ep.so"}, {initfile,"/etc/opt/membase/1.6.5.4r/init.sql"}]}, {memcached, [{engine,"bin/memcached/default_engine.so"}]}]}, {verbosity,[]}]}, {buckets, [{'_vclock', [{'ns_1@10.1.4.243',{10,63472034172}}, {'ns_1@127.0.0.1',{3,63472032417}}]}, {configs, [{"default", [{num_replicas,1}, {ram_quota,839909376}, {auth_type,sasl}, {sasl_password,[]}, {type,membase}, {num_vbuckets,1024}, {db_shards,4}, {ht_size,3079}, {tap_keepalive,0}, {tap_noop_interval,20}, {max_txn_size,1000}, {ht_locks,5}, {servers,['ns_1@10.1.4.243','ns_1@10.1.4.244']}, {map, [['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined]]}]}]}]}, {{node,'ns_1@127.0.0.1',membership},active}, {rest_creds, [{'_vclock',[{'ns_1@127.0.0.1',{1,63472032422}}]}, {creds, [{"Administrator",[{password,'filtered-out'}]}]}]}, {port_servers, [{moxi,"./bin/moxi/moxi", ["-Z", {"port_listen=~B,default_bucket_name=default,downstream_max=1024,downstream_conn_max=4,connect_max_errors=5,connect_retry_interval=30000,connect_timeout=400,auth_timeout=100,cycle=200,downstream_conn_queue_timeout=200,downstream_timeout=5000,wait_queue_timeout=200", [port]}, "-z", {"url=http://127.0.0.1:~B/pools/default/saslBucketsStreaming", [{rest,port}]}, "-p","0","-Y","y","-O","stderr", {"~s",[verbosity]}], [{env, [{"EVENT_NOSELECT","1"}, {"MOXI_SASL_PLAIN_USR", {"~s",[{ns_moxi_sup,rest_user,[]}]}}, {"MOXI_SASL_PLAIN_PWD", {"~s",[{ns_moxi_sup,rest_pass,[]}]}}]}, use_stdio,stderr_to_stdout,stream]}, {memcached,"./bin/memcached/memcached", ["-X","./bin/memcached/stdin_term_handler.so","-p", {"~B",[port]}, "-E","./bin/bucket_engine/bucket_engine.so","-B", "binary","-r","-c","10000","-e", {"admin=~s;default_bucket_name=default;auto_create=false", [admin_user]}, {"~s",[verbosity]}], [{env, [{"EVENT_NOSELECT","1"}, {"MEMCACHED_TOP_KEYS","100"}, {"ISASL_PWFILE",{"~s",[{isasl,path}]}}, {"ISASL_DB_CHECK_TIME","1"}]}, use_stdio,stderr_to_stdout,stream]}]}, {alerts, [{email,[]}, {email_alerts,false}, {email_server, [{user,undefined}, {pass,'filtered-out'}, {addr,undefined}, {port,undefined}, {encrypt,false}]}, {alerts, [server_down,server_unresponsive,server_up, server_joined,server_left,bucket_created, bucket_deleted,bucket_auth_failed]}]}, {rebalance_status,running}, {{node,'ns_1@10.1.4.244',ns_log}, [{'_vclock',[{'ns_1@10.1.4.244',{1,63472033663}}]}, {filename, "/var/opt/membase/1.6.5.4r/data/ns_1/ns_log"}]}, {nodes_wanted, [{'_vclock',[{'ns_1@10.1.4.243',{6,63472034158}}]}, 'ns_1@10.1.4.243','ns_1@10.1.4.244']}, {rest, [{'_vclock',[{'ns_1@127.0.0.1',{1,63472032422}}]}, {port,8091}]}, {{node,'ns_1@10.1.4.243',membership},active}, {{node,'ns_1@10.1.4.244',membership},active}, {{node,'ns_1@127.0.0.1',isasl}, [{path,"/var/opt/membase/1.6.5.4r/data/ns_1/isasl.pw"}]}, {moxi,[{port,11211},{verbosity,[]}]}, {{node,'ns_1@10.1.4.243',isasl}, [{'_vclock',[{'ns_1@10.1.4.243',{1,63472033663}}]}, {path,"/var/opt/membase/1.6.5.4r/data/ns_1/isasl.pw"}]}, {{node,'ns_1@10.1.4.244',isasl}, [{'_vclock',[{'ns_1@10.1.4.244',{1,63472033663}}]}, {path,"/var/opt/membase/1.6.5.4r/data/ns_1/isasl.pw"}]}, {replication,[{enabled,true}]}, {{node,'ns_1@127.0.0.1',memcached}, [{port,11210}, {dbdir,"/var/opt/membase/1.6.5.4r/data/ns_1"}, {admin_user,"_admin"}, {admin_pass,"_admin"}, {bucket_engine,"./bin/bucket_engine/bucket_engine.so"}, {engines, [{membase, [{engine,"bin/ep_engine/ep.so"}, {initfile,"/etc/opt/membase/1.6.5.4r/init.sql"}]}, {memcached, [{engine,"bin/memcached/default_engine.so"}]}]}, {verbosity,[]}]}]}, {basic_info, [{version, [{kernel,"2.13.4"}, {mnesia,"4.4.12"}, {stdlib,"1.16.4"}, {os_mon,"2.2.4"}, {inets,"5.2"}, {ns_server,"1.6.5.4r"}, {menelaus,"1.6.5.4r"}, {sasl,"2.1.8"}]}, {system_arch,"i686-pc-linux-gnu"}, {wall_clock,321}, {memory_data,{1050726400,789147648,{<0.171.0>,4114276}}}, {disk_data, [{"/",7583436,18}, {"/dev",508460,1}, {"/dev/shm",513048,0}, {"/var/run",513048,1}, {"/var/lock",513048,0}, {"/lib/init/rw",513048,0}, {"/boot",233191,8}]}]}, {processes, [{<0.0.0>, [{registered_name,init}, {status,waiting}, {initial_call,{otp_ring0,start,2}}, {backtrace, ["Program counter: 0xb7322408 (init:loop/1 + 20)", "CP: 0x00000000 (invalid)","arity = 0", "0xb6d28080 Return addr 0x08229d74 ()", "y(0) {state,[{'-root',[<<45 bytes>>]},{'-progname',[<<3 bytes>>]},{'-home',[<<12 bytes"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,66}]}, {heap_size,2584}, {total_heap_size,5168}, {links,[<0.6.0>,<0.7.0>,<0.3.0>]}, {memory,21096}, {message_queue_len,0}, {reductions,26784}, {trap_exit,true}]}, {<0.3.0>, [{registered_name,erl_prim_loader}, {status,waiting}, {initial_call,{erlang,apply,2}}, {backtrace, ["Program counter: 0xb7529328 (erl_prim_loader:loop/3 + 92)", "CP: 0x00000000 (invalid)","arity = 0", "0xb73d4414 Return addr 0x08229d74 ()", "y(0) []", "y(1) [\"/opt/membase/1.6.5.4r/bin/ns_server/ebin\",\"/opt/membase/1.6.5.4r/bin/ns_server/", "y(2) <0.2.0>", "y(3) {state,efile,[],none,#Port<0.2>,infinity,undefined,true,{prim_state,false,undefin", "y(4) infinity"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,344}]}, {heap_size,4181}, {total_heap_size,21892}, {links,[#Port<0.2>,<0.0.0>]}, {memory,87972}, {message_queue_len,0}, {reductions,1458410}, {trap_exit,true}]}, {<0.6.0>, [{registered_name,error_logger}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb753925c (gen_event:fetch_msg/5 + 44)", "CP: 0x00000000 (invalid)","arity = 0", "0xb69c2620 Return addr 0xb7543298 (proc_lib:init_p_do_apply/3 + 28)", "y(0) false","y(1) []", "y(2) [{handler,ns_log_mf_h,false,{state,\"/var/opt/membase/1.6.5.4r/logs\",10485760,10,3", "y(3) error_logger","y(4) <0.2.0>", "0xb69c2638 Return addr 0x08229d74 ()", "y(0) Catch 0xb75432a8 (proc_lib:init_p_do_apply/3 + 44)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,15}]}, {heap_size,2584}, {total_heap_size,5168}, {links,[<0.0.0>,<0.25.0>,#Port<0.1102>]}, {memory,21156}, {message_queue_len,0}, {reductions,219855}, {trap_exit,true}]}, {<0.7.0>, [{registered_name,application_controller}, {status,waiting}, {initial_call,{erlang,apply,2}}, {backtrace, ["Program counter: 0xb757151c (gen_server:loop/6 + 144)", "CP: 0x00000000 (invalid)","arity = 0", "0xb6db0288 Return addr 0x08229d74 ()", "y(0) []","y(1) infinity", "y(2) application_controller", "y(3) {state,[],[],[],[{inets,<0.273.0>},{ns_server,<0.51.0>},{mnesia,<0.117.0>},{os_mo", "y(4) application_controller", "y(5) <0.2.0>"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,5}]}, {heap_size,4181}, {total_heap_size,21892}, {links, [<0.41.0>,<0.117.0>,<0.273.0>,<0.51.0>,<0.9.0>, <0.32.0>,<0.0.0>]}, {memory,88132}, {message_queue_len,0}, {reductions,36094}, {trap_exit,true}]}, {<0.9.0>, [{registered_name,[]}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb75a43b0 (application_master:main_loop/2 + 32)", "CP: 0x00000000 (invalid)","arity = 0", "0xb7390f7c Return addr 0xb7543298 (proc_lib:init_p_do_apply/3 + 28)", "y(0) []", "y(1) {state,<0.10.0>,{appl_data,kernel,[application_controller,erl_reply,auth,boot_ser", "y(2) <0.7.0>", "0xb7390f8c Return addr 0x08229d74 ()", "y(0) Catch 0xb75432a8 (proc_lib:init_p_do_apply/3 + 44)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,1}]}, {heap_size,377}, {total_heap_size,754}, {links,[<0.7.0>,<0.10.0>]}, {memory,3480}, {message_queue_len,0}, {reductions,44}, {trap_exit,true}]}, {<0.10.0>, [{registered_name,[]}, {status,waiting}, {initial_call,{application_master,start_it,4}}, {backtrace, ["Program counter: 0xb75a56f4 (application_master:loop_it/4 + 40)", "CP: 0x00000000 (invalid)","arity = 0", "0xb739b320 Return addr 0x08229d74 ()", "y(0) []","y(1) kernel","y(2) <0.11.0>", "y(3) <0.9.0>"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,0}]}, {heap_size,233}, {total_heap_size,233}, {links,[<0.9.0>,<0.11.0>]}, {memory,1336}, {message_queue_len,0}, {reductions,72}, {trap_exit,true}]}, {<0.11.0>, [{registered_name,kernel_sup}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb757151c (gen_server:loop/6 + 144)", "CP: 0x00000000 (invalid)","arity = 0", "0xb73c2b8c Return addr 0xb7543298 (proc_lib:init_p_do_apply/3 + 28)", "y(0) []","y(1) infinity", "y(2) supervisor", "y(3) {state,{local,kernel_sup},one_for_all,[{child,<0.58.0>,net_sup_dynamic,{erl_distr", "y(4) kernel_sup","y(5) <0.10.0>", "0xb73c2ba8 Return addr 0x08229d74 ()", "y(0) Catch 0xb75432a8 (proc_lib:init_p_do_apply/3 + 44)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,13}]}, {heap_size,610}, {total_heap_size,987}, {links, [<0.22.0>,<0.26.0>,<0.27.0>,<0.58.0>,<0.24.0>, <0.17.0>,<0.20.0>,<0.21.0>,<0.19.0>,<0.12.0>, <0.13.0>,<0.10.0>]}, {memory,4612}, {message_queue_len,0}, {reductions,1492}, {trap_exit,true}]}, {<0.12.0>, [{registered_name,rex}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb757151c (gen_server:loop/6 + 144)", "CP: 0x00000000 (invalid)","arity = 0", "0xb57ba5c0 Return addr 0xb7543298 (proc_lib:init_p_do_apply/3 + 28)", "y(0) []","y(1) infinity","y(2) rpc", "y(3) {1,{<0.5349.0>,{<0.5348.0>,{#Ref<0.0.0.103227>,'ns_1@10.1.4.243'}},nil,nil}}", "y(4) rex","y(5) <0.11.0>", "0xb57ba5dc Return addr 0x08229d74 ()", "y(0) Catch 0xb75432a8 (proc_lib:init_p_do_apply/3 + 44)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,0}]}, {heap_size,317811}, {total_heap_size,317811}, {links,[<0.11.0>]}, {memory,1271768}, {message_queue_len,0}, {reductions,5592}, {trap_exit,true}]}, {<0.13.0>, [{registered_name,global_name_server}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb757151c (gen_server:loop/6 + 144)", "CP: 0x00000000 (invalid)","arity = 0", "0xb6db159c Return addr 0xb7543298 (proc_lib:init_p_do_apply/3 + 28)", "y(0) []","y(1) infinity","y(2) global", "y(3) {state,true,['ns_1@10.1.4.244'],['ns_1@10.1.4.244'],[],[],'ns_1@10.1.4.243',<0.14", "y(4) global_name_server","y(5) <0.11.0>", "0xb6db15b8 Return addr 0x08229d74 ()", "y(0) Catch 0xb75432a8 (proc_lib:init_p_do_apply/3 + 44)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,189}]}, {heap_size,610}, {total_heap_size,987}, {links,[<0.14.0>,<0.16.0>,<0.15.0>,<0.11.0>]}, {memory,4532}, {message_queue_len,0}, {reductions,36599}, {trap_exit,true}]}, {<0.14.0>, [{registered_name,[]}, {status,waiting}, {initial_call,{erlang,apply,2}}, {backtrace, ["Program counter: 0xb75db094 (global:loop_the_locker/1 + 588)", "CP: 0x00000000 (invalid)","arity = 0", "0xb6e1289c Return addr 0xb75dae28 (global:init_the_locker/1 + 192)", "y(0) {multi,[],[],['ns_1@10.1.4.244'],'ns_1@10.1.4.244',false,false}", "y(1) infinity", "0xb6e128a8 Return addr 0x08229d74 ()"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,3}]}, {heap_size,233}, {total_heap_size,610}, {links,[<0.13.0>]}, {memory,2824}, {message_queue_len,0}, {reductions,277}, {trap_exit,true}]}, {<0.15.0>, [{registered_name,[]}, {status,waiting}, {initial_call,{erlang,apply,2}}, {backtrace, ["Program counter: 0xb75df140 (global:collect_deletions/2 + 76)", "CP: 0x00000000 (invalid)","arity = 0", "0xb7391324 Return addr 0xb75df0c0 (global:loop_the_deleter/1 + 36)", "y(0) infinity","y(1) []", "y(2) <0.13.0>", "0xb7391334 Return addr 0x08229d74 ()", "y(0) <0.13.0>"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,0}]}, {heap_size,233}, {total_heap_size,233}, {links,[<0.13.0>]}, {memory,1316}, {message_queue_len,0}, {reductions,4}, {trap_exit,false}]}, {<0.16.0>, [{registered_name,[]}, {status,waiting}, {initial_call,{erlang,apply,2}}, {backtrace, ["Program counter: 0xb75df260 (global:loop_the_registrar/0 + 12)", "CP: 0x00000000 (invalid)","arity = 0", "0xb68715c4 Return addr 0x08229d74 ()", "y(0) []"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,3}]}, {heap_size,377}, {total_heap_size,754}, {links,[<0.13.0>]}, {memory,3400}, {message_queue_len,0}, {reductions,246}, {trap_exit,false}]}, {<0.17.0>, [{registered_name,inet_db}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb757151c (gen_server:loop/6 + 144)", "CP: 0x00000000 (invalid)","arity = 0", "0xb73a6f58 Return addr 0xb7543298 (proc_lib:init_p_do_apply/3 + 28)", "y(0) []","y(1) infinity","y(2) inet_db", "y(3) {state,inet_db,inet_cache,inet_hosts_byname,inet_hosts_byaddr,inet_hosts_file_byn", "y(4) inet_db","y(5) <0.11.0>", "0xb73a6f74 Return addr 0x08229d74 ()", "y(0) Catch 0xb75432a8 (proc_lib:init_p_do_apply/3 + 44)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,2}]}, {heap_size,1597}, {total_heap_size,3194}, {links,[<0.11.0>]}, {memory,13220}, {message_queue_len,0}, {reductions,2005}, {trap_exit,true}]}, {<0.19.0>, [{registered_name,global_group}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb757151c (gen_server:loop/6 + 144)", "CP: 0x00000000 (invalid)","arity = 0", "0xb73a0588 Return addr 0xb7543298 (proc_lib:init_p_do_apply/3 + 28)", "y(0) []","y(1) infinity", "y(2) global_group", "y(3) {state,no_conf,true,[],[],[],[],[],'nonode@nohost',[],normal,normal}", "y(4) global_group","y(5) <0.11.0>", "0xb73a05a4 Return addr 0x08229d74 ()", "y(0) Catch 0xb75432a8 (proc_lib:init_p_do_apply/3 + 44)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,0}]}, {heap_size,233}, {total_heap_size,233}, {links,[<0.11.0>]}, {memory,1376}, {message_queue_len,0}, {reductions,76}, {trap_exit,true}]}, {<0.20.0>, [{registered_name,file_server_2}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb757151c (gen_server:loop/6 + 144)", "CP: 0x00000000 (invalid)","arity = 0", "0xb6e0f68c Return addr 0xb7543298 (proc_lib:init_p_do_apply/3 + 28)", "y(0) []","y(1) infinity", "y(2) file_server","y(3) #Port<0.262>", "y(4) file_server_2","y(5) <0.11.0>", "0xb6e0f6a8 Return addr 0x08229d74 ()", "y(0) Catch 0xb75432a8 (proc_lib:init_p_do_apply/3 + 44)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,744}]}, {heap_size,2584}, {total_heap_size,4181}, {links,[#Port<0.262>,<0.11.0>]}, {memory,17188}, {message_queue_len,0}, {reductions,835312}, {trap_exit,true}]}, {<0.21.0>, [{registered_name,code_server}, {status,waiting}, {initial_call,{erlang,apply,2}}, {backtrace, ["Program counter: 0xb6c3d7a4 (code_server:loop/1 + 64)", "CP: 0x00000000 (invalid)","arity = 0", "0xb73ec9f0 Return addr 0x08229d74 ()", "y(0) {state,<0.11.0>,\"/opt/membase/1.6.5.4r/erlang-13b03/lib/erlang\",[\"/opt/membase/1.", "y(1) <0.11.0>"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,366}]}, {heap_size,6765}, {total_heap_size,24476}, {links,[<0.11.0>]}, {memory,98288}, {message_queue_len,0}, {reductions,315390}, {trap_exit,true}]}, {<0.22.0>, [{registered_name,standard_error_sup}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb757151c (gen_server:loop/6 + 144)", "CP: 0x00000000 (invalid)","arity = 0", "0xb73a7300 Return addr 0xb7543298 (proc_lib:init_p_do_apply/3 + 28)", "y(0) []","y(1) infinity", "y(2) supervisor_bridge", "y(3) {state,standard_error,<0.23.0>,<0.23.0>,{local,standard_error_sup}}", "y(4) standard_error_sup","y(5) <0.11.0>", "0xb73a731c Return addr 0x08229d74 ()", "y(0) Catch 0xb75432a8 (proc_lib:init_p_do_apply/3 + 44)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,0}]}, {heap_size,233}, {total_heap_size,233}, {links,[<0.11.0>,<0.23.0>]}, {memory,1396}, {message_queue_len,0}, {reductions,40}, {trap_exit,true}]}, {<0.23.0>, [{registered_name,standard_error}, {status,waiting}, {initial_call,{standard_error,server,2}}, {backtrace, ["Program counter: 0xb6c2d88c (standard_error:server_loop/1 + 20)", "CP: 0x00000000 (invalid)","arity = 0", "0xb73a76c4 Return addr 0x08229d74 ()", "y(0) #Port<0.568>"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,0}]}, {heap_size,233}, {total_heap_size,233}, {links,[<0.22.0>,#Port<0.568>]}, {memory,1396}, {message_queue_len,0}, {reductions,7}, {trap_exit,true}]}, {<0.24.0>, [{registered_name,[]}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb757151c (gen_server:loop/6 + 144)", "CP: 0x00000000 (invalid)","arity = 0", "0xb73a5660 Return addr 0xb7543298 (proc_lib:init_p_do_apply/3 + 28)", "y(0) []","y(1) infinity", "y(2) supervisor_bridge", "y(3) {state,user_sup,<0.25.0>,<0.25.0>,{<0.24.0>,user_sup}}", "y(4) <0.24.0>","y(5) <0.11.0>", "0xb73a567c Return addr 0x08229d74 ()", "y(0) Catch 0xb75432a8 (proc_lib:init_p_do_apply/3 + 44)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,0}]}, {heap_size,1597}, {total_heap_size,1597}, {links,[<0.11.0>,<0.25.0>]}, {memory,6852}, {message_queue_len,0}, {reductions,71}, {trap_exit,true}]}, {<0.25.0>, [{registered_name,user}, {status,waiting}, {initial_call,{user,server,2}}, {backtrace, ["Program counter: 0xb6c51cfc (user:server_loop/2 + 28)", "CP: 0x00000000 (invalid)","arity = 0", "0xb651c6f0 Return addr 0x08229d74 ()", "y(0) []","y(1) []","y(2) {[],[]}", "y(3) #Port<0.587>"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,12}]}, {heap_size,6765}, {total_heap_size,13530}, {links,[<0.6.0>,<0.24.0>,#Port<0.587>]}, {memory,54604}, {message_queue_len,0}, {reductions,33453}, {trap_exit,true}]}, {<0.26.0>, [{registered_name,[]}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb757151c (gen_server:loop/6 + 144)", "CP: 0x00000000 (invalid)","arity = 0", "0xb73903a0 Return addr 0xb7543298 (proc_lib:init_p_do_apply/3 + 28)", "y(0) []","y(1) infinity", "y(2) kernel_config","y(3) []", "y(4) <0.26.0>","y(5) <0.11.0>", "0xb73903bc Return addr 0x08229d74 ()", "y(0) Catch 0xb75432a8 (proc_lib:init_p_do_apply/3 + 44)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,0}]}, {heap_size,233}, {total_heap_size,233}, {links,[<0.11.0>]}, {memory,1376}, {message_queue_len,0}, {reductions,176}, {trap_exit,true}]}, {<0.27.0>, [{registered_name,kernel_safe_sup}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb757151c (gen_server:loop/6 + 144)", "CP: 0x00000000 (invalid)","arity = 0", "0xb6db5938 Return addr 0xb7543298 (proc_lib:init_p_do_apply/3 + 28)", "y(0) []","y(1) infinity", "y(2) supervisor", "y(3) {state,{local,kernel_safe_sup},one_for_one,[{child,<0.133.0>,disk_log_server,{dis", "y(4) kernel_safe_sup","y(5) <0.11.0>", "0xb6db5954 Return addr 0x08229d74 ()", "y(0) Catch 0xb75432a8 (proc_lib:init_p_do_apply/3 + 44)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,4}]}, {heap_size,377}, {total_heap_size,754}, {links, [<0.127.0>,<0.132.0>,<0.133.0>,<0.128.0>,<0.53.0>, <0.56.0>,<0.11.0>]}, {memory,3580}, {message_queue_len,0}, {reductions,438}, {trap_exit,true}]}, {<0.32.0>, [{registered_name,[]}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb75a43b0 (application_master:main_loop/2 + 32)", "CP: 0x00000000 (invalid)","arity = 0", "0xb6d2841c Return addr 0xb7543298 (proc_lib:init_p_do_apply/3 + 28)", "y(0) []", "y(1) {state,<0.33.0>,{appl_data,sasl,[sasl_sup,alarm_handler,overload,release_handler]", "y(2) <0.7.0>", "0xb6d2842c Return addr 0x08229d74 ()", "y(0) Catch 0xb75432a8 (proc_lib:init_p_do_apply/3 + 44)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,0}]}, {heap_size,233}, {total_heap_size,233}, {links,[<0.7.0>,<0.33.0>]}, {memory,1396}, {message_queue_len,0}, {reductions,31}, {trap_exit,true}]}, {<0.33.0>, [{registered_name,[]}, {status,waiting}, {initial_call,{application_master,start_it,4}}, {backtrace, ["Program counter: 0xb75a56f4 (application_master:loop_it/4 + 40)", "CP: 0x00000000 (invalid)","arity = 0", "0xb739ba70 Return addr 0x08229d74 ()", "y(0) {state,tty,undefined}","y(1) sasl", "y(2) <0.34.0>","y(3) <0.32.0>"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,0}]}, {heap_size,233}, {total_heap_size,233}, {links,[<0.32.0>,<0.34.0>]}, {memory,1336}, {message_queue_len,0}, {reductions,70}, {trap_exit,true}]}, {<0.34.0>, [{registered_name,sasl_sup}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb757151c (gen_server:loop/6 + 144)", "CP: 0x00000000 (invalid)","arity = 0", "0xb739c9d4 Return addr 0xb7543298 (proc_lib:init_p_do_apply/3 + 28)", "y(0) []","y(1) infinity", "y(2) supervisor", "y(3) {state,{local,sasl_sup},one_for_one,[{child,<0.38.0>,release_handler,{release_han", "y(4) sasl_sup","y(5) <0.33.0>", "0xb739c9f0 Return addr 0x08229d74 ()", "y(0) Catch 0xb75432a8 (proc_lib:init_p_do_apply/3 + 44)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,2}]}, {heap_size,610}, {total_heap_size,1220}, {links,[<0.35.0>,<0.38.0>,<0.33.0>]}, {memory,5364}, {message_queue_len,0}, {reductions,284}, {trap_exit,true}]}, {<0.35.0>, [{registered_name,sasl_safe_sup}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb757151c (gen_server:loop/6 + 144)", "CP: 0x00000000 (invalid)","arity = 0", "0xb73997e0 Return addr 0xb7543298 (proc_lib:init_p_do_apply/3 + 28)", "y(0) []","y(1) infinity", "y(2) supervisor", "y(3) {state,{local,sasl_safe_sup},one_for_one,[{child,<0.37.0>,overload,{overload,star", "y(4) sasl_safe_sup","y(5) <0.34.0>", "0xb73997fc Return addr 0x08229d74 ()", "y(0) Catch 0xb75432a8 (proc_lib:init_p_do_apply/3 + 44)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,1}]}, {heap_size,233}, {total_heap_size,610}, {links,[<0.36.0>,<0.37.0>,<0.34.0>]}, {memory,2924}, {message_queue_len,0}, {reductions,165}, {trap_exit,true}]}, {<0.36.0>, [{registered_name,alarm_handler}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb753925c (gen_event:fetch_msg/5 + 44)", "CP: 0x00000000 (invalid)","arity = 0", "0xb73a1c54 Return addr 0xb7543298 (proc_lib:init_p_do_apply/3 + 28)", "y(0) false","y(1) []", "y(2) [{handler,alarm_handler,false,[],false}]", "y(3) alarm_handler","y(4) <0.35.0>", "0xb73a1c6c Return addr 0x08229d74 ()", "y(0) Catch 0xb75432a8 (proc_lib:init_p_do_apply/3 + 44)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,0}]}, {heap_size,233}, {total_heap_size,233}, {links,[<0.35.0>]}, {memory,1376}, {message_queue_len,0}, {reductions,28}, {trap_exit,true}]}, {<0.37.0>, [{registered_name,overload}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb757151c (gen_server:loop/6 + 144)", "CP: 0x00000000 (invalid)","arity = 0", "0xb73a23a0 Return addr 0xb7543298 (proc_lib:init_p_do_apply/3 + 28)", "y(0) []","y(1) infinity", "y(2) overload", "y(3) {state,0,0,8.000000e-01,183,1.000000e-01,{0,0},clear}", "y(4) overload","y(5) <0.35.0>", "0xb73a23bc Return addr 0x08229d74 ()", "y(0) Catch 0xb75432a8 (proc_lib:init_p_do_apply/3 + 44)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,0}]}, {heap_size,233}, {total_heap_size,233}, {links,[<0.35.0>]}, {memory,1376}, {message_queue_len,0}, {reductions,39}, {trap_exit,false}]}, {<0.38.0>, [{registered_name,release_handler}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb757151c (gen_server:loop/6 + 144)", "CP: 0x00000000 (invalid)","arity = 0", "0xb7395e98 Return addr 0xb7543298 (proc_lib:init_p_do_apply/3 + 28)", "y(0) []","y(1) infinity", "y(2) release_handler", "y(3) {state,[],\"/opt/membase/1.6.5.4r/erlang-13b03/lib/erlang\",\"/opt/membase/1.6.5.4r/", "y(4) release_handler","y(5) <0.34.0>", "0xb7395eb4 Return addr 0x08229d74 ()", "y(0) Catch 0xb75432a8 (proc_lib:init_p_do_apply/3 + 44)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,3}]}, {heap_size,1597}, {total_heap_size,4181}, {links,[<0.34.0>]}, {memory,17168}, {message_queue_len,0}, {reductions,2503}, {trap_exit,false}]}, {<0.41.0>, [{registered_name,[]}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb75a43b0 (application_master:main_loop/2 + 32)", "CP: 0x00000000 (invalid)","arity = 0", "0xb73bf184 Return addr 0xb7543298 (proc_lib:init_p_do_apply/3 + 28)", "y(0) []", "y(1) {state,<0.42.0>,{appl_data,os_mon,[os_mon_sup,os_mon_sysinfo,disksup,memsup,cpu_s", "y(2) <0.7.0>", "0xb73bf194 Return addr 0x08229d74 ()", "y(0) Catch 0xb75432a8 (proc_lib:init_p_do_apply/3 + 44)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,0}]}, {heap_size,233}, {total_heap_size,233}, {links,[<0.7.0>,<0.42.0>]}, {memory,1396}, {message_queue_len,0}, {reductions,23}, {trap_exit,true}]}, {<0.42.0>, [{registered_name,[]}, {status,waiting}, {initial_call,{application_master,start_it,4}}, {backtrace, ["Program counter: 0xb75a56f4 (application_master:loop_it/4 + 40)", "CP: 0x00000000 (invalid)","arity = 0", "0xb73bf530 Return addr 0x08229d74 ()", "y(0) []","y(1) os_mon","y(2) <0.43.0>", "y(3) <0.41.0>"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,0}]}, {heap_size,233}, {total_heap_size,233}, {links,[<0.41.0>,<0.43.0>]}, {memory,1336}, {message_queue_len,0}, {reductions,40}, {trap_exit,true}]}, {<0.43.0>, [{registered_name,os_mon_sup}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb757151c (gen_server:loop/6 + 144)", "CP: 0x00000000 (invalid)","arity = 0", "0xb739ab00 Return addr 0xb7543298 (proc_lib:init_p_do_apply/3 + 28)", "y(0) []","y(1) infinity", "y(2) supervisor", "y(3) {state,{local,os_mon_sup},one_for_one,[{child,<0.47.0>,cpu_sup,{cpu_sup,start_lin", "y(4) os_mon_sup","y(5) <0.42.0>", "0xb739ab1c Return addr 0x08229d74 ()", "y(0) Catch 0xb75432a8 (proc_lib:init_p_do_apply/3 + 44)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,1}]}, {heap_size,377}, {total_heap_size,754}, {links,[<0.44.0>,<0.45.0>,<0.47.0>,<0.42.0>]}, {memory,3520}, {message_queue_len,0}, {reductions,273}, {trap_exit,true}]}, {<0.44.0>, [{registered_name,disksup}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb757151c (gen_server:loop/6 + 144)", "CP: 0x00000000 (invalid)","arity = 0", "0xb6dff710 Return addr 0xb7543298 (proc_lib:init_p_do_apply/3 + 28)", "y(0) []","y(1) infinity","y(2) disksup", "y(3) {state,80,60000,{unix,linux},[{\"/\",7583436,18},{\"/dev\",508460,1},{\"/dev/shm\",5130", "y(4) disksup","y(5) <0.43.0>", "0xb6dff72c Return addr 0x08229d74 ()", "y(0) Catch 0xb75432a8 (proc_lib:init_p_do_apply/3 + 44)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,12}]}, {heap_size,4181}, {total_heap_size,8362}, {links,[<0.43.0>,#Port<0.988>]}, {memory,33912}, {message_queue_len,0}, {reductions,14798}, {trap_exit,true}]}, {<0.45.0>, [{registered_name,memsup}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb757151c (gen_server:loop/6 + 144)", "CP: 0x00000000 (invalid)","arity = 0", "0xb6870400 Return addr 0xb7543298 (proc_lib:init_p_do_apply/3 + 28)", "y(0) []","y(1) infinity","y(2) memsup", "y(3) {state,{unix,linux},true,{789147648,1050726400},{<0.171.0>,4114276},false,60000,3", "y(4) memsup","y(5) <0.43.0>", "0xb687041c Return addr 0x08229d74 ()", "y(0) Catch 0xb75432a8 (proc_lib:init_p_do_apply/3 + 44)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,19}]}, {heap_size,1597}, {total_heap_size,2207}, {links,[<0.43.0>,<0.46.0>]}, {memory,9292}, {message_queue_len,0}, {reductions,13760}, {trap_exit,true}]}, {<0.46.0>, [{registered_name,[]}, {status,waiting}, {initial_call,{erlang,apply,2}}, {backtrace, ["Program counter: 0xb6cb365c (memsup:port_idle/1 + 20)", "CP: 0x00000000 (invalid)","arity = 0", "0xb68c58b8 Return addr 0x08229d74 ()", "y(0) []","y(1) #Port<0.1020>"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,10}]}, {heap_size,1597}, {total_heap_size,1974}, {links,[<0.45.0>,#Port<0.1020>]}, {memory,8300}, {message_queue_len,0}, {reductions,3872}, {trap_exit,true}]}, {<0.47.0>, [{registered_name,cpu_sup}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb757151c (gen_server:loop/6 + 144)", "CP: 0x00000000 (invalid)","arity = 0", "0xb739aea8 Return addr 0xb7543298 (proc_lib:init_p_do_apply/3 + 28)", "y(0) []","y(1) infinity","y(2) cpu_sup", "y(3) {state,<0.48.0>,{unix,linux}}", "y(4) cpu_sup","y(5) <0.43.0>", "0xb739aec4 Return addr 0x08229d74 ()", "y(0) Catch 0xb75432a8 (proc_lib:init_p_do_apply/3 + 44)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,0}]}, {heap_size,233}, {total_heap_size,233}, {links,[<0.43.0>]}, {memory,1376}, {message_queue_len,0}, {reductions,33}, {trap_exit,true}]}, {<0.48.0>, [{registered_name,[]}, {status,waiting}, {initial_call,{erlang,apply,2}}, {backtrace, ["Program counter: 0xb6cc5964 (cpu_sup:measurement_server_loop/1 + 20)", "CP: 0x00000000 (invalid)","arity = 0", "0xb73bf8c4 Return addr 0x08229d74 ()", "y(0) []","y(1) []","y(2) []", "y(3) []","y(4) []","y(5) []", "y(6) []","y(7) []", "y(8) {internal,<0.49.0>,[],{unix,linux}}"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,0}]}, {heap_size,233}, {total_heap_size,233}, {links,[<0.49.0>]}, {memory,1316}, {message_queue_len,0}, {reductions,11}, {trap_exit,true}]}, {<0.49.0>, [{registered_name,[]}, {status,waiting}, {initial_call,{erlang,apply,2}}, {backtrace, ["Program counter: 0xb6cc6250 (cpu_sup:port_server_loop/2 + 32)", "CP: 0x00000000 (invalid)","arity = 0", "0xb73ab5e8 Return addr 0x08229d74 ()", "y(0) []","y(1) 6000", "y(2) #Port<0.1052>"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,3}]}, {heap_size,610}, {total_heap_size,987}, {links,[<0.48.0>,#Port<0.1052>]}, {memory,4352}, {message_queue_len,0}, {reductions,335}, {trap_exit,false}]}, {<0.51.0>, [{registered_name,[]}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb75a43b0 (application_master:main_loop/2 + 32)", "CP: 0x00000000 (invalid)","arity = 0", "0xb73a093c Return addr 0xb7543298 (proc_lib:init_p_do_apply/3 + 28)", "y(0) []", "y(1) {state,<0.52.0>,{appl_data,ns_server,[ns_server_sup,ns_config,ns_config_sup,ns_co", "y(2) <0.7.0>", "0xb73a094c Return addr 0x08229d74 ()", "y(0) Catch 0xb75432a8 (proc_lib:init_p_do_apply/3 + 44)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,0}]}, {heap_size,233}, {total_heap_size,233}, {links,[<0.7.0>,<0.52.0>]}, {memory,1396}, {message_queue_len,0}, {reductions,27}, {trap_exit,true}]}, {<0.52.0>, [{registered_name,[]}, {status,waiting}, {initial_call,{application_master,start_it,4}}, {backtrace, ["Program counter: 0xb75a56f4 (application_master:loop_it/4 + 40)", "CP: 0x00000000 (invalid)","arity = 0", "0xb7399f40 Return addr 0x08229d74 ()", "y(0) []","y(1) ns_server", "y(2) <0.54.0>","y(3) <0.51.0>"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,0}]}, {heap_size,233}, {total_heap_size,233}, {links,[<0.51.0>,<0.54.0>]}, {memory,1336}, {message_queue_len,0}, {reductions,50}, {trap_exit,true}]}, {<0.53.0>, [{registered_name,timer_server}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb757151c (gen_server:loop/6 + 144)", "CP: 0x00000000 (invalid)","arity = 0", "0xb6a19720 Return addr 0xb7543298 (proc_lib:init_p_do_apply/3 + 28)", "y(0) []","y(1) 575","y(2) timer", "y(3) []","y(4) timer_server", "y(5) <0.27.0>", "0xb6a1973c Return addr 0x08229d74 ()", "y(0) Catch 0xb75432a8 (proc_lib:init_p_do_apply/3 + 44)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,704}]}, {heap_size,1597}, {total_heap_size,1974}, {links, [<0.100.0>,<0.125.0>,<0.168.0>,<0.177.0>,<0.156.0>, <0.113.0>,<0.112.0>,<0.92.0>,<0.94.0>,<0.73.0>, <0.80.0>,<0.27.0>]}, {memory,8560}, {message_queue_len,0}, {reductions,188135}, {trap_exit,true}]}, {<0.54.0>, [{registered_name,ns_server_cluster_sup}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb757151c (gen_server:loop/6 + 144)", "CP: 0x00000000 (invalid)","arity = 0", "0xb73db98c Return addr 0xb7543298 (proc_lib:init_p_do_apply/3 + 28)", "y(0) []","y(1) infinity", "y(2) supervisor", "y(3) {state,{local,ns_server_cluster_sup},one_for_one,[{child,<0.72.0>,ns_server_sup,{", "y(4) ns_server_cluster_sup","y(5) <0.52.0>", "0xb73db9a8 Return addr 0x08229d74 ()", "y(0) Catch 0xb75432a8 (proc_lib:init_p_do_apply/3 + 44)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,1}]}, {heap_size,2584}, {total_heap_size,5168}, {links,[<0.55.0>,<0.65.0>,<0.72.0>,<0.64.0>,<0.52.0>]}, {memory,21196}, {message_queue_len,0}, {reductions,1799}, {trap_exit,true}]}, {<0.55.0>, [{registered_name,dist_manager}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb757151c (gen_server:loop/6 + 144)", "CP: 0x00000000 (invalid)","arity = 0", "0xb73dc0e0 Return addr 0xb7543298 (proc_lib:init_p_do_apply/3 + 28)", "y(0) []","y(1) infinity", "y(2) dist_manager", "y(3) {state,true,\"10.1.4.243\"}", "y(4) dist_manager","y(5) <0.54.0>", "0xb73dc0fc Return addr 0x08229d74 ()", "y(0) Catch 0xb75432a8 (proc_lib:init_p_do_apply/3 + 44)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,5}]}, {heap_size,233}, {total_heap_size,610}, {links,[<0.54.0>]}, {memory,2884}, {message_queue_len,0}, {reductions,429}, {trap_exit,false}]}, {<0.56.0>, [{registered_name,inet_gethost_native_sup}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb757151c (gen_server:loop/6 + 144)", "CP: 0x00000000 (invalid)","arity = 0", "0xb73aa2b8 Return addr 0xb7543298 (proc_lib:init_p_do_apply/3 + 28)", "y(0) []","y(1) infinity", "y(2) supervisor_bridge", "y(3) {state,inet_gethost_native,<0.57.0>,<0.57.0>,{local,inet_gethost_native_sup}}", "y(4) inet_gethost_native_sup", "y(5) <0.27.0>", "0xb73aa2d4 Return addr 0x08229d74 ()", "y(0) Catch 0xb75432a8 (proc_lib:init_p_do_apply/3 + 44)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,0}]}, {heap_size,233}, {total_heap_size,233}, {links,[<0.27.0>,<0.57.0>]}, {memory,1396}, {message_queue_len,0}, {reductions,41}, {trap_exit,true}]}, {<0.57.0>, [{registered_name,inet_gethost_native}, {status,waiting}, {initial_call,{inet_gethost_native,server_init,2}}, {backtrace, ["Program counter: 0xb75f1eac (inet_gethost_native:main_loop/1 + 20)", "CP: 0x00000000 (invalid)","arity = 0", "0xb6db0c28 Return addr 0x08229d74 ()", "y(0) {state,#Port<0.1138>,8000,8209,12306,<0.56.0>,4,{statistics,0,0,0,0,0,0,0,0}}"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,3}]}, {heap_size,610}, {total_heap_size,987}, {links,[<0.56.0>,#Port<0.1138>]}, {memory,4412}, {message_queue_len,0}, {reductions,282}, {trap_exit,true}]}, {<0.58.0>, [{registered_name,net_sup}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb757151c (gen_server:loop/6 + 144)", "CP: 0x00000000 (invalid)","arity = 0", "0xb73be44c Return addr 0xb7543298 (proc_lib:init_p_do_apply/3 + 28)", "y(0) []","y(1) infinity", "y(2) supervisor", "y(3) {state,{local,net_sup},one_for_all,[{child,<0.61.0>,net_kernel,{net_kernel,start_", "y(4) net_sup","y(5) <0.11.0>", "0xb73be468 Return addr 0x08229d74 ()", "y(0) Catch 0xb75432a8 (proc_lib:init_p_do_apply/3 + 44)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,1}]}, {heap_size,610}, {total_heap_size,1220}, {links,[<0.59.0>,<0.60.0>,<0.61.0>,<0.11.0>]}, {memory,5384}, {message_queue_len,0}, {reductions,252}, {trap_exit,true}]}, {<0.59.0>, [{registered_name,erl_epmd}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb757151c (gen_server:loop/6 + 144)", "CP: 0x00000000 (invalid)","arity = 0", "0xb73e58d0 Return addr 0xb7543298 (proc_lib:init_p_do_apply/3 + 28)", "y(0) []","y(1) infinity", "y(2) erl_epmd", "y(3) {state,#Port<0.1178>,21100,ns_1}", "y(4) erl_epmd","y(5) <0.58.0>", "0xb73e58ec Return addr 0x08229d74 ()", "y(0) Catch 0xb75432a8 (proc_lib:init_p_do_apply/3 + 44)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,0}]}, {heap_size,233}, {total_heap_size,233}, {links,[<0.58.0>,#Port<0.1178>]}, {memory,1396}, {message_queue_len,0}, {reductions,127}, {trap_exit,false}]}, {<0.60.0>, [{registered_name,auth}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb757151c (gen_server:loop/6 + 144)", "CP: 0x00000000 (invalid)","arity = 0", "0xb7408be0 Return addr 0xb7543298 (proc_lib:init_p_do_apply/3 + 28)", "y(0) []","y(1) infinity","y(2) auth", "y(3) {state,elyrdmyctqykhepq,16403}", "y(4) auth","y(5) <0.58.0>", "0xb7408bfc Return addr 0x08229d74 ()", "y(0) Catch 0xb75432a8 (proc_lib:init_p_do_apply/3 + 44)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,1}]}, {heap_size,233}, {total_heap_size,610}, {links,[<0.58.0>]}, {memory,2884}, {message_queue_len,0}, {reductions,171}, {trap_exit,true}]}, {<0.61.0>, [{registered_name,net_kernel}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb757151c (gen_server:loop/6 + 144)", "CP: 0x00000000 (invalid)","arity = 0", "0xb7391900 Return addr 0xb7543298 (proc_lib:init_p_do_apply/3 + 28)", "y(0) []","y(1) infinity", "y(2) net_kernel", "y(3) {state,'ns_1@10.1.4.243','ns_1@10.1.4.243',longnames,{tick,<0.63.0>,15000},7000,s", "y(4) net_kernel","y(5) <0.58.0>", "0xb739191c Return addr 0x08229d74 ()", "y(0) Catch 0xb75432a8 (proc_lib:init_p_do_apply/3 + 44)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,9}]}, {heap_size,377}, {total_heap_size,754}, {links, [<0.58.0>,<0.63.0>,<0.294.0>,<0.62.0>,#Port<0.1176>]}, {memory,3556}, {message_queue_len,0}, {reductions,993}, {trap_exit,true}]}, {<0.62.0>, [{registered_name,[]}, {status,waiting}, {initial_call,{inet_tcp_dist,accept_loop,2}}, {backtrace, ["Program counter: 0xb7512a08 (prim_inet:accept0/2 + 92)", "CP: 0x00000000 (invalid)","arity = 0", "0xb73e5c78 Return addr 0xb6cdd81c (inet_tcp:accept/1 + 20)", "y(0) 9949","y(1) #Port<0.1176>", "0xb73e5c84 Return addr 0xb6cde1e8 (inet_tcp_dist:accept_loop/2 + 48)", "y(0) []", "0xb73e5c8c Return addr 0x08229d74 ()", "y(0) []","y(1) #Port<0.1176>", "y(2) <0.61.0>"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,0}]}, {heap_size,233}, {total_heap_size,233}, {links,[<0.61.0>]}, {memory,1356}, {message_queue_len,0}, {reductions,268}, {trap_exit,false}]}, {<0.63.0>, [{registered_name,[]}, {status,waiting}, {initial_call,{net_kernel,ticker,2}}, {backtrace, ["Program counter: 0xb6c208fc (net_kernel:ticker_loop/2 + 28)", "CP: 0x00000000 (invalid)","arity = 0", "0xb73e6038 Return addr 0x08229d74 ()", "y(0) 15000","y(1) <0.61.0>"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,0}]}, {heap_size,233}, {total_heap_size,233}, {links,[<0.61.0>]}, {memory,1316}, {message_queue_len,0}, {reductions,45}, {trap_exit,false}]}, {<0.64.0>, [{registered_name,ns_cluster}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb757151c (gen_server:loop/6 + 144)", "CP: 0x00000000 (invalid)","arity = 0", "0xb6d31280 Return addr 0xb7543298 (proc_lib:init_p_do_apply/3 + 28)", "y(0) []","y(1) infinity", "y(2) ns_cluster","y(3) {state}", "y(4) ns_cluster","y(5) <0.54.0>", "0xb6d3129c Return addr 0x08229d74 ()", "y(0) Catch 0xb75432a8 (proc_lib:init_p_do_apply/3 + 44)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,1}]}, {heap_size,6765}, {total_heap_size,35422}, {links,[<0.54.0>]}, {memory,142132}, {message_queue_len,0}, {reductions,13400}, {trap_exit,false}]}, {<0.65.0>, [{registered_name,ns_config_sup}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb757151c (gen_server:loop/6 + 144)", "CP: 0x00000000 (invalid)","arity = 0", "0xb6d4d6c0 Return addr 0xb7543298 (proc_lib:init_p_do_apply/3 + 28)", "y(0) []","y(1) infinity", "y(2) supervisor", "y(3) {state,{local,ns_config_sup},rest_for_one,[{child,undefined,ns_config_log,{ns_con", "y(4) ns_config_sup","y(5) <0.54.0>", "0xb6d4d6dc Return addr 0x08229d74 ()", "y(0) Catch 0xb75432a8 (proc_lib:init_p_do_apply/3 + 44)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,0}]}, {heap_size,17711}, {total_heap_size,17711}, {links,[<0.66.0>,<0.67.0>,<0.54.0>]}, {memory,71328}, {message_queue_len,0}, {reductions,582}, {trap_exit,true}]}, {<0.66.0>, [{registered_name,ns_config_events}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb753925c (gen_event:fetch_msg/5 + 44)", "CP: 0x00000000 (invalid)","arity = 0", "0xb73e2aa8 Return addr 0xb7543308 (proc_lib:wake_up/3 + 60)", "y(0) false","y(1) []", "y(2) [{handler,ns_pubsub,#Ref<0.0.0.759>,{state,#Fun,ok},<0.17", "y(3) ns_config_events","y(4) <0.65.0>", "0xb73e2ac0 Return addr 0x08229d74 ()", "y(0) Catch 0xb7543318 (proc_lib:wake_up/3 + 76)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,21}]}, {heap_size,2584}, {total_heap_size,20295}, {links,[<0.110.0>,<0.162.0>,<0.174.0>,<0.65.0>]}, {memory,81844}, {message_queue_len,0}, {reductions,73987}, {trap_exit,true}]}, {<0.67.0>, [{registered_name,ns_config}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb757151c (gen_server:loop/6 + 144)", "CP: 0x00000000 (invalid)","arity = 0", "0xb63f9e24 Return addr 0xb7543298 (proc_lib:init_p_do_apply/3 + 28)", "y(0) []","y(1) infinity", "y(2) ns_config", "y(3) {config,{full,\"/etc/opt/membase/1.6.5.4r/config\",undefined,ns_config_default},[[{", "y(4) ns_config","y(5) <0.65.0>", "0xb63f9e40 Return addr 0x08229d74 ()", "y(0) Catch 0xb75432a8 (proc_lib:init_p_do_apply/3 + 44)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,14}]}, {heap_size,46368}, {total_heap_size,92736}, {links,[<0.65.0>]}, {memory,371388}, {message_queue_len,0}, {reductions,119124}, {trap_exit,true}]}, {<0.72.0>, [{registered_name,ns_server_sup}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb757151c (gen_server:loop/6 + 144)", "CP: 0x00000000 (invalid)","arity = 0", "0xb69dfe98 Return addr 0xb7543298 (proc_lib:init_p_do_apply/3 + 28)", "y(0) []","y(1) infinity", "y(2) supervisor", "y(3) {state,{local,ns_server_sup},one_for_one,[{child,<0.177.0>,ns_tick,{ns_tick,start", "y(4) ns_server_sup","y(5) <0.54.0>", "0xb69dfeb4 Return addr 0x08229d74 ()", "y(0) Catch 0xb75432a8 (proc_lib:init_p_do_apply/3 + 44)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,1}]}, {heap_size,28657}, {total_heap_size,225075}, {links, [<0.96.0>,<0.109.0>,<0.161.0>,<0.174.0>,<0.177.0>, <0.162.0>,<0.113.0>,<0.115.0>,<0.110.0>,<0.107.0>, <0.108.0>,<0.101.0>,<0.75.0>,<0.92.0>,<0.94.0>, <0.78.0>,<0.73.0>,<0.74.0>,<0.54.0>]}, {memory,901104}, {message_queue_len,0}, {reductions,99280}, {trap_exit,true}]}, {<0.73.0>, [{registered_name,ns_log}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb757151c (gen_server:loop/6 + 144)", "CP: 0x00000000 (invalid)","arity = 0", "0xb6dba078 Return addr 0xb7543298 (proc_lib:init_p_do_apply/3 + 28)", "y(0) []","y(1) infinity","y(2) ns_log", "y(3) {state,[{log_entry,{1304,813102,949990},'ns_1@127.0.0.1',ns_node_disco,3,\"Initial", "y(4) ns_log","y(5) <0.72.0>", "0xb6dba094 Return addr 0x08229d74 ()", "y(0) Catch 0xb75432a8 (proc_lib:init_p_do_apply/3 + 44)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,8}]}, {heap_size,4181}, {total_heap_size,32838}, {links,[<0.53.0>,<0.72.0>]}, {memory,131816}, {message_queue_len,0}, {reductions,3271}, {trap_exit,false}]}, {<0.74.0>, [{registered_name,ns_log_events}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x08229d70 (unknown function)", "CP: 0x08229d74 ()", "arity = 3"," proc_lib"," wake_up", " [gen_event,wake_hib,[<0.72.0>,ns_log_events,[{handler,ns_mail_log,false,{state},false}]"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,0}]}, {heap_size,34}, {total_heap_size,34}, {links,[<0.72.0>]}, {memory,580}, {message_queue_len,0}, {reductions,252}, {trap_exit,true}]}, {<0.75.0>, [{registered_name,ns_mail_sup}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb757151c (gen_server:loop/6 + 144)", "CP: 0x00000000 (invalid)","arity = 0", "0xb73d02b0 Return addr 0xb7543298 (proc_lib:init_p_do_apply/3 + 28)", "y(0) []","y(1) infinity", "y(2) supervisor", "y(3) {state,{local,ns_mail_sup},one_for_all,[{child,undefined,ns_mail_log,{ns_mail_log", "y(4) ns_mail_sup","y(5) <0.72.0>", "0xb73d02cc Return addr 0x08229d74 ()", "y(0) Catch 0xb75432a8 (proc_lib:init_p_do_apply/3 + 44)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,1}]}, {heap_size,233}, {total_heap_size,610}, {links,[<0.76.0>,<0.72.0>]}, {memory,2904}, {message_queue_len,0}, {reductions,661}, {trap_exit,true}]}, {<0.76.0>, [{registered_name,ns_mail}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb757151c (gen_server:loop/6 + 144)", "CP: 0x00000000 (invalid)","arity = 0", "0xb73dbd38 Return addr 0xb7543298 (proc_lib:init_p_do_apply/3 + 28)", "y(0) []","y(1) infinity","y(2) ns_mail", "y(3) empty_state","y(4) ns_mail", "y(5) <0.75.0>", "0xb73dbd54 Return addr 0x08229d74 ()", "y(0) Catch 0xb75432a8 (proc_lib:init_p_do_apply/3 + 44)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,0}]}, {heap_size,233}, {total_heap_size,233}, {links,[<0.75.0>]}, {memory,1376}, {message_queue_len,0}, {reductions,27}, {trap_exit,true}]}, {<0.78.0>, [{registered_name,ns_node_disco_sup}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb757151c (gen_server:loop/6 + 144)", "CP: 0x00000000 (invalid)","arity = 0", "0xb73d5728 Return addr 0xb7543298 (proc_lib:init_p_do_apply/3 + 28)", "y(0) []","y(1) infinity", "y(2) supervisor", "y(3) {state,{local,ns_node_disco_sup},rest_for_one,[{child,<0.86.0>,ns_config_rep,{ns_", "y(4) ns_node_disco_sup","y(5) <0.72.0>", "0xb73d5744 Return addr 0x08229d74 ()", "y(0) Catch 0xb75432a8 (proc_lib:init_p_do_apply/3 + 44)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,3}]}, {heap_size,377}, {total_heap_size,754}, {links,[<0.79.0>,<0.80.0>,<0.86.0>,<0.72.0>]}, {memory,3520}, {message_queue_len,0}, {reductions,832}, {trap_exit,true}]}, {<0.79.0>, [{registered_name,ns_node_disco_events}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb753925c (gen_event:fetch_msg/5 + 44)", "CP: 0x00000000 (invalid)","arity = 0", "0xb6d24664 Return addr 0xb7543308 (proc_lib:wake_up/3 + 60)", "y(0) false","y(1) []", "y(2) [{handler,menelaus_event,ns_node_disco_events,{state,ns_node_disco_events,undefin", "y(3) ns_node_disco_events","y(4) <0.78.0>", "0xb6d2467c Return addr 0x08229d74 ()", "y(0) Catch 0xb7543318 (proc_lib:wake_up/3 + 76)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,50}]}, {heap_size,233}, {total_heap_size,377}, {links,[<0.78.0>]}, {memory,2112}, {message_queue_len,0}, {reductions,7986}, {trap_exit,true}]}, {<0.80.0>, [{registered_name,ns_node_disco}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb757151c (gen_server:loop/6 + 144)", "CP: 0x00000000 (invalid)","arity = 0", "0xb68f96e0 Return addr 0xb7543298 (proc_lib:init_p_do_apply/3 + 28)", "y(0) []","y(1) infinity", "y(2) ns_node_disco", "y(3) {state,['ns_1@10.1.4.243','ns_1@10.1.4.244'],{interval,#Ref<0.0.0.153>}}", "y(4) ns_node_disco","y(5) <0.78.0>", "0xb68f96fc Return addr 0x08229d74 ()", "y(0) Catch 0xb75432a8 (proc_lib:init_p_do_apply/3 + 44)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,1595}]}, {heap_size,17711}, {total_heap_size,46368}, {links,[<0.78.0>,<0.53.0>]}, {memory,185936}, {message_queue_len,0}, {reductions,43376}, {trap_exit,false}]}, {<0.86.0>, [{registered_name,ns_config_rep}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb757151c (gen_server:loop/6 + 144)", "CP: 0x00000000 (invalid)","arity = 0", "0xb633788c Return addr 0xb7543298 (proc_lib:init_p_do_apply/3 + 28)", "y(0) []","y(1) infinity", "y(2) ns_config_rep","y(3) {state}", "y(4) ns_config_rep","y(5) <0.78.0>", "0xb63378a8 Return addr 0x08229d74 ()", "y(0) Catch 0xb75432a8 (proc_lib:init_p_do_apply/3 + 44)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,5}]}, {heap_size,46368}, {total_heap_size,75025}, {links,[<0.78.0>]}, {memory,300544}, {message_queue_len,0}, {reductions,6535}, {trap_exit,false}]}, {<0.92.0>, [{registered_name,ns_heart}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb757151c (gen_server:loop/6 + 144)", "CP: 0x00000000 (invalid)","arity = 0", "0xb74070b8 Return addr 0xb7543298 (proc_lib:init_p_do_apply/3 + 28)", "y(0) []","y(1) infinity", "y(2) ns_heart", "y(3) [{meminfo,<<1248 bytes>>},{replication,[{\"default\",1.000000e+00}]},{system_memory", "y(4) ns_heart","y(5) <0.72.0>", "0xb74070d4 Return addr 0x08229d74 ()", "y(0) Catch 0xb75432a8 (proc_lib:init_p_do_apply/3 + 44)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,0}]}, {heap_size,987}, {total_heap_size,987}, {links,[<0.72.0>,<0.53.0>]}, {memory,4412}, {message_queue_len,0}, {reductions,622406}, {trap_exit,false}]}, {<0.94.0>, [{registered_name,ns_doctor}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb757151c (gen_server:loop/6 + 144)", "CP: 0x00000000 (invalid)","arity = 0", "0xb7401ff0 Return addr 0xb7543298 (proc_lib:init_p_do_apply/3 + 28)", "y(0) []","y(1) infinity", "y(2) ns_doctor", "y(3) {state,{dict,2,16,16,8,80,48,{[],[],[],[],[],[],[],[],[],[],[],[],[],[],[],[]},{{", "y(4) ns_doctor","y(5) <0.72.0>", "0xb740200c Return addr 0x08229d74 ()", "y(0) Catch 0xb75432a8 (proc_lib:init_p_do_apply/3 + 44)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,2}]}, {heap_size,4181}, {total_heap_size,10946}, {links,[<0.72.0>,<0.53.0>]}, {memory,44248}, {message_queue_len,0}, {reductions,67155}, {trap_exit,false}]}, {<0.96.0>, [{registered_name,menelaus_sup}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb757151c (gen_server:loop/6 + 144)", "CP: 0x00000000 (invalid)","arity = 0", "0xb68444b0 Return addr 0xb7543298 (proc_lib:init_p_do_apply/3 + 28)", "y(0) []","y(1) infinity", "y(2) supervisor", "y(3) {state,{local,menelaus_sup},one_for_one,[{child,<0.100.0>,hot_keys_keeper,{hot_ke", "y(4) menelaus_sup","y(5) <0.72.0>", "0xb68444cc Return addr 0x08229d74 ()", "y(0) Catch 0xb75432a8 (proc_lib:init_p_do_apply/3 + 44)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,0}]}, {heap_size,17711}, {total_heap_size,17711}, {links,[<0.97.0>,<0.100.0>,<0.72.0>]}, {memory,71328}, {message_queue_len,0}, {reductions,1850}, {trap_exit,true}]}, {<0.97.0>, [{registered_name,menelaus_web}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb757151c (gen_server:loop/6 + 144)", "CP: 0x00000000 (invalid)","arity = 0", "0xb6db4d6c Return addr 0xb7543298 (proc_lib:init_p_do_apply/3 + 28)", "y(0) []","y(1) infinity", "y(2) mochiweb_socket_server", "y(3) {mochiweb_socket_server,8091,#Fun,{local,menelaus_web}", "y(4) menelaus_web","y(5) <0.96.0>", "0xb6db4d88 Return addr 0x08229d74 ()", "y(0) Catch 0xb75432a8 (proc_lib:init_p_do_apply/3 + 44)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,5}]}, {heap_size,610}, {total_heap_size,1220}, {links, [<0.567.0>,<0.5104.0>,<0.5219.0>,<0.5290.0>, <0.5185.0>,<0.4993.0>,<0.96.0>,<0.204.0>,<0.98.0>, #Port<0.1400>]}, {memory,5504}, {message_queue_len,0}, {reductions,917}, {trap_exit,true}]}, {<0.98.0>, [{registered_name,[]}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb6a46008 (menelaus_web:handle_streaming/4 + 196)", "CP: 0x00000000 (invalid)","arity = 0", "0xb6df4bd8 Return addr 0xb6a44360 (menelaus_web:loop/3 + 12136)", "y(0) {struct,[{buckets,[{struct,[{name,<<7 bytes>>},{nodeLocator,vbucket},{saslPasswor", "y(1) {struct,[{buckets,[{struct,[{name,<<7 bytes>>},{nodeLocator,vbucket},{saslPasswor", "y(2) {mochiweb_response,{mochiweb_request,#Port<0.1489>,'GET',\"/pools/default/saslBuck", "y(3) {mochiweb_request,#Port<0.1489>,'GET',\"/pools/default/saslBucketsStreaming\",{1,1}", "y(4) #Fun", "0xb6df4bf0 Return addr 0xb6a50ed0 (mochiweb_http:headers/5 + 680)", "y(0) []","y(1) []","y(2) []", "y(3) []", "y(4) {mochiweb_request,#Port<0.1489>,'GET',\"/pools/default/saslBucketsStreaming\",{1,1}", "y(5) Catch 0xb6a44370 (menelaus_web:loop/3 + 12152)", "0xb6df4c0c Return addr 0xb7543298 (proc_lib:init_p_do_apply/3 + 28)", "y(0) #Fun", "y(1) []","y(2) []", "y(3) {mochiweb_request,#Port<0.1489>,'GET',\"/pools/default/saslBucketsStreaming\",{1,1}", "0xb6df4c20 Return addr 0x08229d74 ()", "y(0) Catch 0xb75432a8 (proc_lib:init_p_do_apply/3 + 44)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,2}]}, {heap_size,10946}, {total_heap_size,85971}, {links,[<0.97.0>,#Port<0.1489>]}, {memory,344428}, {message_queue_len,0}, {reductions,1086183}, {trap_exit,false}]}, {<0.100.0>, [{registered_name,hot_keys_keeper}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb757151c (gen_server:loop/6 + 144)", "CP: 0x00000000 (invalid)","arity = 0", "0xb739d364 Return addr 0xb7543298 (proc_lib:init_p_do_apply/3 + 28)", "y(0) []","y(1) infinity", "y(2) hot_keys_keeper", "y(3) {state,[{\"default\",[]}],[{\"default\",[]}],<0.5337.0>}", "y(4) hot_keys_keeper","y(5) <0.96.0>", "0xb739d380 Return addr 0x08229d74 ()", "y(0) Catch 0xb75432a8 (proc_lib:init_p_do_apply/3 + 44)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,12}]}, {heap_size,610}, {total_heap_size,987}, {links,[<0.96.0>,<0.53.0>]}, {memory,4412}, {message_queue_len,0}, {reductions,2004}, {trap_exit,false}]}, {<0.101.0>, [{registered_name,ns_port_sup}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb757151c (gen_server:loop/6 + 144)", "CP: 0x00000000 (invalid)","arity = 0", "0xb69fbe60 Return addr 0xb7543298 (proc_lib:init_p_do_apply/3 + 28)", "y(0) []","y(1) infinity", "y(2) supervisor", "y(3) {state,{local,ns_port_sup},one_for_one,[{child,<0.105.0>,{memcached,\"./bin/memcac", "y(4) ns_port_sup","y(5) <0.72.0>", "0xb69fbe7c Return addr 0x08229d74 ()", "y(0) Catch 0xb75432a8 (proc_lib:init_p_do_apply/3 + 44)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,1}]}, {heap_size,28657}, {total_heap_size,75025}, {links,[<0.103.0>,<0.105.0>,<0.72.0>]}, {memory,300584}, {message_queue_len,0}, {reductions,3050}, {trap_exit,true}]}, {<0.103.0>, [{registered_name,[]}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb757151c (gen_server:loop/6 + 144)", "CP: 0x00000000 (invalid)","arity = 0", "0xb73a9f10 Return addr 0xb7543298 (proc_lib:init_p_do_apply/3 + 28)", "y(0) []","y(1) infinity", "y(2) supervisor_cushion", "y(3) {state,moxi,5000,{1304,814945,958438},<0.104.0>}", "y(4) <0.103.0>","y(5) <0.101.0>", "0xb73a9f2c Return addr 0x08229d74 ()", "y(0) Catch 0xb75432a8 (proc_lib:init_p_do_apply/3 + 44)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,1}]}, {heap_size,987}, {total_heap_size,2584}, {links,[<0.101.0>,<0.104.0>]}, {memory,10800}, {message_queue_len,0}, {reductions,63}, {trap_exit,true}]}, {<0.104.0>, [{registered_name,[]}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb757151c (gen_server:loop/6 + 144)", "CP: 0x00000000 (invalid)","arity = 0", "0xb73e0240 Return addr 0xb7543298 (proc_lib:init_p_do_apply/3 + 28)", "y(0) []","y(1) infinity", "y(2) ns_port_server", "y(3) {state,#Port<0.1439>,moxi,{[\"2011-05-07 17:35:45: (cproxy_config.c.325) env: MOXI", "y(4) <0.104.0>","y(5) <0.103.0>", "0xb73e025c Return addr 0x08229d74 ()", "y(0) Catch 0xb75432a8 (proc_lib:init_p_do_apply/3 + 44)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,1}]}, {heap_size,1597}, {total_heap_size,4181}, {links,[<0.103.0>,#Port<0.1439>]}, {memory,17188}, {message_queue_len,0}, {reductions,317}, {trap_exit,true}]}, {<0.105.0>, [{registered_name,[]}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb757151c (gen_server:loop/6 + 144)", "CP: 0x00000000 (invalid)","arity = 0", "0xb6d2a8c8 Return addr 0xb7543298 (proc_lib:init_p_do_apply/3 + 28)", "y(0) []","y(1) infinity", "y(2) supervisor_cushion", "y(3) {state,memcached,5000,{1304,814945,961795},<0.106.0>}", "y(4) <0.105.0>","y(5) <0.101.0>", "0xb6d2a8e4 Return addr 0x08229d74 ()", "y(0) Catch 0xb75432a8 (proc_lib:init_p_do_apply/3 + 44)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,0}]}, {heap_size,987}, {total_heap_size,987}, {links,[<0.101.0>,<0.106.0>]}, {memory,4412}, {message_queue_len,0}, {reductions,55}, {trap_exit,true}]}, {<0.106.0>, [{registered_name,[]}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb757151c (gen_server:loop/6 + 144)", "CP: 0x00000000 (invalid)","arity = 0", "0xb6d242b4 Return addr 0xb7543298 (proc_lib:init_p_do_apply/3 + 28)", "y(0) []","y(1) infinity", "y(2) ns_port_server", "y(3) {state,#Port<0.1446>,memcached,{[\"Vbucket ","y(5) <0.105.0>", "0xb6d242d0 Return addr 0x08229d74 ()", "y(0) Catch 0xb75432a8 (proc_lib:init_p_do_apply/3 + 44)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,1}]}, {heap_size,610}, {total_heap_size,1597}, {links,[<0.105.0>,#Port<0.1446>]}, {memory,6852}, {message_queue_len,0}, {reductions,329}, {trap_exit,true}]}, {<0.107.0>, [{registered_name,ns_tick_event}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb753925c (gen_event:fetch_msg/5 + 44)", "CP: 0x00000000 (invalid)","arity = 0", "0xb739b6bc Return addr 0xb7543298 (proc_lib:init_p_do_apply/3 + 28)", "y(0) false","y(1) []", "y(2) [{handler,ns_pubsub,#Ref<0.0.0.687>,{state,#Fun,ignored},<0", "y(3) ns_tick_event","y(4) <0.72.0>", "0xb739b6d4 Return addr 0x08229d74 ()", "y(0) Catch 0xb75432a8 (proc_lib:init_p_do_apply/3 + 44)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,45}]}, {heap_size,233}, {total_heap_size,610}, {links,[<0.72.0>,<0.167.0>]}, {memory,2904}, {message_queue_len,0}, {reductions,3560}, {trap_exit,true}]}, {<0.108.0>, [{registered_name,ns_stats_event}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb753925c (gen_event:fetch_msg/5 + 44)", "CP: 0x00000000 (invalid)","arity = 0", "0xb739c04c Return addr 0xb7543298 (proc_lib:init_p_do_apply/3 + 28)", "y(0) false","y(1) []", "y(2) [{handler,ns_pubsub,#Ref<0.0.0.740>,{state,#Fun,ignored},<0", "y(3) ns_stats_event","y(4) <0.72.0>", "0xb739c064 Return addr 0x08229d74 ()", "y(0) Catch 0xb75432a8 (proc_lib:init_p_do_apply/3 + 44)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,635}]}, {heap_size,377}, {total_heap_size,987}, {links,[<0.168.0>,<0.4993.0>,<0.5185.0>,<0.72.0>]}, {memory,4452}, {message_queue_len,0}, {reductions,7842}, {trap_exit,true}]}, {<0.109.0>, [{registered_name,ns_good_bucket_worker}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb757151c (gen_server:loop/6 + 144)", "CP: 0x00000000 (invalid)","arity = 0", "0xb6d3c138 Return addr 0xb7543298 (proc_lib:init_p_do_apply/3 + 28)", "y(0) []","y(1) infinity", "y(2) work_queue","y(3) []", "y(4) ns_good_bucket_worker","y(5) <0.72.0>", "0xb6d3c154 Return addr 0x08229d74 ()", "y(0) Catch 0xb75432a8 (proc_lib:init_p_do_apply/3 + 44)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,0}]}, {heap_size,233}, {total_heap_size,233}, {links,[<0.72.0>]}, {memory,1376}, {message_queue_len,0}, {reductions,59}, {trap_exit,false}]}, {<0.110.0>, [{registered_name,ns_good_bucket_sup}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb757151c (gen_server:loop/6 + 144)", "CP: 0x00000000 (invalid)","arity = 0", "0xb6855970 Return addr 0xb7543298 (proc_lib:init_p_do_apply/3 + 28)", "y(0) []","y(1) infinity", "y(2) supervisor", "y(3) {state,{local,ns_good_bucket_sup},one_for_one,[{child,<0.112.0>,{ns_memcached,\"de", "y(4) ns_good_bucket_sup","y(5) <0.72.0>", "0xb685598c Return addr 0x08229d74 ()", "y(0) Catch 0xb75432a8 (proc_lib:init_p_do_apply/3 + 44)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,0}]}, {heap_size,17711}, {total_heap_size,17711}, {links,[<0.72.0>,<0.111.0>,<0.112.0>,<0.66.0>]}, {memory,71348}, {message_queue_len,0}, {reductions,217}, {trap_exit,true}]}, {<0.111.0>, [{registered_name,'ns_vbm_sup-default'}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb757151c (gen_server:loop/6 + 144)", "CP: 0x00000000 (invalid)","arity = 0", "0xb6e0c498 Return addr 0xb7543298 (proc_lib:init_p_do_apply/3 + 28)", "y(0) []","y(1) infinity", "y(2) supervisor", "y(3) {state,{local,'ns_vbm_sup-default'},one_for_one,[],{dict,0,16,16,8,80,48,{[],[],[", "y(4) 'ns_vbm_sup-default'","y(5) <0.110.0>", "0xb6e0c4b4 Return addr 0x08229d74 ()", "y(0) Catch 0xb75432a8 (proc_lib:init_p_do_apply/3 + 44)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,0}]}, {heap_size,233}, {total_heap_size,233}, {links,[<0.110.0>]}, {memory,1376}, {message_queue_len,0}, {reductions,586}, {trap_exit,true}]}, {<0.112.0>, [{registered_name,'ns_memcached-default'}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb757151c (gen_server:loop/6 + 144)", "CP: 0x00000000 (invalid)","arity = 0", "0xb69bfdb8 Return addr 0xb7543298 (proc_lib:init_p_do_apply/3 + 28)", "y(0) []","y(1) infinity", "y(2) ns_memcached", "y(3) {state,\"default\",#Port<0.2576>}", "y(4) <0.112.0>","y(5) <0.110.0>", "0xb69bfdd4 Return addr 0x08229d74 ()", "y(0) Catch 0xb75432a8 (proc_lib:init_p_do_apply/3 + 44)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,46}]}, {heap_size,6765}, {total_heap_size,35422}, {links,[<0.53.0>,<0.110.0>,#Port<0.2576>]}, {memory,142172}, {message_queue_len,0}, {reductions,2087570}, {trap_exit,true}]}, {<0.113.0>, [{registered_name,[]}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb6a6d1fc (gen_fsm:loop/7 + 156)", "CP: 0x00000000 (invalid)","arity = 0", "0xb6de8214 Return addr 0xb7543298 (proc_lib:init_p_do_apply/3 + 28)", "y(0) []","y(1) infinity", "y(2) ns_orchestrator", "y(3) {rebalancing_state,<0.375.0>,{dict,2,16,16,8,80,48,{[],[],[],[],[],[],[],[],[],[]", "y(4) rebalancing","y(5) ns_orchestrator", "y(6) <0.72.0>", "0xb6de8234 Return addr 0x08229d74 ()", "y(0) Catch 0xb75432a8 (proc_lib:init_p_do_apply/3 + 44)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,28}]}, {heap_size,4181}, {total_heap_size,32838}, {links,[<0.72.0>,<6517.312.0>,<0.375.0>,<0.53.0>]}, {memory,131976}, {message_queue_len,0}, {reductions,146309}, {trap_exit,true}]}, {<0.115.0>, [{registered_name,ns_mnesia}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb757151c (gen_server:loop/6 + 144)", "CP: 0x00000000 (invalid)","arity = 0", "0xb740efc8 Return addr 0xb7543298 (proc_lib:init_p_do_apply/3 + 28)", "y(0) []","y(1) infinity", "y(2) ns_mnesia","y(3) {state}", "y(4) ns_mnesia","y(5) <0.72.0>", "0xb740efe4 Return addr 0x08229d74 ()", "y(0) Catch 0xb75432a8 (proc_lib:init_p_do_apply/3 + 44)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,4}]}, {heap_size,1597}, {total_heap_size,1974}, {links,[<0.72.0>,<0.123.0>]}, {memory,8360}, {message_queue_len,0}, {reductions,1185}, {trap_exit,true}]}, {<0.117.0>, [{registered_name,[]}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb75a43b0 (application_master:main_loop/2 + 32)", "CP: 0x00000000 (invalid)","arity = 0", "0xb6d23934 Return addr 0xb7543298 (proc_lib:init_p_do_apply/3 + 28)", "y(0) []", "y(1) {state,<0.118.0>,{appl_data,mnesia,[mnesia_dumper_load_regulator,mnesia_event,mne", "y(2) <0.7.0>", "0xb6d23944 Return addr 0x08229d74 ()", "y(0) Catch 0xb75432a8 (proc_lib:init_p_do_apply/3 + 44)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,6}]}, {heap_size,377}, {total_heap_size,754}, {links,[<0.7.0>,<0.118.0>]}, {memory,3480}, {message_queue_len,0}, {reductions,47}, {trap_exit,true}]}, {<0.118.0>, [{registered_name,[]}, {status,waiting}, {initial_call,{application_master,start_it,4}}, {backtrace, ["Program counter: 0xb75a56f4 (application_master:loop_it/4 + 40)", "CP: 0x00000000 (invalid)","arity = 0", "0xb6d7dd50 Return addr 0x08229d74 ()", "y(0) {normal,[]}","y(1) mnesia_sup", "y(2) <0.119.0>","y(3) <0.117.0>"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,0}]}, {heap_size,233}, {total_heap_size,233}, {links,[<0.117.0>,<0.119.0>]}, {memory,1336}, {message_queue_len,0}, {reductions,40}, {trap_exit,true}]}, {<0.119.0>, [{registered_name,mnesia_sup}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb757151c (gen_server:loop/6 + 144)", "CP: 0x00000000 (invalid)","arity = 0", "0xb73e3c40 Return addr 0xb7543298 (proc_lib:init_p_do_apply/3 + 28)", "y(0) []","y(1) infinity", "y(2) supervisor", "y(3) {state,{local,mnesia_sup},one_for_all,[{child,<0.121.0>,mnesia_kernel_sup,{mnesia", "y(4) mnesia_sup","y(5) <0.118.0>", "0xb73e3c5c Return addr 0x08229d74 ()", "y(0) Catch 0xb75432a8 (proc_lib:init_p_do_apply/3 + 44)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,1}]}, {heap_size,377}, {total_heap_size,754}, {links,[<0.120.0>,<0.121.0>,<0.118.0>]}, {memory,3500}, {message_queue_len,0}, {reductions,207}, {trap_exit,true}]}, {<0.120.0>, [{registered_name,mnesia_event}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb753925c (gen_event:fetch_msg/5 + 44)", "CP: 0x00000000 (invalid)","arity = 0", "0xb6dbd5cc Return addr 0xb7543298 (proc_lib:init_p_do_apply/3 + 28)", "y(0) false","y(1) []", "y(2) [{handler,mnesia_event,false,{state,[],false,[]},false}]", "y(3) mnesia_event","y(4) <0.119.0>", "0xb6dbd5e4 Return addr 0x08229d74 ()", "y(0) Catch 0xb75432a8 (proc_lib:init_p_do_apply/3 + 44)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,6}]}, {heap_size,377}, {total_heap_size,754}, {links,[<0.119.0>,<0.123.0>]}, {memory,3480}, {message_queue_len,0}, {reductions,425}, {trap_exit,true}]}, {<0.121.0>, [{registered_name,mnesia_kernel_sup}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb757151c (gen_server:loop/6 + 144)", "CP: 0x00000000 (invalid)","arity = 0", "0xb6d24fec Return addr 0xb7543298 (proc_lib:init_p_do_apply/3 + 28)", "y(0) []","y(1) infinity", "y(2) supervisor", "y(3) {state,{local,mnesia_kernel_sup},one_for_all,[{child,<0.157.0>,mnesia_late_loader", "y(4) mnesia_kernel_sup","y(5) <0.119.0>", "0xb6d25008 Return addr 0x08229d74 ()", "y(0) Catch 0xb75432a8 (proc_lib:init_p_do_apply/3 + 44)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,4}]}, {heap_size,610}, {total_heap_size,987}, {links, [<0.124.0>,<0.155.0>,<0.156.0>,<0.157.0>,<0.126.0>, <0.154.0>,<0.125.0>,<0.122.0>,<0.123.0>,<0.119.0>]}, {memory,4572}, {message_queue_len,0}, {reductions,593}, {trap_exit,true}]}, {<0.122.0>, [{registered_name,mnesia_monitor}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb757151c (gen_server:loop/6 + 144)", "CP: 0x00000000 (invalid)","arity = 0", "0xb7396480 Return addr 0xb7543298 (proc_lib:init_p_do_apply/3 + 28)", "y(0) []","y(1) infinity", "y(2) mnesia_monitor", "y(3) {state,<0.121.0>,[],[],true,[],undefined,[]}", "y(4) mnesia_monitor","y(5) <0.121.0>", "0xb739649c Return addr 0x08229d74 ()", "y(0) Catch 0xb75432a8 (proc_lib:init_p_do_apply/3 + 44)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,14}]}, {heap_size,377}, {total_heap_size,987}, {links,[<0.152.0>,<0.121.0>]}, {memory,4412}, {message_queue_len,0}, {reductions,3405}, {trap_exit,true}]}, {<0.123.0>, [{registered_name,mnesia_subscr}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb757151c (gen_server:loop/6 + 144)", "CP: 0x00000000 (invalid)","arity = 0", "0xb7396828 Return addr 0xb7543298 (proc_lib:init_p_do_apply/3 + 28)", "y(0) []","y(1) infinity", "y(2) mnesia_subscr", "y(3) {state,<0.121.0>,20503}", "y(4) mnesia_subscr","y(5) <0.121.0>", "0xb7396844 Return addr 0x08229d74 ()", "y(0) Catch 0xb75432a8 (proc_lib:init_p_do_apply/3 + 44)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,0}]}, {heap_size,233}, {total_heap_size,233}, {links,[<0.120.0>,<0.121.0>,<0.115.0>]}, {memory,1416}, {message_queue_len,0}, {reductions,111}, {trap_exit,true}]}, {<0.124.0>, [{registered_name,mnesia_locker}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb6b05d28 (mnesia_locker:loop/1 + 20)", "CP: 0x00000000 (invalid)","arity = 0", "0xb6dbc2e4 Return addr 0xb6ad46b0 (mnesia_sp:init_proc/4 + 132)", "y(0) []","y(1) []","y(2) []", "y(3) []","y(4) []", "y(5) {state,<0.121.0>}", "0xb6dbc300 Return addr 0xb7543298 (proc_lib:init_p_do_apply/3 + 28)", "y(0) Catch 0xb6ad46b0 (mnesia_sp:init_proc/4 + 132)", "y(1) mnesia_locker","y(2) []","y(3) []", "y(4) [<0.121.0>]", "0xb6dbc318 Return addr 0x08229d74 ()", "y(0) Catch 0xb75432a8 (proc_lib:init_p_do_apply/3 + 44)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,133}]}, {heap_size,610}, {total_heap_size,987}, {links,[<0.121.0>]}, {memory,4392}, {message_queue_len,0}, {reductions,48493}, {trap_exit,true}]}, {<0.125.0>, [{registered_name,mnesia_recover}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb757151c (gen_server:loop/6 + 144)", "CP: 0x00000000 (invalid)","arity = 0", "0xb655b028 Return addr 0xb7543298 (proc_lib:init_p_do_apply/3 + 28)", "y(0) []","y(1) infinity", "y(2) mnesia_recover", "y(3) {state,<0.121.0>,undefined,undefined,undefined,0,true,[]}", "y(4) mnesia_recover","y(5) <0.121.0>", "0xb655b044 Return addr 0x08229d74 ()", "y(0) Catch 0xb75432a8 (proc_lib:init_p_do_apply/3 + 44)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,2}]}, {heap_size,6765}, {total_heap_size,13530}, {links,[<0.121.0>,<0.53.0>]}, {memory,54584}, {message_queue_len,0}, {reductions,11763}, {trap_exit,true}]}, {<0.126.0>, [{registered_name,mnesia_tm}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb6572d44 (mnesia_tm:doit_loop/1 + 108)", "CP: 0x00000000 (invalid)","arity = 0", "0xb740d1ac Return addr 0xb6ad46b0 (mnesia_sp:init_proc/4 + 132)", "y(0) []","y(1) []", "y(2) {state,{0,nil},{0,nil},<0.121.0>,[],[],[]}", "y(3) []","y(4) []","y(5) <0.121.0>", "y(6) {0,nil}","y(7) {0,nil}", "0xb740d1d0 Return addr 0xb7543298 (proc_lib:init_p_do_apply/3 + 28)", "y(0) Catch 0xb6ad46b0 (mnesia_sp:init_proc/4 + 132)", "y(1) mnesia_tm","y(2) []","y(3) []", "y(4) [<0.121.0>]", "0xb740d1e8 Return addr 0x08229d74 ()", "y(0) Catch 0xb75432a8 (proc_lib:init_p_do_apply/3 + 44)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,26}]}, {heap_size,2584}, {total_heap_size,20295}, {links,[<0.121.0>]}, {memory,81624}, {message_queue_len,0}, {reductions,139055}, {trap_exit,true}]}, {<0.127.0>, [{registered_name,dets_sup}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb757151c (gen_server:loop/6 + 144)", "CP: 0x00000000 (invalid)","arity = 0", "0xb73e5308 Return addr 0xb7543298 (proc_lib:init_p_do_apply/3 + 28)", "y(0) []","y(1) infinity", "y(2) supervisor", "y(3) {state,{local,dets_sup},simple_one_for_one,[{child,undefined,dets,{dets,istart_li", "y(4) dets_sup","y(5) <0.27.0>", "0xb73e5324 Return addr 0x08229d74 ()", "y(0) Catch 0xb75432a8 (proc_lib:init_p_do_apply/3 + 44)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,0}]}, {heap_size,377}, {total_heap_size,377}, {links,[<0.27.0>]}, {memory,1952}, {message_queue_len,0}, {reductions,120}, {trap_exit,true}]}, {<0.128.0>, [{registered_name,dets}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb757151c (gen_server:loop/6 + 144)", "CP: 0x00000000 (invalid)","arity = 0", "0xb6db3470 Return addr 0xb7543298 (proc_lib:init_p_do_apply/3 + 28)", "y(0) []","y(1) infinity", "y(2) dets_server", "y(3) {state,28704,[<0.27.0>],[]}", "y(4) dets","y(5) <0.27.0>", "0xb6db348c Return addr 0x08229d74 ()", "y(0) Catch 0xb75432a8 (proc_lib:init_p_do_apply/3 + 44)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,1}]}, {heap_size,233}, {total_heap_size,610}, {links,[<0.27.0>]}, {memory,2884}, {message_queue_len,0}, {reductions,140}, {trap_exit,true}]}, {<0.132.0>, [{registered_name,disk_log_sup}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb757151c (gen_server:loop/6 + 144)", "CP: 0x00000000 (invalid)","arity = 0", "0xb6dbb970 Return addr 0xb7543298 (proc_lib:init_p_do_apply/3 + 28)", "y(0) []","y(1) infinity", "y(2) supervisor", "y(3) {state,{local,disk_log_sup},simple_one_for_one,[{child,undefined,disk_log,{disk_l", "y(4) disk_log_sup","y(5) <0.27.0>", "0xb6dbb98c Return addr 0x08229d74 ()", "y(0) Catch 0xb75432a8 (proc_lib:init_p_do_apply/3 + 44)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,8}]}, {heap_size,1597}, {total_heap_size,2207}, {links,[<0.152.0>,<0.27.0>]}, {memory,9292}, {message_queue_len,0}, {reductions,1910}, {trap_exit,true}]}, {<0.133.0>, [{registered_name,disk_log_server}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb757151c (gen_server:loop/6 + 144)", "CP: 0x00000000 (invalid)","arity = 0", "0xb6d2972c Return addr 0xb7543298 (proc_lib:init_p_do_apply/3 + 28)", "y(0) []","y(1) infinity", "y(2) disk_log_server","y(3) {state,[]}", "y(4) disk_log_server","y(5) <0.27.0>", "0xb6d29748 Return addr 0x08229d74 ()", "y(0) Catch 0xb75432a8 (proc_lib:init_p_do_apply/3 + 44)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,33}]}, {heap_size,610}, {total_heap_size,1220}, {links,[<0.152.0>,<0.27.0>]}, {memory,5344}, {message_queue_len,0}, {reductions,1930}, {trap_exit,true}]}, {<0.152.0>, [{registered_name,[]}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb6611198 (disk_log:loop/1 + 84)", "CP: 0x00000000 (invalid)","arity = 0", "0xb6e18694 Return addr 0xb7543298 (proc_lib:init_p_do_apply/3 + 28)", "y(0) {state,[],[],<0.132.0>,<0.133.0>,357,{arg,latest_log,undefined,\"/var/opt/membase/", "0xb6e1869c Return addr 0x08229d74 ()", "y(0) Catch 0xb75432a8 (proc_lib:init_p_do_apply/3 + 44)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,4}]}, {heap_size,1597}, {total_heap_size,3194}, {links,[<0.132.0>,<0.133.0>,<0.122.0>,#Port<0.7516>]}, {memory,13280}, {message_queue_len,0}, {reductions,34662}, {trap_exit,true}]}, {<0.154.0>, [{registered_name,mnesia_checkpoint_sup}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb757151c (gen_server:loop/6 + 144)", "CP: 0x00000000 (invalid)","arity = 0", "0xb740fb88 Return addr 0xb7543298 (proc_lib:init_p_do_apply/3 + 28)", "y(0) []","y(1) infinity", "y(2) supervisor", "y(3) {state,{local,mnesia_checkpoint_sup},simple_one_for_one,[{child,undefined,mnesia_", "y(4) mnesia_checkpoint_sup", "y(5) <0.121.0>", "0xb740fba4 Return addr 0x08229d74 ()", "y(0) Catch 0xb75432a8 (proc_lib:init_p_do_apply/3 + 44)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,0}]}, {heap_size,233}, {total_heap_size,233}, {links,[<0.121.0>]}, {memory,1376}, {message_queue_len,0}, {reductions,61}, {trap_exit,true}]}, {<0.155.0>, [{registered_name,mnesia_snmp_sup}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb757151c (gen_server:loop/6 + 144)", "CP: 0x00000000 (invalid)","arity = 0", "0xb73e3070 Return addr 0xb7543298 (proc_lib:init_p_do_apply/3 + 28)", "y(0) []","y(1) infinity", "y(2) supervisor", "y(3) {state,{local,mnesia_snmp_sup},simple_one_for_one,[{child,undefined,mnesia_snmp_s", "y(4) mnesia_snmp_sup","y(5) <0.121.0>", "0xb73e308c Return addr 0x08229d74 ()", "y(0) Catch 0xb75432a8 (proc_lib:init_p_do_apply/3 + 44)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,0}]}, {heap_size,233}, {total_heap_size,233}, {links,[<0.121.0>]}, {memory,1376}, {message_queue_len,0}, {reductions,61}, {trap_exit,true}]}, {<0.156.0>, [{registered_name,mnesia_controller}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb757151c (gen_server:loop/6 + 144)", "CP: 0x00000000 (invalid)","arity = 0", "0xb686e758 Return addr 0xb7543298 (proc_lib:init_p_do_apply/3 + 28)", "y(0) []","y(1) infinity", "y(2) mnesia_controller", "y(3) {state,<0.121.0>,true,[],[],{0,nil},[],[],{0,nil},undefined,[],[],{interval,#Ref<", "y(4) mnesia_controller","y(5) <0.121.0>", "0xb686e774 Return addr 0x08229d74 ()", "y(0) Catch 0xb75432a8 (proc_lib:init_p_do_apply/3 + 44)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,8}]}, {heap_size,987}, {total_heap_size,1364}, {links,[<0.121.0>,<0.53.0>]}, {memory,5920}, {message_queue_len,0}, {reductions,1501}, {trap_exit,true}]}, {<0.157.0>, [{registered_name,mnesia_late_loader}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb65d4da4 (mnesia_late_loader:loop/1 + 20)", "CP: 0x00000000 (invalid)","arity = 0", "0xb6db43d4 Return addr 0xb6ad46b0 (mnesia_sp:init_proc/4 + 132)", "y(0) []","y(1) []", "y(2) {state,<0.121.0>}", "0xb6db43e4 Return addr 0xb7543298 (proc_lib:init_p_do_apply/3 + 28)", "y(0) Catch 0xb6ad46b0 (mnesia_sp:init_proc/4 + 132)", "y(1) mnesia_late_loader","y(2) []", "y(3) []","y(4) [<0.121.0>]", "0xb6db43fc Return addr 0x08229d74 ()", "y(0) Catch 0xb75432a8 (proc_lib:init_p_do_apply/3 + 44)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,2}]}, {heap_size,987}, {total_heap_size,1597}, {links,[<0.121.0>]}, {memory,6832}, {message_queue_len,0}, {reductions,509}, {trap_exit,false}]}, {<0.161.0>, [{registered_name,ns_bad_bucket_worker}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb757151c (gen_server:loop/6 + 144)", "CP: 0x00000000 (invalid)","arity = 0", "0xb740ff30 Return addr 0xb7543298 (proc_lib:init_p_do_apply/3 + 28)", "y(0) []","y(1) infinity", "y(2) work_queue","y(3) []", "y(4) ns_bad_bucket_worker","y(5) <0.72.0>", "0xb740ff4c Return addr 0x08229d74 ()", "y(0) Catch 0xb75432a8 (proc_lib:init_p_do_apply/3 + 44)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,1}]}, {heap_size,233}, {total_heap_size,610}, {links,[<0.72.0>]}, {memory,2884}, {message_queue_len,0}, {reductions,61}, {trap_exit,false}]}, {<0.162.0>, [{registered_name,ns_bad_bucket_sup}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb757151c (gen_server:loop/6 + 144)", "CP: 0x00000000 (invalid)","arity = 0", "0xb69a7f40 Return addr 0xb7543298 (proc_lib:init_p_do_apply/3 + 28)", "y(0) []","y(1) infinity", "y(2) supervisor", "y(3) {state,{local,ns_bad_bucket_sup},one_for_one,[{child,<0.171.0>,{stats_reader,\"def", "y(4) ns_bad_bucket_sup","y(5) <0.72.0>", "0xb69a7f5c Return addr 0x08229d74 ()", "y(0) Catch 0xb75432a8 (proc_lib:init_p_do_apply/3 + 44)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,0}]}, {heap_size,17711}, {total_heap_size,17711}, {links, [<0.72.0>,<0.168.0>,<0.171.0>,<0.167.0>,<0.66.0>]}, {memory,71368}, {message_queue_len,0}, {reductions,291}, {trap_exit,true}]}, {<0.167.0>, [{registered_name,[]}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb757151c (gen_server:loop/6 + 144)", "CP: 0x00000000 (invalid)","arity = 0", "0xb6e16d88 Return addr 0xb7543298 (proc_lib:init_p_do_apply/3 + 28)", "y(0) []","y(1) infinity", "y(2) stats_collector", "y(3) {state,\"default\",[44285,2615692,0,0,0,0,0,0,0,0,0,0,0,0,0,205000,0,0,0,0,0,0,200,", "y(4) <0.167.0>","y(5) <0.162.0>", "0xb6e16da4 Return addr 0x08229d74 ()", "y(0) Catch 0xb75432a8 (proc_lib:init_p_do_apply/3 + 44)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,3}]}, {heap_size,4181}, {total_heap_size,10946}, {links,[<0.162.0>,<0.107.0>]}, {memory,44248}, {message_queue_len,0}, {reductions,2301741}, {trap_exit,false}]}, {<0.168.0>, [{registered_name,'stats_archiver-default'}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb757151c (gen_server:loop/6 + 144)", "CP: 0x00000000 (invalid)","arity = 0", "0xb655c928 Return addr 0xb7543298 (proc_lib:init_p_do_apply/3 + 28)", "y(0) []","y(1) infinity", "y(2) stats_archiver", "y(3) {state,\"default\"}", "y(4) 'stats_archiver-default'", "y(5) <0.162.0>", "0xb655c944 Return addr 0x08229d74 ()", "y(0) Catch 0xb75432a8 (proc_lib:init_p_do_apply/3 + 44)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,44}]}, {heap_size,1597}, {total_heap_size,12543}, {links,[<0.108.0>,<0.162.0>,<0.53.0>]}, {memory,50656}, {message_queue_len,0}, {reductions,994247}, {trap_exit,false}]}, {<0.171.0>, [{registered_name,'stats_reader-default'}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb757151c (gen_server:loop/6 + 144)", "CP: 0x00000000 (invalid)","arity = 0", "0xb53662c8 Return addr 0xb7543298 (proc_lib:init_p_do_apply/3 + 28)", "y(0) []","y(1) infinity", "y(2) stats_reader", "y(3) {state,\"default\"}", "y(4) 'stats_reader-default'", "y(5) <0.162.0>", "0xb53662e4 Return addr 0x08229d74 ()", "y(0) Catch 0xb75432a8 (proc_lib:init_p_do_apply/3 + 44)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,21}]}, {heap_size,514229}, {total_heap_size,1028458}, {links,[<0.162.0>]}, {memory,4114276}, {message_queue_len,0}, {reductions,265929}, {trap_exit,false}]}, {<0.174.0>, [{registered_name,ns_moxi_sup}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb757151c (gen_server:loop/6 + 144)", "CP: 0x00000000 (invalid)","arity = 0", "0xb694d638 Return addr 0xb7543298 (proc_lib:init_p_do_apply/3 + 28)", "y(0) []","y(1) infinity", "y(2) supervisor", "y(3) {state,{local,ns_moxi_sup},one_for_one,[],{dict,0,16,16,8,80,48,{[],[],[],[],[],[", "y(4) ns_moxi_sup","y(5) <0.72.0>", "0xb694d654 Return addr 0x08229d74 ()", "y(0) Catch 0xb75432a8 (proc_lib:init_p_do_apply/3 + 44)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,0}]}, {heap_size,17711}, {total_heap_size,17711}, {links,[<0.72.0>,<0.66.0>]}, {memory,71308}, {message_queue_len,0}, {reductions,684}, {trap_exit,true}]}, {<0.177.0>, [{registered_name,[]}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb757151c (gen_server:loop/6 + 144)", "CP: 0x00000000 (invalid)","arity = 0", "0xb7398710 Return addr 0xb7543298 (proc_lib:init_p_do_apply/3 + 28)", "y(0) []","y(1) infinity","y(2) ns_tick", "y(3) {state,1304815267293}","y(4) ns_tick", "y(5) <0.72.0>", "0xb739872c Return addr 0x08229d74 ()", "y(0) Catch 0xb75432a8 (proc_lib:init_p_do_apply/3 + 44)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,106}]}, {heap_size,233}, {total_heap_size,610}, {links,[<0.72.0>,<6517.348.0>,<0.53.0>]}, {memory,3044}, {message_queue_len,0}, {reductions,10800}, {trap_exit,false}]}, {<0.204.0>, [{registered_name,[]}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb6a44e3c (menelaus_web:handle_pool_info_wait/6 + 140)", "CP: 0x00000000 (invalid)","arity = 0", "0xb6915664 Return addr 0xb6a44b7c (menelaus_web:check_and_handle_pool_info/2 + 124)", "y(0) \"34002712\"","y(1) \"34002712\"", "y(2) 20000","y(3) \"10.1.4.243\"", "y(4) {\"Administrator\",\"111111\"}", "y(5) \"default\"", "y(6) {mochiweb_request,#Port<0.2596>,'GET',\"/pools/default?waitChange=20000&etag=34002", "0xb6915684 Return addr 0xb6a44328 (menelaus_web:loop/3 + 12080)", "y(0) Catch 0xb6a44ba8 (menelaus_web:check_and_handle_pool_info/2 + 168)", "y(1) []","y(2) <0.5309.0>", "0xb6915694 Return addr 0xb6a50ed0 (mochiweb_http:headers/5 + 680)", "y(0) []","y(1) []","y(2) []", "y(3) []", "y(4) {mochiweb_request,#Port<0.2596>,'GET',\"/pools/default?waitChange=20000&etag=34002", "y(5) Catch 0xb6a44370 (menelaus_web:loop/3 + 12152)", "0xb69156b0 Return addr 0xb7543298 (proc_lib:init_p_do_apply/3 + 28)", "y(0) #Fun", "y(1) []","y(2) []", "y(3) {mochiweb_request,#Port<0.2596>,'GET',\"/pools/default?waitChange=20000&etag=34002", "0xb69156c4 Return addr 0x08229d74 ()", "y(0) Catch 0xb75432a8 (proc_lib:init_p_do_apply/3 + 44)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,4}]}, {heap_size,17711}, {total_heap_size,46368}, {links,[<0.97.0>,<0.5309.0>,#Port<0.2596>]}, {memory,186036}, {message_queue_len,0}, {reductions,3367686}, {trap_exit,false}]}, {<0.273.0>, [{registered_name,[]}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb75a43b0 (application_master:main_loop/2 + 32)", "CP: 0x00000000 (invalid)","arity = 0", "0xb6d28dac Return addr 0xb7543298 (proc_lib:init_p_do_apply/3 + 28)", "y(0) []", "y(1) {state,<0.274.0>,{appl_data,inets,[inets_sup,httpc_manager],undefined,{inets_app,", "y(2) <0.7.0>", "0xb6d28dbc Return addr 0x08229d74 ()", "y(0) Catch 0xb75432a8 (proc_lib:init_p_do_apply/3 + 44)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,0}]}, {heap_size,233}, {total_heap_size,233}, {links,[<0.7.0>,<0.274.0>]}, {memory,1396}, {message_queue_len,0}, {reductions,23}, {trap_exit,true}]}, {<0.274.0>, [{registered_name,[]}, {status,waiting}, {initial_call,{application_master,start_it,4}}, {backtrace, ["Program counter: 0xb75a56f4 (application_master:loop_it/4 + 40)", "CP: 0x00000000 (invalid)","arity = 0", "0xb686eb18 Return addr 0x08229d74 ()", "y(0) []","y(1) inets_app", "y(2) <0.275.0>","y(3) <0.273.0>"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,0}]}, {heap_size,233}, {total_heap_size,233}, {links,[<0.273.0>,<0.275.0>]}, {memory,1336}, {message_queue_len,0}, {reductions,42}, {trap_exit,true}]}, {<0.275.0>, [{registered_name,inets_sup}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb757151c (gen_server:loop/6 + 144)", "CP: 0x00000000 (invalid)","arity = 0", "0xb6d289f8 Return addr 0xb7543298 (proc_lib:init_p_do_apply/3 + 28)", "y(0) []","y(1) infinity", "y(2) supervisor", "y(3) {state,{local,inets_sup},one_for_one,[{child,<0.282.0>,tftp_sup,{tftp_sup,start_l", "y(4) inets_sup","y(5) <0.274.0>", "0xb6d28a14 Return addr 0x08229d74 ()", "y(0) Catch 0xb75432a8 (proc_lib:init_p_do_apply/3 + 44)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,4}]}, {heap_size,377}, {total_heap_size,754}, {links, [<0.276.0>,<0.281.0>,<0.282.0>,<0.277.0>,<0.274.0>]}, {memory,3540}, {message_queue_len,0}, {reductions,320}, {trap_exit,true}]}, {<0.276.0>, [{registered_name,ftp_sup}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb757151c (gen_server:loop/6 + 144)", "CP: 0x00000000 (invalid)","arity = 0", "0xb73a1ff8 Return addr 0xb7543298 (proc_lib:init_p_do_apply/3 + 28)", "y(0) []","y(1) infinity", "y(2) supervisor", "y(3) {state,{local,ftp_sup},simple_one_for_one,[{child,undefined,undefined,{ftp,start_", "y(4) ftp_sup","y(5) <0.275.0>", "0xb73a2014 Return addr 0x08229d74 ()", "y(0) Catch 0xb75432a8 (proc_lib:init_p_do_apply/3 + 44)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,0}]}, {heap_size,233}, {total_heap_size,233}, {links,[<0.275.0>]}, {memory,1376}, {message_queue_len,0}, {reductions,54}, {trap_exit,true}]}, {<0.277.0>, [{registered_name,httpc_sup}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb757151c (gen_server:loop/6 + 144)", "CP: 0x00000000 (invalid)","arity = 0", "0xb6db2508 Return addr 0xb7543298 (proc_lib:init_p_do_apply/3 + 28)", "y(0) []","y(1) infinity", "y(2) supervisor", "y(3) {state,{local,httpc_sup},one_for_one,[{child,<0.280.0>,httpc_handler_sup,{httpc_h", "y(4) httpc_sup","y(5) <0.275.0>", "0xb6db2524 Return addr 0x08229d74 ()", "y(0) Catch 0xb75432a8 (proc_lib:init_p_do_apply/3 + 44)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,1}]}, {heap_size,233}, {total_heap_size,610}, {links,[<0.278.0>,<0.280.0>,<0.275.0>]}, {memory,2924}, {message_queue_len,0}, {reductions,173}, {trap_exit,true}]}, {<0.278.0>, [{registered_name,httpc_profile_sup}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb757151c (gen_server:loop/6 + 144)", "CP: 0x00000000 (invalid)","arity = 0", "0xb6db5118 Return addr 0xb7543298 (proc_lib:init_p_do_apply/3 + 28)", "y(0) []","y(1) infinity", "y(2) supervisor", "y(3) {state,{local,httpc_profile_sup},one_for_one,[{child,<0.279.0>,httpc_manager,{htt", "y(4) httpc_profile_sup","y(5) <0.277.0>", "0xb6db5134 Return addr 0x08229d74 ()", "y(0) Catch 0xb75432a8 (proc_lib:init_p_do_apply/3 + 44)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,0}]}, {heap_size,233}, {total_heap_size,233}, {links,[<0.277.0>,<0.279.0>]}, {memory,1396}, {message_queue_len,0}, {reductions,124}, {trap_exit,true}]}, {<0.279.0>, [{registered_name,httpc_manager}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb757151c (gen_server:loop/6 + 144)", "CP: 0x00000000 (invalid)","arity = 0", "0xb6d3bd90 Return addr 0xb7543298 (proc_lib:init_p_do_apply/3 + 28)", "y(0) []","y(1) infinity", "y(2) httpc_manager", "y(3) {state,[],151595,{undefined,155692},httpc_manager_session_db,httpc_manager,{optio", "y(4) httpc_manager","y(5) <0.278.0>", "0xb6d3bdac Return addr 0x08229d74 ()", "y(0) Catch 0xb75432a8 (proc_lib:init_p_do_apply/3 + 44)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,1}]}, {heap_size,4181}, {total_heap_size,10946}, {links,[<0.278.0>]}, {memory,44228}, {message_queue_len,0}, {reductions,288}, {trap_exit,true}]}, {<0.280.0>, [{registered_name,httpc_handler_sup}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb757151c (gen_server:loop/6 + 144)", "CP: 0x00000000 (invalid)","arity = 0", "0xb7406148 Return addr 0xb7543298 (proc_lib:init_p_do_apply/3 + 28)", "y(0) []","y(1) infinity", "y(2) supervisor", "y(3) {state,{local,httpc_handler_sup},simple_one_for_one,[{child,undefined,undefined,{", "y(4) httpc_handler_sup","y(5) <0.277.0>", "0xb7406164 Return addr 0x08229d74 ()", "y(0) Catch 0xb75432a8 (proc_lib:init_p_do_apply/3 + 44)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,0}]}, {heap_size,4181}, {total_heap_size,4181}, {links,[<0.277.0>]}, {memory,17168}, {message_queue_len,0}, {reductions,120}, {trap_exit,true}]}, {<0.281.0>, [{registered_name,httpd_sup}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb757151c (gen_server:loop/6 + 144)", "CP: 0x00000000 (invalid)","arity = 0", "0xb6e12c30 Return addr 0xb7543298 (proc_lib:init_p_do_apply/3 + 28)", "y(0) []","y(1) infinity", "y(2) supervisor", "y(3) {state,{local,httpd_sup},one_for_one,[],{dict,0,16,16,8,80,48,{[],[],[],[],[],[],", "y(4) httpd_sup","y(5) <0.275.0>", "0xb6e12c4c Return addr 0x08229d74 ()", "y(0) Catch 0xb75432a8 (proc_lib:init_p_do_apply/3 + 44)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,0}]}, {heap_size,233}, {total_heap_size,233}, {links,[<0.275.0>]}, {memory,1376}, {message_queue_len,0}, {reductions,45}, {trap_exit,true}]}, {<0.282.0>, [{registered_name,tftp_sup}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb757151c (gen_server:loop/6 + 144)", "CP: 0x00000000 (invalid)","arity = 0", "0xb6db28b0 Return addr 0xb7543298 (proc_lib:init_p_do_apply/3 + 28)", "y(0) []","y(1) infinity", "y(2) supervisor", "y(3) {state,{local,tftp_sup},one_for_one,[],{dict,0,16,16,8,80,48,{[],[],[],[],[],[],[", "y(4) tftp_sup","y(5) <0.275.0>", "0xb6db28cc Return addr 0x08229d74 ()", "y(0) Catch 0xb75432a8 (proc_lib:init_p_do_apply/3 + 44)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,0}]}, {heap_size,233}, {total_heap_size,233}, {links,[<0.275.0>]}, {memory,1376}, {message_queue_len,0}, {reductions,46}, {trap_exit,true}]}, {<0.294.0>, [{registered_name,[]}, {status,waiting}, {initial_call,{inet_tcp_dist,do_accept,6}}, {backtrace, ["Program counter: 0xb61fd274 (dist_util:con_loop/9 + 72)", "CP: 0x00000000 (invalid)","arity = 0", "0xb73a01a0 Return addr 0x08229d74 ()", "y(0) []", "y(1) #Fun", "y(2) #Fun", "y(3) {tick,1149,2321,1,1}","y(4) normal", "y(5) 'ns_1@10.1.4.243'", "y(6) {net_address,{{10,1,4,244},52011},\"10.1.4.244\",tcp,inet}", "y(7) #Port<0.3490>", "y(8) 'ns_1@10.1.4.244'","y(9) <0.61.0>"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,6}]}, {heap_size,987}, {total_heap_size,1364}, {links,[<0.61.0>,#Port<0.3490>]}, {memory,5860}, {message_queue_len,0}, {reductions,1068}, {trap_exit,false}]}, {<0.375.0>, [{registered_name,[]}, {status,waiting}, {initial_call,{erlang,apply,2}}, {backtrace, ["Program counter: 0xb665237c (ns_rebalancer:wait_for_mover/1 + 20)", "CP: 0x00000000 (invalid)","arity = 0", "0xb656b538 Return addr 0xb66520f8 (ns_rebalancer:perform_moves/4 + 152)", "y(0) <0.409.0>", "0xb656b540 Return addr 0xb6650f68 (ns_rebalancer:rebalance/5 + 392)", "y(0) []","y(1) []", "y(2) [{511,'ns_1@10.1.4.243','ns_1@10.1.4.244'},{510,'ns_1@10.1.4.243','ns_1@10.1.4.24", "y(3) [['ns_1@10.1.4.243',undefined],['ns_1@10.1.4.243',undefined],['ns_1@10.1.4.243',u", "0xb656b554 Return addr 0xb754a35c (lists:foreach/2 + 64)", "y(0) []","y(1) []", "y(2) Catch 0xb6651118 (ns_rebalancer:rebalance/5 + 824)", "y(3) []","y(4) []", "y(5) ['ns_1@10.1.4.243','ns_1@10.1.4.244']", "y(6) \"default\"", "0xb656b574 Return addr 0xb6650cf4 (ns_rebalancer:rebalance/3 + 352)", "y(0) #Fun", "y(1) []", "0xb656b580 Return addr 0x08229d74 ()", "y(0) []","y(1) []", "y(2) Catch 0xb6650d04 (ns_rebalancer:rebalance/3 + 368)", "y(3) []","y(4) []","y(5) []", "y(6) []","y(7) []"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,134}]}, {heap_size,6765}, {total_heap_size,53133}, {links,[<0.113.0>,<0.409.0>]}, {memory,212936}, {message_queue_len,0}, {reductions,1416059}, {trap_exit,true}]}, {<0.409.0>, [{registered_name,[]}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb757151c (gen_server:loop/6 + 144)", "CP: 0x00000000 (invalid)","arity = 0", "0xb6564bc8 Return addr 0xb7543298 (proc_lib:init_p_do_apply/3 + 28)", "y(0) []","y(1) infinity", "y(2) ns_vbucket_mover", "y(3) {state,\"default\",{dict,2,16,16,8,80,48,{[],[],[],[],[],[],[],[],[],[],[],[],[],[]", "y(4) <0.409.0>","y(5) <0.375.0>", "0xb6564be4 Return addr 0x08229d74 ()", "y(0) Catch 0xb75432a8 (proc_lib:init_p_do_apply/3 + 44)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,26}]}, {heap_size,6765}, {total_heap_size,13530}, {links,[<0.375.0>,<0.410.0>]}, {memory,54584}, {message_queue_len,0}, {reductions,25790}, {trap_exit,false}]}, {<0.410.0>, [{registered_name,[]}, {status,waiting}, {initial_call,{erlang,apply,2}}, {backtrace, ["Program counter: 0xb6204048 (ns_vbucket_mover:wait_for_mover/5 + 44)", "CP: 0x00000000 (invalid)","arity = 0", "0xb69b9400 Return addr 0xb62046e0 (ns_vbucket_mover:'-spawn_workers/1-fun-0-'/5 + 132)", "y(0) 2","y(1) 'ns_1@10.1.4.244'", "y(2) 'ns_1@10.1.4.243'","y(3) 0", "y(4) \"default\"", "0xb69b9418 Return addr 0x08229d74 ()", "y(0) 'ns_1@10.1.4.243'","y(1) <0.409.0>"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,2}]}, {heap_size,17711}, {total_heap_size,46368}, {links,[<0.409.0>,<0.411.0>]}, {memory,185876}, {message_queue_len,0}, {reductions,495}, {trap_exit,true}]}, {<0.411.0>, [{registered_name,[]}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb757151c (gen_server:loop/6 + 144)", "CP: 0x00000000 (invalid)","arity = 0", "0xb6e124e0 Return addr 0xb7543298 (proc_lib:init_p_do_apply/3 + 28)", "y(0) []","y(1) infinity", "y(2) ns_port_server", "y(3) {state,#Port<0.3510>,vbucketmigrator,{[\"Bucket 0 moved to the next server\",\"Start", "y(4) <0.411.0>","y(5) <0.410.0>", "0xb6e124fc Return addr 0x08229d74 ()", "y(0) Catch 0xb75432a8 (proc_lib:init_p_do_apply/3 + 44)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,2}]}, {heap_size,987}, {total_heap_size,1974}, {links,[<0.410.0>,#Port<0.3510>]}, {memory,8360}, {message_queue_len,0}, {reductions,450}, {trap_exit,true}]}, {<0.567.0>, [{registered_name,[]}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb6a44e3c (menelaus_web:handle_pool_info_wait/6 + 140)", "CP: 0x00000000 (invalid)","arity = 0", "0xb6866dec Return addr 0xb6a44b7c (menelaus_web:check_and_handle_pool_info/2 + 124)", "y(0) \"34002712\"","y(1) \"34002712\"", "y(2) 3000","y(3) \"10.1.4.243\"", "y(4) {\"Administrator\",\"111111\"}", "y(5) \"default\"", "y(6) {mochiweb_request,#Port<0.3523>,'GET',\"/pools/default?waitChange=3000&etag=340027", "0xb6866e0c Return addr 0xb6a44328 (menelaus_web:loop/3 + 12080)", "y(0) Catch 0xb6a44ba8 (menelaus_web:check_and_handle_pool_info/2 + 168)", "y(1) []","y(2) <0.5347.0>", "0xb6866e1c Return addr 0xb6a50ed0 (mochiweb_http:headers/5 + 680)", "y(0) []","y(1) []","y(2) []", "y(3) []", "y(4) {mochiweb_request,#Port<0.3523>,'GET',\"/pools/default?waitChange=3000&etag=340027", "y(5) Catch 0xb6a44370 (menelaus_web:loop/3 + 12152)", "0xb6866e38 Return addr 0xb7543298 (proc_lib:init_p_do_apply/3 + 28)", "y(0) #Fun", "y(1) []","y(2) []", "y(3) {mochiweb_request,#Port<0.3523>,'GET',\"/pools/default?waitChange=3000&etag=340027", "0xb6866e4c Return addr 0x08229d74 ()", "y(0) Catch 0xb75432a8 (proc_lib:init_p_do_apply/3 + 44)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,4}]}, {heap_size,17711}, {total_heap_size,46368}, {links,[<0.97.0>,<0.5347.0>,#Port<0.3523>]}, {memory,186036}, {message_queue_len,0}, {reductions,1241010}, {trap_exit,false}]}, {<0.4993.0>, [{registered_name,[]}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb7513438 (prim_inet:recv0/3 + 112)", "CP: 0x00000000 (invalid)","arity = 0", "0xb62ee440 Return addr 0xb6a50a64 (mochiweb_http:request/2 + 60)", "y(0) 47948","y(1) #Port<0.7544>", "0xb62ee44c Return addr 0xb7543298 (proc_lib:init_p_do_apply/3 + 28)", "y(0) #Fun", "y(1) #Port<0.7544>", "0xb62ee458 Return addr 0x08229d74 ()", "y(0) Catch 0xb75432a8 (proc_lib:init_p_do_apply/3 + 44)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,77}]}, {heap_size,46368}, {total_heap_size,364179}, {links,[<0.97.0>,<0.108.0>,#Port<0.7544>]}, {memory,1457216}, {message_queue_len,1}, {reductions,526443}, {trap_exit,false}]}, {<0.5104.0>, [{registered_name,[]}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb7513438 (prim_inet:recv0/3 + 112)", "CP: 0x00000000 (invalid)","arity = 0", "0xb6969604 Return addr 0xb6a50a64 (mochiweb_http:request/2 + 60)", "y(0) 47965","y(1) #Port<0.7547>", "0xb6969610 Return addr 0xb7543298 (proc_lib:init_p_do_apply/3 + 28)", "y(0) #Fun", "y(1) #Port<0.7547>", "0xb696961c Return addr 0x08229d74 ()", "y(0) Catch 0xb75432a8 (proc_lib:init_p_do_apply/3 + 44)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,1}]}, {heap_size,28657}, {total_heap_size,75025}, {links,[<0.97.0>,#Port<0.7547>]}, {memory,300564}, {message_queue_len,0}, {reductions,472166}, {trap_exit,false}]}, {<0.5185.0>, [{registered_name,[]}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb7571a0c (gen_server:do_multi_call/4 + 384)", "CP: 0x00000000 (invalid)","arity = 0", "0xb6549b08 Return addr 0xb75b9760 (rpc:do_multicall/5 + 160)", "y(0) []","y(1) #Ref<0.0.0.103228>", "y(2) <0.5348.0>","y(3) #Ref<0.0.0.103227>", "0xb6549b1c Return addr 0xb6d17d9c (diag_handler:diag_multicall/3 + 116)", "y(0) []","y(1) []", "0xb6549b28 Return addr 0xb6d183b4 (diag_handler:handle_diag/1 + 192)", "y(0) ['ns_1@10.1.4.243','ns_1@10.1.4.244']", "y(1) []","y(2) []","y(3) []", "0xb6549b3c Return addr 0xb6a440b4 (menelaus_web:loop/3 + 11452)", "y(0) []", "y(1) [[\"2011\",45,[\"0\",\"5\"],45,[\"0\",\"7\"],32,\"17\",58,[\"0\",\"5\"],58,[\"0\",\"2\"],46,\"949\"],32", "y(2) [{\"default\",[{num_replicas,1},{ram_quota,839909376},{auth_type,sasl},{sasl_passwo", "y(3) {mochiweb_request,#Port<0.7549>,'GET',\"/diag\",{1,1},{9,{\"host\",{'Host',\"10.1.4.24", "0xb6549b50 Return addr 0xb6a50ed0 (mochiweb_http:headers/5 + 680)", "y(0) []","y(1) []","y(2) []", "y(3) []", "y(4) {mochiweb_request,#Port<0.7549>,'GET',\"/diag\",{1,1},{9,{\"host\",{'Host',\"10.1.4.24", "y(5) Catch 0xb6a44370 (menelaus_web:loop/3 + 12152)", "0xb6549b6c Return addr 0xb7543298 (proc_lib:init_p_do_apply/3 + 28)", "y(0) #Fun", "y(1) []","y(2) []", "y(3) {mochiweb_request,#Port<0.7549>,'GET',\"/diag\",{1,1},{9,{\"host\",{'Host',\"10.1.4.24", "0xb6549b80 Return addr 0x08229d74 ()", "y(0) Catch 0xb75432a8 (proc_lib:init_p_do_apply/3 + 44)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,8}]}, {heap_size,46368}, {total_heap_size,364179}, {links,[<0.97.0>,<0.108.0>,#Port<0.7549>]}, {memory,1457312}, {message_queue_len,2}, {reductions,947570}, {trap_exit,false}]}, {<0.5219.0>, [{registered_name,[]}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb6a44e3c (menelaus_web:handle_pool_info_wait/6 + 140)", "CP: 0x00000000 (invalid)","arity = 0", "0xb6d5eb3c Return addr 0xb6a44b7c (menelaus_web:check_and_handle_pool_info/2 + 124)", "y(0) \"34002712\"","y(1) \"34002712\"", "y(2) 3000","y(3) \"10.1.4.243\"", "y(4) {\"Administrator\",\"111111\"}", "y(5) \"default\"", "y(6) {mochiweb_request,#Port<0.7552>,'GET',\"/pools/default?waitChange=3000&etag=340027", "0xb6d5eb5c Return addr 0xb6a44328 (menelaus_web:loop/3 + 12080)", "y(0) Catch 0xb6a44ba8 (menelaus_web:check_and_handle_pool_info/2 + 168)", "y(1) []","y(2) <0.5343.0>", "0xb6d5eb6c Return addr 0xb6a50ed0 (mochiweb_http:headers/5 + 680)", "y(0) []","y(1) []","y(2) []", "y(3) []", "y(4) {mochiweb_request,#Port<0.7552>,'GET',\"/pools/default?waitChange=3000&etag=340027", "y(5) Catch 0xb6a44370 (menelaus_web:loop/3 + 12152)", "0xb6d5eb88 Return addr 0xb7543298 (proc_lib:init_p_do_apply/3 + 28)", "y(0) #Fun", "y(1) []","y(2) []", "y(3) {mochiweb_request,#Port<0.7552>,'GET',\"/pools/default?waitChange=3000&etag=340027", "0xb6d5eb9c Return addr 0x08229d74 ()", "y(0) Catch 0xb75432a8 (proc_lib:init_p_do_apply/3 + 44)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,4}]}, {heap_size,17711}, {total_heap_size,46368}, {links,[<0.97.0>,<0.5343.0>,#Port<0.7552>]}, {memory,186036}, {message_queue_len,0}, {reductions,167547}, {trap_exit,false}]}, {<0.5290.0>, [{registered_name,[]}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb7512a08 (prim_inet:accept0/2 + 92)", "CP: 0x00000000 (invalid)","arity = 0", "0xb6d7e908 Return addr 0xb6cdd81c (inet_tcp:accept/1 + 20)", "y(0) 39028","y(1) #Port<0.1400>", "0xb6d7e914 Return addr 0xb6a58234 (mochiweb_socket_server:acceptor_loop/1 + 80)", "y(0) []", "0xb6d7e91c Return addr 0xb7543298 (proc_lib:init_p_do_apply/3 + 28)", "y(0) #Fun", "y(1) <0.97.0>", "y(2) Catch 0xb6a58234 (mochiweb_socket_server:acceptor_loop/1 + 80)", "0xb6d7e92c Return addr 0x08229d74 ()", "y(0) Catch 0xb75432a8 (proc_lib:init_p_do_apply/3 + 44)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,1}]}, {heap_size,377}, {total_heap_size,754}, {links,[<0.97.0>]}, {memory,3500}, {message_queue_len,0}, {reductions,39}, {trap_exit,false}]}, {<0.5309.0>, [{registered_name,[]}, {status,waiting}, {initial_call,{erlang,apply,2}}, {backtrace, ["Program counter: 0xb6d18b48 (diag_handler:'-arm_timeout/2-fun-0-'/3 + 32)", "CP: 0x00000000 (invalid)","arity = 0", "0xb73d5ae4 Return addr 0x08229d74 ()", "y(0) <0.204.0>", "y(1) #Fun", "y(2) 23000"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,0}]}, {heap_size,233}, {total_heap_size,233}, {links,[<0.204.0>]}, {memory,1316}, {message_queue_len,0}, {reductions,2}, {trap_exit,false}]}, {<0.5343.0>, [{registered_name,[]}, {status,waiting}, {initial_call,{erlang,apply,2}}, {backtrace, ["Program counter: 0xb6d18b48 (diag_handler:'-arm_timeout/2-fun-0-'/3 + 32)", "CP: 0x00000000 (invalid)","arity = 0", "0xb73a18bc Return addr 0x08229d74 ()", "y(0) <0.5219.0>", "y(1) #Fun", "y(2) 23000"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,0}]}, {heap_size,233}, {total_heap_size,233}, {links,[<0.5219.0>]}, {memory,1316}, {message_queue_len,0}, {reductions,2}, {trap_exit,false}]}, {<0.5347.0>, [{registered_name,[]}, {status,waiting}, {initial_call,{erlang,apply,2}}, {backtrace, ["Program counter: 0xb6d18b48 (diag_handler:'-arm_timeout/2-fun-0-'/3 + 32)", "CP: 0x00000000 (invalid)","arity = 0", "0xb73aa674 Return addr 0x08229d74 ()", "y(0) <0.567.0>", "y(1) #Fun", "y(2) 23000"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,0}]}, {heap_size,233}, {total_heap_size,233}, {links,[<0.567.0>]}, {memory,1316}, {message_queue_len,0}, {reductions,2}, {trap_exit,false}]}, {<0.5348.0>, [{registered_name,[]}, {status,waiting}, {initial_call,{erlang,apply,2}}, {backtrace, ["Program counter: 0xb7571d1c (gen_server:rec_nodes/7 + 120)", "CP: 0x00000000 (invalid)","arity = 0", "0xb7399b6c Return addr 0xb7574db8 (gen_server:'-do_multi_call/4-fun-0-'/6 + 356)", "y(0) #Ref<0.0.0.103232>","y(1) 2000", "y(2) []","y(3) []","y(4) rex", "y(5) #Ref<0.0.0.103227>", "y(6) [{'ns_1@10.1.4.243',#Ref<0.0.0.103230>}]", "y(7) #Ref<0.0.0.103231>", "y(8) 'ns_1@10.1.4.244'", "0xb7399b94 Return addr 0x08229d74 ()", "y(0) #Ref<0.0.0.103227>","y(1) []", "y(2) []","y(3) []","y(4) []"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,0}]}, {heap_size,233}, {total_heap_size,233}, {links,[]}, {memory,1456}, {message_queue_len,0}, {reductions,35}, {trap_exit,true}]}, {<0.5349.0>, [{registered_name,[]}, {status,running}, {initial_call,{erlang,apply,2}}, {backtrace, ["Program counter: 0xb74a2da4 (unknown function)", "CP: 0xb6d17af4 (diag_handler:grab_process_info/1 + 32)", "0xb5943f48 Return addr 0xb6d18da4 (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 44)", "y(0) []", "0xb5943f50 Return addr 0xb6d18dc4 (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) []","y(1) <0.5349.0>", "0xb5943f5c Return addr 0xb6d18dc4 (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,[]},{status,waiting},{initial_call,{erlang,apply,2}},{backtrace", "y(1) <0.5348.0>", "0xb5943f68 Return addr 0xb6d18dc4 (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,[]},{status,waiting},{initial_call,{erlang,apply,2}},{backtrace", "y(1) <0.5347.0>", "0xb5943f74 Return addr 0xb6d18dc4 (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,[]},{status,waiting},{initial_call,{erlang,apply,2}},{backtrace", "y(1) <0.5343.0>", "0xb5943f80 Return addr 0xb6d18dc4 (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,[]},{status,waiting},{initial_call,{erlang,apply,2}},{backtrace", "y(1) <0.5309.0>", "0xb5943f8c Return addr 0xb6d18dc4 (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,[]},{status,waiting},{initial_call,{proc_lib,init_p,5}},{backtr", "y(1) <0.5290.0>", "0xb5943f98 Return addr 0xb6d18dc4 (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,[]},{status,waiting},{initial_call,{proc_lib,init_p,5}},{backtr", "y(1) <0.5219.0>", "0xb5943fa4 Return addr 0xb6d18dc4 (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,[]},{status,waiting},{initial_call,{proc_lib,init_p,5}},{backtr", "y(1) <0.5185.0>", "0xb5943fb0 Return addr 0xb6d18dc4 (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,[]},{status,waiting},{initial_call,{proc_lib,init_p,5}},{backtr", "y(1) <0.5104.0>", "0xb5943fbc Return addr 0xb6d18dc4 (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,[]},{status,waiting},{initial_call,{proc_lib,init_p,5}},{backtr", "y(1) <0.4993.0>", "0xb5943fc8 Return addr 0xb6d18dc4 (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,[]},{status,waiting},{initial_call,{proc_lib,init_p,5}},{backtr", "y(1) <0.567.0>", "0xb5943fd4 Return addr 0xb6d18dc4 (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,[]},{status,waiting},{initial_call,{proc_lib,init_p,5}},{backtr", "y(1) <0.411.0>", "0xb5943fe0 Return addr 0xb6d18dc4 (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,[]},{status,waiting},{initial_call,{erlang,apply,2}},{backtrace", "y(1) <0.410.0>", "0xb5943fec Return addr 0xb6d18dc4 (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,[]},{status,waiting},{initial_call,{proc_lib,init_p,5}},{backtr", "y(1) <0.409.0>", "0xb5943ff8 Return addr 0xb6d18dc4 (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,[]},{status,waiting},{initial_call,{erlang,apply,2}},{backtrace", "y(1) <0.375.0>", "0xb5944004 Return addr 0xb6d18dc4 (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,[]},{status,waiting},{initial_call,{inet_tcp_dist,do_accept,6}}", "y(1) <0.294.0>", "0xb5944010 Return addr 0xb6d18dc4 (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,tftp_sup},{status,waiting},{initial_call,{proc_lib,init_p,5}},{", "y(1) <0.282.0>", "0xb594401c Return addr 0xb6d18dc4 (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,httpd_sup},{status,waiting},{initial_call,{proc_lib,init_p,5}},", "y(1) <0.281.0>", "0xb5944028 Return addr 0xb6d18dc4 (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,httpc_handler_sup},{status,waiting},{initial_call,{proc_lib,ini", "y(1) <0.280.0>", "0xb5944034 Return addr 0xb6d18dc4 (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,httpc_manager},{status,waiting},{initial_call,{proc_lib,init_p,", "y(1) <0.279.0>", "0xb5944040 Return addr 0xb6d18dc4 (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,httpc_profile_sup},{status,waiting},{initial_call,{proc_lib,ini", "y(1) <0.278.0>", "0xb594404c Return addr 0xb6d18dc4 (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,httpc_sup},{status,waiting},{initial_call,{proc_lib,init_p,5}},", "y(1) <0.277.0>", "0xb5944058 Return addr 0xb6d18dc4 (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,ftp_sup},{status,waiting},{initial_call,{proc_lib,init_p,5}},{b", "y(1) <0.276.0>", "0xb5944064 Return addr 0xb6d18dc4 (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,inets_sup},{status,waiting},{initial_call,{proc_lib,init_p,5}},", "y(1) <0.275.0>", "0xb5944070 Return addr 0xb6d18dc4 (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,[]},{status,waiting},{initial_call,{application_master,start_it", "y(1) <0.274.0>", "0xb594407c Return addr 0xb6d18dc4 (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,[]},{status,waiting},{initial_call,{proc_lib,init_p,5}},{backtr", "y(1) <0.273.0>", "0xb5944088 Return addr 0xb6d18dc4 (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,[]},{status,waiting},{initial_call,{proc_lib,init_p,5}},{backtr", "y(1) <0.204.0>", "0xb5944094 Return addr 0xb6d18dc4 (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,[]},{status,waiting},{initial_call,{proc_lib,init_p,5}},{backtr", "y(1) <0.177.0>", "0xb59440a0 Return addr 0xb6d18dc4 (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,ns_moxi_sup},{status,waiting},{initial_call,{proc_lib,init_p,5}", "y(1) <0.174.0>", "0xb59440ac Return addr 0xb6d18dc4 (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,'stats_reader-default'},{status,waiting},{initial_call,{proc_li", "y(1) <0.171.0>", "0xb59440b8 Return addr 0xb6d18dc4 (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,'stats_archiver-default'},{status,waiting},{initial_call,{proc_", "y(1) <0.168.0>", "0xb59440c4 Return addr 0xb6d18dc4 (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,[]},{status,waiting},{initial_call,{proc_lib,init_p,5}},{backtr", "y(1) <0.167.0>", "0xb59440d0 Return addr 0xb6d18dc4 (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,ns_bad_bucket_sup},{status,waiting},{initial_call,{proc_lib,ini", "y(1) <0.162.0>", "0xb59440dc Return addr 0xb6d18dc4 (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,ns_bad_bucket_worker},{status,waiting},{initial_call,{proc_lib,", "y(1) <0.161.0>", "0xb59440e8 Return addr 0xb6d18dc4 (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,mnesia_late_loader},{status,waiting},{initial_call,{proc_lib,in", "y(1) <0.157.0>", "0xb59440f4 Return addr 0xb6d18dc4 (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,mnesia_controller},{status,waiting},{initial_call,{proc_lib,ini", "y(1) <0.156.0>", "0xb5944100 Return addr 0xb6d18dc4 (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,mnesia_snmp_sup},{status,waiting},{initial_call,{proc_lib,init_", "y(1) <0.155.0>", "0xb594410c Return addr 0xb6d18dc4 (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,mnesia_checkpoint_sup},{status,waiting},{initial_call,{proc_lib", "y(1) <0.154.0>", "0xb5944118 Return addr 0xb6d18dc4 (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,[]},{status,waiting},{initial_call,{proc_lib,init_p,5}},{backtr", "y(1) <0.152.0>", "0xb5944124 Return addr 0xb6d18dc4 (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,disk_log_server},{status,waiting},{initial_call,{proc_lib,init_", "y(1) <0.133.0>", "0xb5944130 Return addr 0xb6d18dc4 (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,disk_log_sup},{status,waiting},{initial_call,{proc_lib,init_p,5", "y(1) <0.132.0>", "0xb594413c Return addr 0xb6d18dc4 (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,dets},{status,waiting},{initial_call,{proc_lib,init_p,5}},{back", "y(1) <0.128.0>", "0xb5944148 Return addr 0xb6d18dc4 (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,dets_sup},{status,waiting},{initial_call,{proc_lib,init_p,5}},{", "y(1) <0.127.0>", "0xb5944154 Return addr 0xb6d18dc4 (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,mnesia_tm},{status,waiting},{initial_call,{proc_lib,init_p,5}},", "y(1) <0.126.0>", "0xb5944160 Return addr 0xb6d18dc4 (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,mnesia_recover},{status,waiting},{initial_call,{proc_lib,init_p", "y(1) <0.125.0>", "0xb594416c Return addr 0xb6d18dc4 (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,mnesia_locker},{status,waiting},{initial_call,{proc_lib,init_p,", "y(1) <0.124.0>", "0xb5944178 Return addr 0xb6d18dc4 (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,mnesia_subscr},{status,waiting},{initial_call,{proc_lib,init_p,", "y(1) <0.123.0>", "0xb5944184 Return addr 0xb6d18dc4 (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,mnesia_monitor},{status,waiting},{initial_call,{proc_lib,init_p", "y(1) <0.122.0>", "0xb5944190 Return addr 0xb6d18dc4 (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,mnesia_kernel_sup},{status,waiting},{initial_call,{proc_lib,ini", "y(1) <0.121.0>", "0xb594419c Return addr 0xb6d18dc4 (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,mnesia_event},{status,waiting},{initial_call,{proc_lib,init_p,5", "y(1) <0.120.0>", "0xb59441a8 Return addr 0xb6d18dc4 (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,mnesia_sup},{status,waiting},{initial_call,{proc_lib,init_p,5}}", "y(1) <0.119.0>", "0xb59441b4 Return addr 0xb6d18dc4 (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,[]},{status,waiting},{initial_call,{application_master,start_it", "y(1) <0.118.0>", "0xb59441c0 Return addr 0xb6d18dc4 (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,[]},{status,waiting},{initial_call,{proc_lib,init_p,5}},{backtr", "y(1) <0.117.0>", "0xb59441cc Return addr 0xb6d18dc4 (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,ns_mnesia},{status,waiting},{initial_call,{proc_lib,init_p,5}},", "y(1) <0.115.0>", "0xb59441d8 Return addr 0xb6d18dc4 (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,[]},{status,waiting},{initial_call,{proc_lib,init_p,5}},{backtr", "y(1) <0.113.0>", "0xb59441e4 Return addr 0xb6d18dc4 (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,'ns_memcached-default'},{status,waiting},{initial_call,{proc_li", "y(1) <0.112.0>", "0xb59441f0 Return addr 0xb6d18dc4 (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,'ns_vbm_sup-default'},{status,waiting},{initial_call,{proc_lib,", "y(1) <0.111.0>", "0xb59441fc Return addr 0xb6d18dc4 (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,ns_good_bucket_sup},{status,waiting},{initial_call,{proc_lib,in", "y(1) <0.110.0>", "0xb5944208 Return addr 0xb6d18dc4 (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,ns_good_bucket_worker},{status,waiting},{initial_call,{proc_lib", "y(1) <0.109.0>", "0xb5944214 Return addr 0xb6d18dc4 (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,ns_stats_event},{status,waiting},{initial_call,{proc_lib,init_p", "y(1) <0.108.0>", "0xb5944220 Return addr 0xb6d18dc4 (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,ns_tick_event},{status,waiting},{initial_call,{proc_lib,init_p,", "y(1) <0.107.0>", "0xb594422c Return addr 0xb6d18dc4 (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,[]},{status,waiting},{initial_call,{proc_lib,init_p,5}},{backtr", "y(1) <0.106.0>", "0xb5944238 Return addr 0xb6d18dc4 (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,[]},{status,waiting},{initial_call,{proc_lib,init_p,5}},{backtr", "y(1) <0.105.0>", "0xb5944244 Return addr 0xb6d18dc4 (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,[]},{status,waiting},{initial_call,{proc_lib,init_p,5}},{backtr", "y(1) <0.104.0>", "0xb5944250 Return addr 0xb6d18dc4 (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,[]},{status,waiting},{initial_call,{proc_lib,init_p,5}},{backtr", "y(1) <0.103.0>", "0xb594425c Return addr 0xb6d18dc4 (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,ns_port_sup},{status,waiting},{initial_call,{proc_lib,init_p,5}", "y(1) <0.101.0>", "0xb5944268 Return addr 0xb6d18dc4 (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,hot_keys_keeper},{status,waiting},{initial_call,{proc_lib,init_", "y(1) <0.100.0>", "0xb5944274 Return addr 0xb6d18dc4 (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,[]},{status,waiting},{initial_call,{proc_lib,init_p,5}},{backtr", "y(1) <0.98.0>", "0xb5944280 Return addr 0xb6d18dc4 (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,menelaus_web},{status,waiting},{initial_call,{proc_lib,init_p,5", "y(1) <0.97.0>", "0xb594428c Return addr 0xb6d18dc4 (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,menelaus_sup},{status,waiting},{initial_call,{proc_lib,init_p,5", "y(1) <0.96.0>", "0xb5944298 Return addr 0xb6d18dc4 (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,ns_doctor},{status,waiting},{initial_call,{proc_lib,init_p,5}},", "y(1) <0.94.0>", "0xb59442a4 Return addr 0xb6d18dc4 (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,ns_heart},{status,waiting},{initial_call,{proc_lib,init_p,5}},{", "y(1) <0.92.0>", "0xb59442b0 Return addr 0xb6d18dc4 (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,ns_config_rep},{status,waiting},{initial_call,{proc_lib,init_p,", "y(1) <0.86.0>", "0xb59442bc Return addr 0xb6d18dc4 (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,ns_node_disco},{status,waiting},{initial_call,{proc_lib,init_p,", "y(1) <0.80.0>", "0xb59442c8 Return addr 0xb6d18dc4 (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,ns_node_disco_events},{status,waiting},{initial_call,{proc_lib,", "y(1) <0.79.0>", "0xb59442d4 Return addr 0xb6d18dc4 (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,ns_node_disco_sup},{status,waiting},{initial_call,{proc_lib,ini", "y(1) <0.78.0>", "0xb59442e0 Return addr 0xb6d18dc4 (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,ns_mail},{status,waiting},{initial_call,{proc_lib,init_p,5}},{b", "y(1) <0.76.0>", "0xb59442ec Return addr 0xb6d18dc4 (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,ns_mail_sup},{status,waiting},{initial_call,{proc_lib,init_p,5}", "y(1) <0.75.0>", "0xb59442f8 Return addr 0xb6d18dc4 (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,ns_log_events},{status,waiting},{initial_call,{proc_lib,init_p,", "y(1) <0.74.0>", "0xb5944304 Return addr 0xb6d18dc4 (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,ns_log},{status,waiting},{initial_call,{proc_lib,init_p,5}},{ba", "y(1) <0.73.0>", "0xb5944310 Return addr 0xb6d18dc4 (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,ns_server_sup},{status,waiting},{initial_call,{proc_lib,init_p,", "y(1) <0.72.0>", "0xb594431c Return addr 0xb6d18dc4 (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,ns_config},{status,waiting},{initial_call,{proc_lib,init_p,5}},", "y(1) <0.67.0>", "0xb5944328 Return addr 0xb6d18dc4 (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,ns_config_events},{status,waiting},{initial_call,{proc_lib,init", "y(1) <0.66.0>", "0xb5944334 Return addr 0xb6d18dc4 (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,ns_config_sup},{status,waiting},{initial_call,{proc_lib,init_p,", "y(1) <0.65.0>", "0xb5944340 Return addr 0xb6d18dc4 (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,ns_cluster},{status,waiting},{initial_call,{proc_lib,init_p,5}}", "y(1) <0.64.0>", "0xb594434c Return addr 0xb6d18dc4 (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,[]},{status,waiting},{initial_call,{net_kernel,ticker,2}},{back", "y(1) <0.63.0>", "0xb5944358 Return addr 0xb6d18dc4 (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,[]},{status,waiting},{initial_call,{inet_tcp_dist,accept_loop,2", "y(1) <0.62.0>", "0xb5944364 Return addr 0xb6d18dc4 (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,net_kernel},{status,waiting},{initial_call,{proc_lib,init_p,5}}", "y(1) <0.61.0>", "0xb5944370 Return addr 0xb6d18dc4 (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,auth},{status,waiting},{initial_call,{proc_lib,init_p,5}},{back", "y(1) <0.60.0>", "0xb594437c Return addr 0xb6d18dc4 (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,erl_epmd},{status,waiting},{initial_call,{proc_lib,init_p,5}},{", "y(1) <0.59.0>", "0xb5944388 Return addr 0xb6d18dc4 (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,net_sup},{status,waiting},{initial_call,{proc_lib,init_p,5}},{b", "y(1) <0.58.0>", "0xb5944394 Return addr 0xb6d18dc4 (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,inet_gethost_native},{status,waiting},{initial_call,{inet_getho", "y(1) <0.57.0>", "0xb59443a0 Return addr 0xb6d18dc4 (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,inet_gethost_native_sup},{status,waiting},{initial_call,{proc_l", "y(1) <0.56.0>", "0xb59443ac Return addr 0xb6d18dc4 (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,dist_manager},{status,waiting},{initial_call,{proc_lib,init_p,5", "y(1) <0.55.0>", "0xb59443b8 Return addr 0xb6d18dc4 (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,ns_server_cluster_sup},{status,waiting},{initial_call,{proc_lib", "y(1) <0.54.0>", "0xb59443c4 Return addr 0xb6d18dc4 (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,timer_server},{status,waiting},{initial_call,{proc_lib,init_p,5", "y(1) <0.53.0>", "0xb59443d0 Return addr 0xb6d18dc4 (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,[]},{status,waiting},{initial_call,{application_master,start_it", "y(1) <0.52.0>", "0xb59443dc Return addr 0xb6d18dc4 (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,[]},{status,waiting},{initial_call,{proc_lib,init_p,5}},{backtr", "y(1) <0.51.0>", "0xb59443e8 Return addr 0xb6d18dc4 (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,[]},{status,waiting},{initial_call,{erlang,apply,2}},{backtrace", "y(1) <0.49.0>", "0xb59443f4 Return addr 0xb6d18dc4 (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,[]},{status,waiting},{initial_call,{erlang,apply,2}},{backtrace", "y(1) <0.48.0>", "0xb5944400 Return addr 0xb6d18dc4 (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,cpu_sup},{status,waiting},{initial_call,{proc_lib,init_p,5}},{b", "y(1) <0.47.0>", "0xb594440c Return addr 0xb6d18dc4 (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,[]},{status,waiting},{initial_call,{erlang,apply,2}},{backtrace", "y(1) <0.46.0>", "0xb5944418 Return addr 0xb6d18dc4 (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,memsup},{status,waiting},{initial_call,{proc_lib,init_p,5}},{ba", "y(1) <0.45.0>", "0xb5944424 Return addr 0xb6d18dc4 (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,disksup},{status,waiting},{initial_call,{proc_lib,init_p,5}},{b", "y(1) <0.44.0>", "0xb5944430 Return addr 0xb6d18dc4 (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,os_mon_sup},{status,waiting},{initial_call,{proc_lib,init_p,5}}", "y(1) <0.43.0>", "0xb594443c Return addr 0xb6d18dc4 (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,[]},{status,waiting},{initial_call,{application_master,start_it", "y(1) <0.42.0>", "0xb5944448 Return addr 0xb6d18dc4 (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,[]},{status,waiting},{initial_call,{proc_lib,init_p,5}},{backtr", "y(1) <0.41.0>", "0xb5944454 Return addr 0xb6d18dc4 (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,release_handler},{status,waiting},{initial_call,{proc_lib,init_", "y(1) <0.38.0>", "0xb5944460 Return addr 0xb6d18dc4 (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,overload},{status,waiting},{initial_call,{proc_lib,init_p,5}},{", "y(1) <0.37.0>", "0xb594446c Return addr 0xb6d18dc4 (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,alarm_handler},{status,waiting},{initial_call,{proc_lib,init_p,", "y(1) <0.36.0>", "0xb5944478 Return addr 0xb6d18dc4 (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,sasl_safe_sup},{status,waiting},{initial_call,{proc_lib,init_p,", "y(1) <0.35.0>", "0xb5944484 Return addr 0xb6d18dc4 (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,sasl_sup},{status,waiting},{initial_call,{proc_lib,init_p,5}},{", "y(1) <0.34.0>", "0xb5944490 Return addr 0xb6d18dc4 (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,[]},{status,waiting},{initial_call,{application_master,start_it", "y(1) <0.33.0>", "0xb594449c Return addr 0xb6d18dc4 (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,[]},{status,waiting},{initial_call,{proc_lib,init_p,5}},{backtr", "y(1) <0.32.0>", "0xb59444a8 Return addr 0xb6d18dc4 (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,kernel_safe_sup},{status,waiting},{initial_call,{proc_lib,init_", "y(1) <0.27.0>", "0xb59444b4 Return addr 0xb6d18dc4 (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,[]},{status,waiting},{initial_call,{proc_lib,init_p,5}},{backtr", "y(1) <0.26.0>", "0xb59444c0 Return addr 0xb6d18dc4 (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,user},{status,waiting},{initial_call,{user,server,2}},{backtrac", "y(1) <0.25.0>", "0xb59444cc Return addr 0xb6d18dc4 (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,[]},{status,waiting},{initial_call,{proc_lib,init_p,5}},{backtr", "y(1) <0.24.0>", "0xb59444d8 Return addr 0xb6d18dc4 (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,standard_error},{status,waiting},{initial_call,{standard_error,", "y(1) <0.23.0>", "0xb59444e4 Return addr 0xb6d18dc4 (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,standard_error_sup},{status,waiting},{initial_call,{proc_lib,in", "y(1) <0.22.0>", "0xb59444f0 Return addr 0xb6d18dc4 (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,code_server},{status,waiting},{initial_call,{erlang,apply,2}},{", "y(1) <0.21.0>", "0xb59444fc Return addr 0xb6d18dc4 (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,file_server_2},{status,waiting},{initial_call,{proc_lib,init_p,", "y(1) <0.20.0>", "0xb5944508 Return addr 0xb6d18dc4 (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,global_group},{status,waiting},{initial_call,{proc_lib,init_p,5", "y(1) <0.19.0>", "0xb5944514 Return addr 0xb6d18dc4 (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,inet_db},{status,waiting},{initial_call,{proc_lib,init_p,5}},{b", "y(1) <0.17.0>", "0xb5944520 Return addr 0xb6d18dc4 (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,[]},{status,waiting},{initial_call,{erlang,apply,2}},{backtrace", "y(1) <0.16.0>", "0xb594452c Return addr 0xb6d18dc4 (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,[]},{status,waiting},{initial_call,{erlang,apply,2}},{backtrace", "y(1) <0.15.0>", "0xb5944538 Return addr 0xb6d18dc4 (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,[]},{status,waiting},{initial_call,{erlang,apply,2}},{backtrace", "y(1) <0.14.0>", "0xb5944544 Return addr 0xb6d18dc4 (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,global_name_server},{status,waiting},{initial_call,{proc_lib,in", "y(1) <0.13.0>", "0xb5944550 Return addr 0xb6d18dc4 (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,rex},{status,waiting},{initial_call,{proc_lib,init_p,5}},{backt", "y(1) <0.12.0>", "0xb594455c Return addr 0xb6d18dc4 (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,kernel_sup},{status,waiting},{initial_call,{proc_lib,init_p,5}}", "y(1) <0.11.0>", "0xb5944568 Return addr 0xb6d18dc4 (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,[]},{status,waiting},{initial_call,{application_master,start_it", "y(1) <0.10.0>", "0xb5944574 Return addr 0xb6d18dc4 (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,[]},{status,waiting},{initial_call,{proc_lib,init_p,5}},{backtr", "y(1) <0.9.0>", "0xb5944580 Return addr 0xb6d18dc4 (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,application_controller},{status,waiting},{initial_call,{erlang,", "y(1) <0.7.0>", "0xb594458c Return addr 0xb6d18dc4 (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,error_logger},{status,waiting},{initial_call,{proc_lib,init_p,5", "y(1) <0.6.0>", "0xb5944598 Return addr 0xb6d18dc4 (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,erl_prim_loader},{status,waiting},{initial_call,{erlang,apply,2", "y(1) <0.3.0>", "0xb59445a4 Return addr 0xb6d17bfc (diag_handler:do_diag_per_node/0 + 112)", "y(0) [{registered_name,init},{status,waiting},{initial_call,{otp_ring0,start,2}},{back", "y(1) <0.0.0>", "0xb59445b0 Return addr 0xb75ba4f4 (rpc:'-handle_call_call/6-fun-0-'/5 + 104)", "y(0) []","y(1) []", "y(2) [{version,[{kernel,\"2.13.4\"},{mnesia,\"4.4.12\"},{stdlib,\"1.16.4\"},{os_mon,\"2.2.4\"}", "y(3) [{{node,'ns_1@127.0.0.1',ns_log},[{filename,\"/var/opt/membase/1.6.5.4r/data/ns_1/", "y(4) [\"bucket_engine_1.6.5.3_1_g5ba5371-Linux.i686.tar.gz\",\"curl-7.21.1-w64_patched.ta", "y(5) [{kernel,\"2.13.4\"},{mnesia,\"4.4.12\"},{stdlib,\"1.16.4\"},{os_mon,\"2.2.4\"},{inets,\"5", "0xb59445cc Return addr 0x08229d74 ()", "y(0) Catch 0xb75ba4f4 (rpc:'-handle_call_call/6-fun-0-'/5 + 104)", "y(1) []","y(2) []","y(3) []", "y(4) <0.12.0>"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,1}]}, {heap_size,317811}, {total_heap_size,832040}, {links,[]}, {memory,3328564}, {message_queue_len,0}, {reductions,1035787}, {trap_exit,false}]}]}, {memory,{1050726400,789147648,{<0.171.0>,4114276}}}, {disk, [{"/",7583436,18}, {"/dev",508460,1}, {"/dev/shm",513048,0}, {"/var/run",513048,1}, {"/var/lock",513048,0}, {"/lib/init/rw",513048,0}, {"/boot",233191,8}]}]}, {'ns_1@10.1.4.244', [{version, [{kernel,"2.13.4"}, {mnesia,"4.4.12"}, {stdlib,"1.16.4"}, {os_mon,"2.2.4"}, {ns_server,"1.6.5.4r"}, {menelaus,"1.6.5.4r"}, {sasl,"2.1.8"}]}, {manifest, ["bucket_engine_1.6.5.3_1_g5ba5371-Linux.i686.tar.gz", "curl-7.21.1-w64_patched.tar.gz", "ep-engine_1.6.5.3.1_8_gce31baa-Linux.i686.tar.gz", "google-perftools-1.6","google-perftools-1.6.tar.gz", "libconflate_1.6.5-Linux.i686.tar.gz", "libevent-2.0.7-rc.tar.gz", "libmemcached-0.41_trond-norbye_mingw32-revno895.tar.gz", "libvbucket_1.6.4-Linux.i686.tar.gz", "membase-cli_1.6.5-Linux.i686.tar.gz", "memcached_1.4.4_382_g9df3289-Linux.i686.tar.gz", "moxi_1.6.5.2-Linux.i686.tar.gz", "ns_server_1.6.5.4r.tar.gz", "vbucketmigrator_1.6.5-Linux.i686.tar.gz", "wallace_1.6.5.4r-4-g8da20bb-Linux.i686"]}, {config, [{{node,'ns_1@127.0.0.1',ns_log}, [{filename, "/var/opt/membase/1.6.5.4r/data/ns_1/ns_log"}]}, {otp, [{'_vclock',[{'ns_1@10.1.4.244',{1,63472033663}}]}, {cookie,elyrdmyctqykhepq}]}, {memory_quota,801}, {{node,'ns_1@10.1.4.243',memcached}, [{'_vclock', [{'ns_1@10.1.4.243',{1,63472033663}}, {'ns_1@127.0.0.1',{1,63472032415}}]}, {dbdir,"/var/opt/membase/1.6.5.4r/data/ns_1"}, {port,11210}, {admin_user,"_admin"}, {admin_pass,"_admin"}, {bucket_engine,"./bin/bucket_engine/bucket_engine.so"}, {engines, [{membase, [{engine,"bin/ep_engine/ep.so"}, {initfile,"/etc/opt/membase/1.6.5.4r/init.sql"}]}, {memcached, [{engine,"bin/memcached/default_engine.so"}]}]}, {verbosity,[]}]}, {{node,'ns_1@10.1.4.243',ns_log}, [{'_vclock',[{'ns_1@10.1.4.243',{1,63472033663}}]}, {filename, "/var/opt/membase/1.6.5.4r/data/ns_1/ns_log"}]}, {{node,'ns_1@10.1.4.244',memcached}, [{'_vclock', [{'ns_1@10.1.4.244',{1,63472033663}}, {'ns_1@127.0.0.1',{1,63472033663}}]}, {dbdir,"/var/opt/membase/1.6.5.4r/data/ns_1"}, {port,11210}, {admin_user,"_admin"}, {admin_pass,"_admin"}, {bucket_engine,"./bin/bucket_engine/bucket_engine.so"}, {engines, [{membase, [{engine,"bin/ep_engine/ep.so"}, {initfile,"/etc/opt/membase/1.6.5.4r/init.sql"}]}, {memcached, [{engine,"bin/memcached/default_engine.so"}]}]}, {verbosity,[]}]}, {buckets, [{'_vclock', [{'ns_1@10.1.4.243',{10,63472034172}}, {'ns_1@127.0.0.1',{3,63472032417}}]}, {configs, [{"default", [{num_replicas,1}, {ram_quota,839909376}, {auth_type,sasl}, {sasl_password,[]}, {type,membase}, {num_vbuckets,1024}, {db_shards,4}, {ht_size,3079}, {tap_keepalive,0}, {tap_noop_interval,20}, {max_txn_size,1000}, {ht_locks,5}, {servers,['ns_1@10.1.4.243','ns_1@10.1.4.244']}, {map, [['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined]]}]}]}]}, {{node,'ns_1@127.0.0.1',membership},active}, {rest_creds, [{'_vclock',[{'ns_1@127.0.0.1',{1,63472032422}}]}, {creds, [{"Administrator",[{password,'filtered-out'}]}]}]}, {port_servers, [{moxi,"./bin/moxi/moxi", ["-Z", {"port_listen=~B,default_bucket_name=default,downstream_max=1024,downstream_conn_max=4,connect_max_errors=5,connect_retry_interval=30000,connect_timeout=400,auth_timeout=100,cycle=200,downstream_conn_queue_timeout=200,downstream_timeout=5000,wait_queue_timeout=200", [port]}, "-z", {"url=http://127.0.0.1:~B/pools/default/saslBucketsStreaming", [{rest,port}]}, "-p","0","-Y","y","-O","stderr", {"~s",[verbosity]}], [{env, [{"EVENT_NOSELECT","1"}, {"MOXI_SASL_PLAIN_USR", {"~s",[{ns_moxi_sup,rest_user,[]}]}}, {"MOXI_SASL_PLAIN_PWD", {"~s",[{ns_moxi_sup,rest_pass,[]}]}}]}, use_stdio,stderr_to_stdout,stream]}, {memcached,"./bin/memcached/memcached", ["-X","./bin/memcached/stdin_term_handler.so","-p", {"~B",[port]}, "-E","./bin/bucket_engine/bucket_engine.so","-B", "binary","-r","-c","10000","-e", {"admin=~s;default_bucket_name=default;auto_create=false", [admin_user]}, {"~s",[verbosity]}], [{env, [{"EVENT_NOSELECT","1"}, {"MEMCACHED_TOP_KEYS","100"}, {"ISASL_PWFILE",{"~s",[{isasl,path}]}}, {"ISASL_DB_CHECK_TIME","1"}]}, use_stdio,stderr_to_stdout,stream]}]}, {alerts, [{email,[]}, {email_alerts,false}, {email_server, [{user,undefined}, {pass,'filtered-out'}, {addr,undefined}, {port,undefined}, {encrypt,false}]}, {alerts, [server_down,server_unresponsive,server_up, server_joined,server_left,bucket_created, bucket_deleted,bucket_auth_failed]}]}, {rebalance_status,running}, {{node,'ns_1@10.1.4.244',ns_log}, [{'_vclock',[{'ns_1@10.1.4.244',{1,63472033663}}]}, {filename, "/var/opt/membase/1.6.5.4r/data/ns_1/ns_log"}]}, {nodes_wanted, [{'_vclock',[{'ns_1@10.1.4.243',{6,63472034158}}]}, 'ns_1@10.1.4.243','ns_1@10.1.4.244']}, {rest, [{'_vclock',[{'ns_1@127.0.0.1',{1,63472032422}}]}, {port,8091}]}, {{node,'ns_1@10.1.4.243',membership},active}, {{node,'ns_1@10.1.4.244',membership},active}, {{node,'ns_1@127.0.0.1',isasl}, [{path,"/var/opt/membase/1.6.5.4r/data/ns_1/isasl.pw"}]}, {moxi,[{port,11211},{verbosity,[]}]}, {{node,'ns_1@10.1.4.243',isasl}, [{'_vclock',[{'ns_1@10.1.4.243',{1,63472033663}}]}, {path,"/var/opt/membase/1.6.5.4r/data/ns_1/isasl.pw"}]}, {{node,'ns_1@10.1.4.244',isasl}, [{'_vclock',[{'ns_1@10.1.4.244',{1,63472033663}}]}, {path,"/var/opt/membase/1.6.5.4r/data/ns_1/isasl.pw"}]}, {replication,[{enabled,true}]}, {{node,'ns_1@127.0.0.1',memcached}, [{port,11210}, {dbdir,"/var/opt/membase/1.6.5.4r/data/ns_1"}, {admin_user,"_admin"}, {admin_pass,"_admin"}, {bucket_engine,"./bin/bucket_engine/bucket_engine.so"}, {engines, [{membase, [{engine,"bin/ep_engine/ep.so"}, {initfile,"/etc/opt/membase/1.6.5.4r/init.sql"}]}, {memcached, [{engine,"bin/memcached/default_engine.so"}]}]}, {verbosity,[]}]}]}, {basic_info, [{version, [{kernel,"2.13.4"}, {mnesia,"4.4.12"}, {stdlib,"1.16.4"}, {os_mon,"2.2.4"}, {ns_server,"1.6.5.4r"}, {menelaus,"1.6.5.4r"}, {sasl,"2.1.8"}]}, {system_arch,"i686-pc-linux-gnu"}, {wall_clock,373}, {memory_data, {1050726400,412069888,{<6517.255.0>,901104}}}, {disk_data, [{"/",7583436,16}, {"/dev",508460,1}, {"/dev/shm",513048,0}, {"/var/run",513048,1}, {"/var/lock",513048,0}, {"/lib/init/rw",513048,0}, {"/boot",233191,8}]}]}, {processes, [{<6517.0.0>, [{registered_name,init}, {status,waiting}, {initial_call,{otp_ring0,start,2}}, {backtrace, ["Program counter: 0xb731a408 (init:loop/1 + 20)", "CP: 0x00000000 (invalid)","arity = 0", "0xb6934184 Return addr 0x08229d74 ()", "y(0) {state,[{'-root',[<<45 bytes>>]},{'-progname',[<<3 bytes>>]},{'-home',[<<12 bytes"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,83}]}, {heap_size,4181}, {total_heap_size,6765}, {links,[<6517.6.0>,<6517.7.0>,<6517.3.0>]}, {memory,27484}, {message_queue_len,0}, {reductions,34102}, {trap_exit,true}]}, {<6517.3.0>, [{registered_name,erl_prim_loader}, {status,waiting}, {initial_call,{erlang,apply,2}}, {backtrace, ["Program counter: 0xb7521328 (erl_prim_loader:loop/3 + 92)", "CP: 0x00000000 (invalid)","arity = 0", "0xb6d36754 Return addr 0x08229d74 ()", "y(0) []", "y(1) [\"/opt/membase/1.6.5.4r/bin/ns_server/ebin\",\"/opt/membase/1.6.5.4r/bin/ns_server/", "y(2) <0.2.0>", "y(3) {state,efile,[],none,#Port<0.2>,infinity,undefined,true,{prim_state,false,undefin", "y(4) infinity"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,11}]}, {heap_size,4181}, {total_heap_size,8362}, {links,[#Port<6517.2>,<6517.0.0>]}, {memory,33852}, {message_queue_len,0}, {reductions,1070073}, {trap_exit,true}]}, {<6517.6.0>, [{registered_name,error_logger}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb753125c (gen_event:fetch_msg/5 + 44)", "CP: 0x00000000 (invalid)","arity = 0", "0xb73a1f10 Return addr 0xb753b298 (proc_lib:init_p_do_apply/3 + 28)", "y(0) false","y(1) []", "y(2) [{handler,ns_log_mf_h,false,{state,\"/var/opt/membase/1.6.5.4r/logs\",10485760,10,2", "y(3) error_logger","y(4) <0.2.0>", "0xb73a1f28 Return addr 0x08229d74 ()", "y(0) Catch 0xb753b2a8 (proc_lib:init_p_do_apply/3 + 44)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,2}]}, {heap_size,2584}, {total_heap_size,5168}, {links,[<6517.0.0>,<6517.25.0>,#Port<6517.1102>]}, {memory,21156}, {message_queue_len,0}, {reductions,79049}, {trap_exit,true}]}, {<6517.7.0>, [{registered_name,application_controller}, {status,waiting}, {initial_call,{erlang,apply,2}}, {backtrace, ["Program counter: 0xb756951c (gen_server:loop/6 + 144)", "CP: 0x00000000 (invalid)","arity = 0", "0xb73ca774 Return addr 0x08229d74 ()", "y(0) []","y(1) infinity", "y(2) application_controller", "y(3) {state,[],[],[],[{mnesia,<0.314.0>},{ns_server,<0.52.0>},{os_mon,<0.41.0>},{sasl,", "y(4) application_controller", "y(5) <0.2.0>"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,4}]}, {heap_size,2584}, {total_heap_size,20295}, {links, [<6517.41.0>,<6517.52.0>,<6517.314.0>,<6517.9.0>, <6517.32.0>,<6517.0.0>]}, {memory,81724}, {message_queue_len,0}, {reductions,26746}, {trap_exit,true}]}, {<6517.9.0>, [{registered_name,[]}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb759c3b0 (application_master:main_loop/2 + 32)", "CP: 0x00000000 (invalid)","arity = 0", "0xb7388f7c Return addr 0xb753b298 (proc_lib:init_p_do_apply/3 + 28)", "y(0) []", "y(1) {state,<0.10.0>,{appl_data,kernel,[application_controller,erl_reply,auth,boot_ser", "y(2) <0.7.0>", "0xb7388f8c Return addr 0x08229d74 ()", "y(0) Catch 0xb753b2a8 (proc_lib:init_p_do_apply/3 + 44)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,1}]}, {heap_size,377}, {total_heap_size,754}, {links,[<6517.7.0>,<6517.10.0>]}, {memory,3480}, {message_queue_len,0}, {reductions,44}, {trap_exit,true}]}, {<6517.10.0>, [{registered_name,[]}, {status,waiting}, {initial_call,{application_master,start_it,4}}, {backtrace, ["Program counter: 0xb759d6f4 (application_master:loop_it/4 + 40)", "CP: 0x00000000 (invalid)","arity = 0", "0xb7393320 Return addr 0x08229d74 ()", "y(0) []","y(1) kernel","y(2) <0.11.0>", "y(3) <0.9.0>"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,0}]}, {heap_size,233}, {total_heap_size,233}, {links,[<6517.9.0>,<6517.11.0>]}, {memory,1336}, {message_queue_len,0}, {reductions,72}, {trap_exit,true}]}, {<6517.11.0>, [{registered_name,kernel_sup}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb756951c (gen_server:loop/6 + 144)", "CP: 0x00000000 (invalid)","arity = 0", "0xb73bab8c Return addr 0xb753b298 (proc_lib:init_p_do_apply/3 + 28)", "y(0) []","y(1) infinity", "y(2) supervisor", "y(3) {state,{local,kernel_sup},one_for_all,[{child,<0.189.0>,net_sup_dynamic,{erl_dist", "y(4) kernel_sup","y(5) <0.10.0>", "0xb73baba8 Return addr 0x08229d74 ()", "y(0) Catch 0xb753b2a8 (proc_lib:init_p_do_apply/3 + 44)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,13}]}, {heap_size,610}, {total_heap_size,987}, {links, [<6517.22.0>,<6517.26.0>,<6517.27.0>,<6517.189.0>, <6517.24.0>,<6517.17.0>,<6517.20.0>,<6517.21.0>, <6517.19.0>,<6517.12.0>,<6517.13.0>,<6517.10.0>]}, {memory,4612}, {message_queue_len,0}, {reductions,1580}, {trap_exit,true}]}, {<6517.12.0>, [{registered_name,rex}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb756951c (gen_server:loop/6 + 144)", "CP: 0x00000000 (invalid)","arity = 0", "0xb5e64cfc Return addr 0xb753b298 (proc_lib:init_p_do_apply/3 + 28)", "y(0) []","y(1) infinity","y(2) rpc", "y(3) {1,{<0.801.0>,{<10111.5348.0>,{#Ref<10111.0.0.103227>,'ns_1@10.1.4.244'}},nil,nil", "y(4) rex","y(5) <0.11.0>", "0xb5e64d18 Return addr 0x08229d74 ()", "y(0) Catch 0xb753b2a8 (proc_lib:init_p_do_apply/3 + 44)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,0}]}, {heap_size,196418}, {total_heap_size,196418}, {links,[<6517.11.0>]}, {memory,786220}, {message_queue_len,0}, {reductions,8848}, {trap_exit,true}]}, {<6517.13.0>, [{registered_name,global_name_server}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb756951c (gen_server:loop/6 + 144)", "CP: 0x00000000 (invalid)","arity = 0", "0xb73b6a34 Return addr 0xb753b298 (proc_lib:init_p_do_apply/3 + 28)", "y(0) []","y(1) infinity","y(2) global", "y(3) {state,true,['ns_1@10.1.4.243'],['ns_1@10.1.4.243'],[],[],'ns_1@10.1.4.244',<0.14", "y(4) global_name_server","y(5) <0.11.0>", "0xb73b6a50 Return addr 0x08229d74 ()", "y(0) Catch 0xb753b2a8 (proc_lib:init_p_do_apply/3 + 44)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,14}]}, {heap_size,610}, {total_heap_size,987}, {links, [<6517.14.0>,<6517.16.0>,<6517.15.0>,<6517.11.0>]}, {memory,4564}, {message_queue_len,0}, {reductions,1992}, {trap_exit,true}]}, {<6517.14.0>, [{registered_name,[]}, {status,waiting}, {initial_call,{erlang,apply,2}}, {backtrace, ["Program counter: 0xb75d3094 (global:loop_the_locker/1 + 588)", "CP: 0x00000000 (invalid)","arity = 0", "0xb73bebac Return addr 0xb75d2e28 (global:init_the_locker/1 + 192)", "y(0) {multi,[],[],['ns_1@10.1.4.243'],'ns_1@10.1.4.244',false,false}", "y(1) infinity", "0xb73bebb8 Return addr 0x08229d74 ()"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,2}]}, {heap_size,377}, {total_heap_size,754}, {links,[<6517.13.0>]}, {memory,3400}, {message_queue_len,0}, {reductions,356}, {trap_exit,true}]}, {<6517.15.0>, [{registered_name,[]}, {status,waiting}, {initial_call,{erlang,apply,2}}, {backtrace, ["Program counter: 0xb75d7140 (global:collect_deletions/2 + 76)", "CP: 0x00000000 (invalid)","arity = 0", "0xb73ddb38 Return addr 0xb75d70c0 (global:loop_the_deleter/1 + 36)", "y(0) infinity","y(1) []", "y(2) <0.13.0>", "0xb73ddb48 Return addr 0x08229d74 ()", "y(0) <0.13.0>"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,4}]}, {heap_size,610}, {total_heap_size,987}, {links,[<6517.13.0>]}, {memory,4332}, {message_queue_len,0}, {reductions,342}, {trap_exit,false}]}, {<6517.16.0>, [{registered_name,[]}, {status,waiting}, {initial_call,{erlang,apply,2}}, {backtrace, ["Program counter: 0xb75d7260 (global:loop_the_registrar/0 + 12)", "CP: 0x00000000 (invalid)","arity = 0", "0xb73df2e4 Return addr 0x08229d74 ()", "y(0) []"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,5}]}, {heap_size,377}, {total_heap_size,754}, {links,[<6517.13.0>]}, {memory,3400}, {message_queue_len,0}, {reductions,362}, {trap_exit,false}]}, {<6517.17.0>, [{registered_name,inet_db}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb756951c (gen_server:loop/6 + 144)", "CP: 0x00000000 (invalid)","arity = 0", "0xb739ef58 Return addr 0xb753b298 (proc_lib:init_p_do_apply/3 + 28)", "y(0) []","y(1) infinity","y(2) inet_db", "y(3) {state,inet_db,inet_cache,inet_hosts_byname,inet_hosts_byaddr,inet_hosts_file_byn", "y(4) inet_db","y(5) <0.11.0>", "0xb739ef74 Return addr 0x08229d74 ()", "y(0) Catch 0xb753b2a8 (proc_lib:init_p_do_apply/3 + 44)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,2}]}, {heap_size,1597}, {total_heap_size,3194}, {links,[<6517.11.0>]}, {memory,13220}, {message_queue_len,0}, {reductions,2005}, {trap_exit,true}]}, {<6517.19.0>, [{registered_name,global_group}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb756951c (gen_server:loop/6 + 144)", "CP: 0x00000000 (invalid)","arity = 0", "0xb7398588 Return addr 0xb753b298 (proc_lib:init_p_do_apply/3 + 28)", "y(0) []","y(1) infinity", "y(2) global_group", "y(3) {state,no_conf,true,[],[],[],[],[],'nonode@nohost',[],normal,normal}", "y(4) global_group","y(5) <0.11.0>", "0xb73985a4 Return addr 0x08229d74 ()", "y(0) Catch 0xb753b2a8 (proc_lib:init_p_do_apply/3 + 44)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,0}]}, {heap_size,233}, {total_heap_size,233}, {links,[<6517.11.0>]}, {memory,1376}, {message_queue_len,0}, {reductions,92}, {trap_exit,true}]}, {<6517.20.0>, [{registered_name,file_server_2}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb756951c (gen_server:loop/6 + 144)", "CP: 0x00000000 (invalid)","arity = 0", "0xb6d938e8 Return addr 0xb753b298 (proc_lib:init_p_do_apply/3 + 28)", "y(0) []","y(1) infinity", "y(2) file_server","y(3) #Port<0.262>", "y(4) file_server_2","y(5) <0.11.0>", "0xb6d93904 Return addr 0x08229d74 ()", "y(0) Catch 0xb753b2a8 (proc_lib:init_p_do_apply/3 + 44)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,117}]}, {heap_size,6765}, {total_heap_size,13530}, {links,[#Port<6517.262>,<6517.11.0>]}, {memory,54584}, {message_queue_len,0}, {reductions,676418}, {trap_exit,true}]}, {<6517.21.0>, [{registered_name,code_server}, {status,waiting}, {initial_call,{erlang,apply,2}}, {backtrace, ["Program counter: 0xb6c357a4 (code_server:loop/1 + 64)", "CP: 0x00000000 (invalid)","arity = 0", "0xb6d8cf48 Return addr 0x08229d74 ()", "y(0) {state,<0.11.0>,\"/opt/membase/1.6.5.4r/erlang-13b03/lib/erlang\",[\"/opt/membase/1.", "y(1) <0.11.0>"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,279}]}, {heap_size,4181}, {total_heap_size,21892}, {links,[<6517.11.0>]}, {memory,87952}, {message_queue_len,0}, {reductions,264455}, {trap_exit,true}]}, {<6517.22.0>, [{registered_name,standard_error_sup}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb756951c (gen_server:loop/6 + 144)", "CP: 0x00000000 (invalid)","arity = 0", "0xb739f300 Return addr 0xb753b298 (proc_lib:init_p_do_apply/3 + 28)", "y(0) []","y(1) infinity", "y(2) supervisor_bridge", "y(3) {state,standard_error,<0.23.0>,<0.23.0>,{local,standard_error_sup}}", "y(4) standard_error_sup","y(5) <0.11.0>", "0xb739f31c Return addr 0x08229d74 ()", "y(0) Catch 0xb753b2a8 (proc_lib:init_p_do_apply/3 + 44)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,0}]}, {heap_size,233}, {total_heap_size,233}, {links,[<6517.11.0>,<6517.23.0>]}, {memory,1396}, {message_queue_len,0}, {reductions,40}, {trap_exit,true}]}, {<6517.23.0>, [{registered_name,standard_error}, {status,waiting}, {initial_call,{standard_error,server,2}}, {backtrace, ["Program counter: 0xb6c2588c (standard_error:server_loop/1 + 20)", "CP: 0x00000000 (invalid)","arity = 0", "0xb739f6c4 Return addr 0x08229d74 ()", "y(0) #Port<0.568>"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,0}]}, {heap_size,233}, {total_heap_size,233}, {links,[<6517.22.0>,#Port<6517.568>]}, {memory,1396}, {message_queue_len,0}, {reductions,7}, {trap_exit,true}]}, {<6517.24.0>, [{registered_name,[]}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb756951c (gen_server:loop/6 + 144)", "CP: 0x00000000 (invalid)","arity = 0", "0xb739d660 Return addr 0xb753b298 (proc_lib:init_p_do_apply/3 + 28)", "y(0) []","y(1) infinity", "y(2) supervisor_bridge", "y(3) {state,user_sup,<0.25.0>,<0.25.0>,{<0.24.0>,user_sup}}", "y(4) <0.24.0>","y(5) <0.11.0>", "0xb739d67c Return addr 0x08229d74 ()", "y(0) Catch 0xb753b2a8 (proc_lib:init_p_do_apply/3 + 44)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,0}]}, {heap_size,1597}, {total_heap_size,1597}, {links,[<6517.11.0>,<6517.25.0>]}, {memory,6852}, {message_queue_len,0}, {reductions,71}, {trap_exit,true}]}, {<6517.25.0>, [{registered_name,user}, {status,waiting}, {initial_call,{user,server,2}}, {backtrace, ["Program counter: 0xb6c49cfc (user:server_loop/2 + 28)", "CP: 0x00000000 (invalid)","arity = 0", "0xb6dd4ab8 Return addr 0x08229d74 ()", "y(0) []","y(1) []","y(2) {[],[]}", "y(3) #Port<0.587>"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,33}]}, {heap_size,6765}, {total_heap_size,13530}, {links,[<6517.6.0>,<6517.24.0>,#Port<6517.587>]}, {memory,54604}, {message_queue_len,0}, {reductions,49878}, {trap_exit,true}]}, {<6517.26.0>, [{registered_name,[]}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb756951c (gen_server:loop/6 + 144)", "CP: 0x00000000 (invalid)","arity = 0", "0xb73883a0 Return addr 0xb753b298 (proc_lib:init_p_do_apply/3 + 28)", "y(0) []","y(1) infinity", "y(2) kernel_config","y(3) []", "y(4) <0.26.0>","y(5) <0.11.0>", "0xb73883bc Return addr 0x08229d74 ()", "y(0) Catch 0xb753b2a8 (proc_lib:init_p_do_apply/3 + 44)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,0}]}, {heap_size,233}, {total_heap_size,233}, {links,[<6517.11.0>]}, {memory,1376}, {message_queue_len,0}, {reductions,176}, {trap_exit,true}]}, {<6517.27.0>, [{registered_name,kernel_safe_sup}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb756951c (gen_server:loop/6 + 144)", "CP: 0x00000000 (invalid)","arity = 0", "0xb6d602a8 Return addr 0xb753b298 (proc_lib:init_p_do_apply/3 + 28)", "y(0) []","y(1) infinity", "y(2) supervisor", "y(3) {state,{local,kernel_safe_sup},one_for_one,[{child,<0.180.0>,inet_gethost_native_", "y(4) kernel_safe_sup","y(5) <0.11.0>", "0xb6d602c4 Return addr 0x08229d74 ()", "y(0) Catch 0xb753b2a8 (proc_lib:init_p_do_apply/3 + 44)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,5}]}, {heap_size,377}, {total_heap_size,754}, {links, [<6517.123.0>,<6517.128.0>,<6517.180.0>,<6517.127.0>, <6517.50.0>,<6517.122.0>,<6517.11.0>]}, {memory,3580}, {message_queue_len,0}, {reductions,446}, {trap_exit,true}]}, {<6517.32.0>, [{registered_name,[]}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb759c3b0 (application_master:main_loop/2 + 32)", "CP: 0x00000000 (invalid)","arity = 0", "0xb7389a74 Return addr 0xb753b298 (proc_lib:init_p_do_apply/3 + 28)", "y(0) []", "y(1) {state,<0.33.0>,{appl_data,sasl,[sasl_sup,alarm_handler,overload,release_handler]", "y(2) <0.7.0>", "0xb7389a84 Return addr 0x08229d74 ()", "y(0) Catch 0xb753b2a8 (proc_lib:init_p_do_apply/3 + 44)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,0}]}, {heap_size,233}, {total_heap_size,233}, {links,[<6517.7.0>,<6517.33.0>]}, {memory,1396}, {message_queue_len,0}, {reductions,23}, {trap_exit,true}]}, {<6517.33.0>, [{registered_name,[]}, {status,waiting}, {initial_call,{application_master,start_it,4}}, {backtrace, ["Program counter: 0xb759d6f4 (application_master:loop_it/4 + 40)", "CP: 0x00000000 (invalid)","arity = 0", "0xb7393a70 Return addr 0x08229d74 ()", "y(0) {state,tty,undefined}","y(1) sasl", "y(2) <0.34.0>","y(3) <0.32.0>"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,0}]}, {heap_size,233}, {total_heap_size,233}, {links,[<6517.32.0>,<6517.34.0>]}, {memory,1336}, {message_queue_len,0}, {reductions,70}, {trap_exit,true}]}, {<6517.34.0>, [{registered_name,sasl_sup}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb756951c (gen_server:loop/6 + 144)", "CP: 0x00000000 (invalid)","arity = 0", "0xb73998a8 Return addr 0xb753b298 (proc_lib:init_p_do_apply/3 + 28)", "y(0) []","y(1) infinity", "y(2) supervisor", "y(3) {state,{local,sasl_sup},one_for_one,[{child,<0.38.0>,release_handler,{release_han", "y(4) sasl_sup","y(5) <0.33.0>", "0xb73998c4 Return addr 0x08229d74 ()", "y(0) Catch 0xb753b2a8 (proc_lib:init_p_do_apply/3 + 44)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,1}]}, {heap_size,233}, {total_heap_size,610}, {links,[<6517.35.0>,<6517.38.0>,<6517.33.0>]}, {memory,2924}, {message_queue_len,0}, {reductions,158}, {trap_exit,true}]}, {<6517.35.0>, [{registered_name,sasl_safe_sup}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb756951c (gen_server:loop/6 + 144)", "CP: 0x00000000 (invalid)","arity = 0", "0xb73917e0 Return addr 0xb753b298 (proc_lib:init_p_do_apply/3 + 28)", "y(0) []","y(1) infinity", "y(2) supervisor", "y(3) {state,{local,sasl_safe_sup},one_for_one,[{child,<0.37.0>,overload,{overload,star", "y(4) sasl_safe_sup","y(5) <0.34.0>", "0xb73917fc Return addr 0x08229d74 ()", "y(0) Catch 0xb753b2a8 (proc_lib:init_p_do_apply/3 + 44)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,1}]}, {heap_size,233}, {total_heap_size,610}, {links,[<6517.36.0>,<6517.37.0>,<6517.34.0>]}, {memory,2924}, {message_queue_len,0}, {reductions,165}, {trap_exit,true}]}, {<6517.36.0>, [{registered_name,alarm_handler}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb753125c (gen_event:fetch_msg/5 + 44)", "CP: 0x00000000 (invalid)","arity = 0", "0xb7399c54 Return addr 0xb753b298 (proc_lib:init_p_do_apply/3 + 28)", "y(0) false","y(1) []", "y(2) [{handler,alarm_handler,false,[],false}]", "y(3) alarm_handler","y(4) <0.35.0>", "0xb7399c6c Return addr 0x08229d74 ()", "y(0) Catch 0xb753b2a8 (proc_lib:init_p_do_apply/3 + 44)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,0}]}, {heap_size,233}, {total_heap_size,233}, {links,[<6517.35.0>]}, {memory,1376}, {message_queue_len,0}, {reductions,28}, {trap_exit,true}]}, {<6517.37.0>, [{registered_name,overload}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb756951c (gen_server:loop/6 + 144)", "CP: 0x00000000 (invalid)","arity = 0", "0xb739a3a0 Return addr 0xb753b298 (proc_lib:init_p_do_apply/3 + 28)", "y(0) []","y(1) infinity", "y(2) overload", "y(3) {state,0,0,8.000000e-01,193,1.000000e-01,{0,0},clear}", "y(4) overload","y(5) <0.35.0>", "0xb739a3bc Return addr 0x08229d74 ()", "y(0) Catch 0xb753b2a8 (proc_lib:init_p_do_apply/3 + 44)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,0}]}, {heap_size,233}, {total_heap_size,233}, {links,[<6517.35.0>]}, {memory,1376}, {message_queue_len,0}, {reductions,39}, {trap_exit,false}]}, {<6517.38.0>, [{registered_name,release_handler}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb756951c (gen_server:loop/6 + 144)", "CP: 0x00000000 (invalid)","arity = 0", "0xb738de98 Return addr 0xb753b298 (proc_lib:init_p_do_apply/3 + 28)", "y(0) []","y(1) infinity", "y(2) release_handler", "y(3) {state,[],\"/opt/membase/1.6.5.4r/erlang-13b03/lib/erlang\",\"/opt/membase/1.6.5.4r/", "y(4) release_handler","y(5) <0.34.0>", "0xb738deb4 Return addr 0x08229d74 ()", "y(0) Catch 0xb753b2a8 (proc_lib:init_p_do_apply/3 + 44)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,3}]}, {heap_size,1597}, {total_heap_size,4181}, {links,[<6517.34.0>]}, {memory,17168}, {message_queue_len,0}, {reductions,2503}, {trap_exit,false}]}, {<6517.41.0>, [{registered_name,[]}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb759c3b0 (application_master:main_loop/2 + 32)", "CP: 0x00000000 (invalid)","arity = 0", "0xb73b7184 Return addr 0xb753b298 (proc_lib:init_p_do_apply/3 + 28)", "y(0) []", "y(1) {state,<0.42.0>,{appl_data,os_mon,[os_mon_sup,os_mon_sysinfo,disksup,memsup,cpu_s", "y(2) <0.7.0>", "0xb73b7194 Return addr 0x08229d74 ()", "y(0) Catch 0xb753b2a8 (proc_lib:init_p_do_apply/3 + 44)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,0}]}, {heap_size,233}, {total_heap_size,233}, {links,[<6517.7.0>,<6517.42.0>]}, {memory,1396}, {message_queue_len,0}, {reductions,23}, {trap_exit,true}]}, {<6517.42.0>, [{registered_name,[]}, {status,waiting}, {initial_call,{application_master,start_it,4}}, {backtrace, ["Program counter: 0xb759d6f4 (application_master:loop_it/4 + 40)", "CP: 0x00000000 (invalid)","arity = 0", "0xb73b7530 Return addr 0x08229d74 ()", "y(0) []","y(1) os_mon","y(2) <0.43.0>", "y(3) <0.41.0>"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,0}]}, {heap_size,233}, {total_heap_size,233}, {links,[<6517.41.0>,<6517.43.0>]}, {memory,1336}, {message_queue_len,0}, {reductions,40}, {trap_exit,true}]}, {<6517.43.0>, [{registered_name,os_mon_sup}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb756951c (gen_server:loop/6 + 144)", "CP: 0x00000000 (invalid)","arity = 0", "0xb7392b00 Return addr 0xb753b298 (proc_lib:init_p_do_apply/3 + 28)", "y(0) []","y(1) infinity", "y(2) supervisor", "y(3) {state,{local,os_mon_sup},one_for_one,[{child,<0.47.0>,cpu_sup,{cpu_sup,start_lin", "y(4) os_mon_sup","y(5) <0.42.0>", "0xb7392b1c Return addr 0x08229d74 ()", "y(0) Catch 0xb753b2a8 (proc_lib:init_p_do_apply/3 + 44)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,1}]}, {heap_size,377}, {total_heap_size,754}, {links, [<6517.44.0>,<6517.45.0>,<6517.47.0>,<6517.42.0>]}, {memory,3520}, {message_queue_len,0}, {reductions,273}, {trap_exit,true}]}, {<6517.44.0>, [{registered_name,disksup}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb756951c (gen_server:loop/6 + 144)", "CP: 0x00000000 (invalid)","arity = 0", "0xb73eda88 Return addr 0xb753b298 (proc_lib:init_p_do_apply/3 + 28)", "y(0) []","y(1) infinity","y(2) disksup", "y(3) {state,80,60000,{unix,linux},[{\"/\",7583436,16},{\"/dev\",508460,1},{\"/dev/shm\",5130", "y(4) disksup","y(5) <0.43.0>", "0xb73edaa4 Return addr 0x08229d74 ()", "y(0) Catch 0xb753b2a8 (proc_lib:init_p_do_apply/3 + 44)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,14}]}, {heap_size,4181}, {total_heap_size,8362}, {links,[<6517.43.0>,#Port<6517.988>]}, {memory,33912}, {message_queue_len,0}, {reductions,17175}, {trap_exit,true}]}, {<6517.45.0>, [{registered_name,memsup}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb756951c (gen_server:loop/6 + 144)", "CP: 0x00000000 (invalid)","arity = 0", "0xb73e4394 Return addr 0xb753b298 (proc_lib:init_p_do_apply/3 + 28)", "y(0) []","y(1) infinity","y(2) memsup", "y(3) {state,{unix,linux},true,{412069888,1050726400},{<0.255.0>,901104},false,60000,30", "y(4) memsup","y(5) <0.43.0>", "0xb73e43b0 Return addr 0x08229d74 ()", "y(0) Catch 0xb753b2a8 (proc_lib:init_p_do_apply/3 + 44)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,22}]}, {heap_size,2584}, {total_heap_size,3194}, {links,[<6517.43.0>,<6517.46.0>]}, {memory,13240}, {message_queue_len,0}, {reductions,15545}, {trap_exit,true}]}, {<6517.46.0>, [{registered_name,[]}, {status,waiting}, {initial_call,{erlang,apply,2}}, {backtrace, ["Program counter: 0xb6cab65c (memsup:port_idle/1 + 20)", "CP: 0x00000000 (invalid)","arity = 0", "0xb6d56ab8 Return addr 0x08229d74 ()", "y(0) []","y(1) #Port<0.1020>"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,10}]}, {heap_size,1597}, {total_heap_size,1974}, {links,[<6517.45.0>,#Port<6517.1020>]}, {memory,8300}, {message_queue_len,0}, {reductions,4392}, {trap_exit,true}]}, {<6517.47.0>, [{registered_name,cpu_sup}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb756951c (gen_server:loop/6 + 144)", "CP: 0x00000000 (invalid)","arity = 0", "0xb7391f30 Return addr 0xb753b298 (proc_lib:init_p_do_apply/3 + 28)", "y(0) []","y(1) infinity","y(2) cpu_sup", "y(3) {state,<0.48.0>,{unix,linux}}", "y(4) cpu_sup","y(5) <0.43.0>", "0xb7391f4c Return addr 0x08229d74 ()", "y(0) Catch 0xb753b2a8 (proc_lib:init_p_do_apply/3 + 44)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,0}]}, {heap_size,233}, {total_heap_size,233}, {links,[<6517.43.0>]}, {memory,1376}, {message_queue_len,0}, {reductions,33}, {trap_exit,true}]}, {<6517.48.0>, [{registered_name,[]}, {status,waiting}, {initial_call,{erlang,apply,2}}, {backtrace, ["Program counter: 0xb6cbd964 (cpu_sup:measurement_server_loop/1 + 20)", "CP: 0x00000000 (invalid)","arity = 0", "0xb7392ea4 Return addr 0x08229d74 ()", "y(0) []","y(1) []","y(2) []", "y(3) []","y(4) []","y(5) []", "y(6) []","y(7) []", "y(8) {internal,<0.49.0>,[],{unix,linux}}"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,0}]}, {heap_size,233}, {total_heap_size,233}, {links,[<6517.49.0>]}, {memory,1316}, {message_queue_len,0}, {reductions,11}, {trap_exit,true}]}, {<6517.49.0>, [{registered_name,[]}, {status,waiting}, {initial_call,{erlang,apply,2}}, {backtrace, ["Program counter: 0xb6cbe250 (cpu_sup:port_server_loop/2 + 32)", "CP: 0x00000000 (invalid)","arity = 0", "0xb73d3f88 Return addr 0x08229d74 ()", "y(0) []","y(1) 6000", "y(2) #Port<0.1062>"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,3}]}, {heap_size,610}, {total_heap_size,987}, {links,[<6517.48.0>,#Port<6517.1062>]}, {memory,4352}, {message_queue_len,0}, {reductions,335}, {trap_exit,false}]}, {<6517.50.0>, [{registered_name,timer_server}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb756951c (gen_server:loop/6 + 144)", "CP: 0x00000000 (invalid)","arity = 0", "0xb6d87d04 Return addr 0xb753b298 (proc_lib:init_p_do_apply/3 + 28)", "y(0) []","y(1) 398","y(2) timer", "y(3) []","y(4) timer_server", "y(5) <0.27.0>", "0xb6d87d20 Return addr 0x08229d74 ()", "y(0) Catch 0xb753b2a8 (proc_lib:init_p_do_apply/3 + 44)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,240}]}, {heap_size,2584}, {total_heap_size,2961}, {links, [<6517.277.0>,<6517.326.0>,<6517.379.0>,<6517.418.0>, <6517.374.0>,<6517.278.0>,<6517.322.0>,<6517.256.0>, <6517.264.0>,<6517.27.0>]}, {memory,12468}, {message_queue_len,0}, {reductions,79202}, {trap_exit,true}]}, {<6517.52.0>, [{registered_name,[]}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb759c3b0 (application_master:main_loop/2 + 32)", "CP: 0x00000000 (invalid)","arity = 0", "0xb6dce0fc Return addr 0xb753b298 (proc_lib:init_p_do_apply/3 + 28)", "y(0) []", "y(1) {state,<0.53.0>,{appl_data,ns_server,[ns_server_sup,ns_config,ns_config_sup,ns_co", "y(2) <0.7.0>", "0xb6dce10c Return addr 0x08229d74 ()", "y(0) Catch 0xb753b2a8 (proc_lib:init_p_do_apply/3 + 44)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,1}]}, {heap_size,233}, {total_heap_size,610}, {links,[<6517.7.0>,<6517.53.0>]}, {memory,2904}, {message_queue_len,0}, {reductions,29}, {trap_exit,true}]}, {<6517.53.0>, [{registered_name,[]}, {status,waiting}, {initial_call,{application_master,start_it,4}}, {backtrace, ["Program counter: 0xb759d6f4 (application_master:loop_it/4 + 40)", "CP: 0x00000000 (invalid)","arity = 0", "0xb73d33b8 Return addr 0x08229d74 ()", "y(0) []","y(1) ns_server", "y(2) <0.54.0>","y(3) <0.52.0>"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,0}]}, {heap_size,233}, {total_heap_size,233}, {links,[<6517.52.0>,<6517.54.0>]}, {memory,1336}, {message_queue_len,0}, {reductions,50}, {trap_exit,true}]}, {<6517.54.0>, [{registered_name,ns_server_cluster_sup}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb756951c (gen_server:loop/6 + 144)", "CP: 0x00000000 (invalid)","arity = 0", "0xb73bd3f4 Return addr 0xb753b298 (proc_lib:init_p_do_apply/3 + 28)", "y(0) []","y(1) infinity", "y(2) supervisor", "y(3) {state,{local,ns_server_cluster_sup},one_for_one,[{child,<0.255.0>,ns_server_sup,", "y(4) ns_server_cluster_sup","y(5) <0.53.0>", "0xb73bd410 Return addr 0x08229d74 ()", "y(0) Catch 0xb753b2a8 (proc_lib:init_p_do_apply/3 + 44)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,1}]}, {heap_size,2584}, {total_heap_size,5168}, {links, [<6517.55.0>,<6517.63.0>,<6517.255.0>,<6517.62.0>, <6517.53.0>]}, {memory,21196}, {message_queue_len,0}, {reductions,1865}, {trap_exit,true}]}, {<6517.55.0>, [{registered_name,dist_manager}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb756951c (gen_server:loop/6 + 144)", "CP: 0x00000000 (invalid)","arity = 0", "0xb73fc360 Return addr 0xb753b298 (proc_lib:init_p_do_apply/3 + 28)", "y(0) []","y(1) infinity", "y(2) dist_manager", "y(3) {state,true,\"10.1.4.244\"}", "y(4) dist_manager","y(5) <0.54.0>", "0xb73fc37c Return addr 0x08229d74 ()", "y(0) Catch 0xb753b2a8 (proc_lib:init_p_do_apply/3 + 44)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,6}]}, {heap_size,377}, {total_heap_size,754}, {links,[<6517.54.0>]}, {memory,3460}, {message_queue_len,0}, {reductions,598}, {trap_exit,false}]}, {<6517.62.0>, [{registered_name,ns_cluster}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb756951c (gen_server:loop/6 + 144)", "CP: 0x00000000 (invalid)","arity = 0", "0xb73d07a0 Return addr 0xb753b298 (proc_lib:init_p_do_apply/3 + 28)", "y(0) []","y(1) infinity", "y(2) ns_cluster","y(3) {state}", "y(4) ns_cluster","y(5) <0.54.0>", "0xb73d07bc Return addr 0x08229d74 ()", "y(0) Catch 0xb753b2a8 (proc_lib:init_p_do_apply/3 + 44)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,12}]}, {heap_size,4181}, {total_heap_size,8362}, {links,[<6517.54.0>]}, {memory,33892}, {message_queue_len,0}, {reductions,12029}, {trap_exit,false}]}, {<6517.63.0>, [{registered_name,ns_config_sup}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb756951c (gen_server:loop/6 + 144)", "CP: 0x00000000 (invalid)","arity = 0", "0xb6d2be68 Return addr 0xb753b298 (proc_lib:init_p_do_apply/3 + 28)", "y(0) []","y(1) infinity", "y(2) supervisor", "y(3) {state,{local,ns_config_sup},rest_for_one,[{child,undefined,ns_config_log,{ns_con", "y(4) ns_config_sup","y(5) <0.54.0>", "0xb6d2be84 Return addr 0x08229d74 ()", "y(0) Catch 0xb753b2a8 (proc_lib:init_p_do_apply/3 + 44)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,0}]}, {heap_size,6765}, {total_heap_size,6765}, {links,[<6517.64.0>,<6517.65.0>,<6517.54.0>]}, {memory,27544}, {message_queue_len,0}, {reductions,582}, {trap_exit,true}]}, {<6517.64.0>, [{registered_name,ns_config_events}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x08229d70 (unknown function)", "CP: 0x08229d74 ()", "arity = 3"," proc_lib"," wake_up", " [gen_event,wake_hib,[<0.63.0>,ns_config_events,[{handler,ns_pubsub,#Ref<0.0.0.2881>,{st"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,5}]}, {heap_size,7010}, {total_heap_size,7010}, {links, [<6517.305.0>,<6517.345.0>,<6517.346.0>,<6517.63.0>]}, {memory,28584}, {message_queue_len,0}, {reductions,102265}, {trap_exit,true}]}, {<6517.65.0>, [{registered_name,ns_config}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb756951c (gen_server:loop/6 + 144)", "CP: 0x00000000 (invalid)","arity = 0", "0xb69be760 Return addr 0xb753b298 (proc_lib:init_p_do_apply/3 + 28)", "y(0) []","y(1) infinity", "y(2) ns_config", "y(3) {config,{full,\"/etc/opt/membase/1.6.5.4r/config\",undefined,ns_config_default},[[{", "y(4) ns_config","y(5) <0.63.0>", "0xb69be77c Return addr 0x08229d74 ()", "y(0) Catch 0xb753b2a8 (proc_lib:init_p_do_apply/3 + 44)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,12}]}, {heap_size,28657}, {total_heap_size,57314}, {links,[<6517.63.0>]}, {memory,229700}, {message_queue_len,0}, {reductions,44740}, {trap_exit,true}]}, {<6517.122.0>, [{registered_name,dets_sup}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb756951c (gen_server:loop/6 + 144)", "CP: 0x00000000 (invalid)","arity = 0", "0xb6d31520 Return addr 0xb753b298 (proc_lib:init_p_do_apply/3 + 28)", "y(0) []","y(1) infinity", "y(2) supervisor", "y(3) {state,{local,dets_sup},simple_one_for_one,[{child,undefined,dets,{dets,istart_li", "y(4) dets_sup","y(5) <0.27.0>", "0xb6d3153c Return addr 0x08229d74 ()", "y(0) Catch 0xb753b2a8 (proc_lib:init_p_do_apply/3 + 44)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,7}]}, {heap_size,377}, {total_heap_size,987}, {links,[<6517.27.0>]}, {memory,4392}, {message_queue_len,0}, {reductions,782}, {trap_exit,true}]}, {<6517.123.0>, [{registered_name,dets}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb756951c (gen_server:loop/6 + 144)", "CP: 0x00000000 (invalid)","arity = 0", "0xb73dc234 Return addr 0xb753b298 (proc_lib:init_p_do_apply/3 + 28)", "y(0) []","y(1) infinity", "y(2) dets_server", "y(3) {state,20510,[<0.27.0>],[]}", "y(4) dets","y(5) <0.27.0>", "0xb73dc250 Return addr 0x08229d74 ()", "y(0) Catch 0xb753b2a8 (proc_lib:init_p_do_apply/3 + 44)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,14}]}, {heap_size,610}, {total_heap_size,987}, {links,[<6517.27.0>]}, {memory,4392}, {message_queue_len,0}, {reductions,1136}, {trap_exit,true}]}, {<6517.127.0>, [{registered_name,disk_log_sup}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb756951c (gen_server:loop/6 + 144)", "CP: 0x00000000 (invalid)","arity = 0", "0xb73d67dc Return addr 0xb753b298 (proc_lib:init_p_do_apply/3 + 28)", "y(0) []","y(1) infinity", "y(2) supervisor", "y(3) {state,{local,disk_log_sup},simple_one_for_one,[{child,undefined,disk_log,{disk_l", "y(4) disk_log_sup","y(5) <0.27.0>", "0xb73d67f8 Return addr 0x08229d74 ()", "y(0) Catch 0xb753b2a8 (proc_lib:init_p_do_apply/3 + 44)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,10}]}, {heap_size,2584}, {total_heap_size,3194}, {links,[<6517.330.0>,<6517.27.0>]}, {memory,13240}, {message_queue_len,0}, {reductions,1637}, {trap_exit,true}]}, {<6517.128.0>, [{registered_name,disk_log_server}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb756951c (gen_server:loop/6 + 144)", "CP: 0x00000000 (invalid)","arity = 0", "0xb6e13d7c Return addr 0xb753b298 (proc_lib:init_p_do_apply/3 + 28)", "y(0) []","y(1) infinity", "y(2) disk_log_server","y(3) {state,[]}", "y(4) disk_log_server","y(5) <0.27.0>", "0xb6e13d98 Return addr 0x08229d74 ()", "y(0) Catch 0xb753b2a8 (proc_lib:init_p_do_apply/3 + 44)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,28}]}, {heap_size,610}, {total_heap_size,1220}, {links,[<6517.330.0>,<6517.27.0>]}, {memory,5344}, {message_queue_len,0}, {reductions,1709}, {trap_exit,true}]}, {<6517.180.0>, [{registered_name,inet_gethost_native_sup}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb756951c (gen_server:loop/6 + 144)", "CP: 0x00000000 (invalid)","arity = 0", "0xb7391b88 Return addr 0xb753b298 (proc_lib:init_p_do_apply/3 + 28)", "y(0) []","y(1) infinity", "y(2) supervisor_bridge", "y(3) {state,inet_gethost_native,<0.181.0>,<0.181.0>,{local,inet_gethost_native_sup}}", "y(4) inet_gethost_native_sup", "y(5) <0.27.0>", "0xb7391ba4 Return addr 0x08229d74 ()", "y(0) Catch 0xb753b2a8 (proc_lib:init_p_do_apply/3 + 44)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,0}]}, {heap_size,233}, {total_heap_size,233}, {links,[<6517.27.0>,<6517.181.0>]}, {memory,1396}, {message_queue_len,0}, {reductions,41}, {trap_exit,true}]}, {<6517.181.0>, [{registered_name,inet_gethost_native}, {status,waiting}, {initial_call,{inet_gethost_native,server_init,2}}, {backtrace, ["Program counter: 0xb75e9eac (inet_gethost_native:main_loop/1 + 20)", "CP: 0x00000000 (invalid)","arity = 0", "0xb7389574 Return addr 0x08229d74 ()", "y(0) {state,#Port<0.2440>,8000,32802,36899,<0.180.0>,4,{statistics,0,0,0,0,0,0,0,0}}"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,1}]}, {heap_size,377}, {total_heap_size,754}, {links,[<6517.180.0>,#Port<6517.2440>]}, {memory,3480}, {message_queue_len,0}, {reductions,132}, {trap_exit,true}]}, {<6517.189.0>, [{registered_name,net_sup}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb756951c (gen_server:loop/6 + 144)", "CP: 0x00000000 (invalid)","arity = 0", "0xb73b5718 Return addr 0xb753b298 (proc_lib:init_p_do_apply/3 + 28)", "y(0) []","y(1) infinity", "y(2) supervisor", "y(3) {state,{local,net_sup},one_for_all,[{child,<0.192.0>,net_kernel,{net_kernel,start", "y(4) net_sup","y(5) <0.11.0>", "0xb73b5734 Return addr 0x08229d74 ()", "y(0) Catch 0xb753b2a8 (proc_lib:init_p_do_apply/3 + 44)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,1}]}, {heap_size,377}, {total_heap_size,987}, {links, [<6517.190.0>,<6517.191.0>,<6517.192.0>,<6517.11.0>]}, {memory,4452}, {message_queue_len,0}, {reductions,244}, {trap_exit,true}]}, {<6517.190.0>, [{registered_name,erl_epmd}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb756951c (gen_server:loop/6 + 144)", "CP: 0x00000000 (invalid)","arity = 0", "0xb73f00b8 Return addr 0xb753b298 (proc_lib:init_p_do_apply/3 + 28)", "y(0) []","y(1) infinity", "y(2) erl_epmd", "y(3) {state,#Port<0.2481>,21100,ns_1}", "y(4) erl_epmd","y(5) <0.189.0>", "0xb73f00d4 Return addr 0x08229d74 ()", "y(0) Catch 0xb753b2a8 (proc_lib:init_p_do_apply/3 + 44)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,0}]}, {heap_size,233}, {total_heap_size,233}, {links,[<6517.189.0>,#Port<6517.2481>]}, {memory,1396}, {message_queue_len,0}, {reductions,127}, {trap_exit,false}]}, {<6517.191.0>, [{registered_name,auth}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb756951c (gen_server:loop/6 + 144)", "CP: 0x00000000 (invalid)","arity = 0", "0xb73dc5e0 Return addr 0xb753b298 (proc_lib:init_p_do_apply/3 + 28)", "y(0) []","y(1) infinity","y(2) auth", "y(3) {state,elyrdmyctqykhepq,53265}", "y(4) auth","y(5) <0.189.0>", "0xb73dc5fc Return addr 0x08229d74 ()", "y(0) Catch 0xb753b2a8 (proc_lib:init_p_do_apply/3 + 44)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,4}]}, {heap_size,233}, {total_heap_size,610}, {links,[<6517.189.0>]}, {memory,2884}, {message_queue_len,0}, {reductions,335}, {trap_exit,true}]}, {<6517.192.0>, [{registered_name,net_kernel}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb756951c (gen_server:loop/6 + 144)", "CP: 0x00000000 (invalid)","arity = 0", "0xb73fe250 Return addr 0xb753b298 (proc_lib:init_p_do_apply/3 + 28)", "y(0) []","y(1) infinity", "y(2) net_kernel", "y(3) {state,'ns_1@10.1.4.244','ns_1@10.1.4.244',longnames,{tick,<0.194.0>,15000},7000,", "y(4) net_kernel","y(5) <0.189.0>", "0xb73fe26c Return addr 0x08229d74 ()", "y(0) Catch 0xb753b2a8 (proc_lib:init_p_do_apply/3 + 44)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,8}]}, {heap_size,377}, {total_heap_size,754}, {links, [<6517.189.0>,<6517.194.0>,<6517.252.0>,<6517.193.0>, #Port<6517.2479>]}, {memory,3556}, {message_queue_len,0}, {reductions,914}, {trap_exit,true}]}, {<6517.193.0>, [{registered_name,[]}, {status,waiting}, {initial_call,{inet_tcp_dist,accept_loop,2}}, {backtrace, ["Program counter: 0xb750aa08 (prim_inet:accept0/2 + 92)", "CP: 0x00000000 (invalid)","arity = 0", "0xb738ebd0 Return addr 0xb6cd08ec (inet_tcp:accept/1 + 20)", "y(0) 32","y(1) #Port<0.2479>", "0xb738ebdc Return addr 0xb6cd2998 (inet_tcp_dist:accept_loop/2 + 48)", "y(0) []", "0xb738ebe4 Return addr 0x08229d74 ()", "y(0) []","y(1) #Port<0.2479>", "y(2) <0.192.0>"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,2}]}, {heap_size,233}, {total_heap_size,610}, {links,[<6517.192.0>]}, {memory,2864}, {message_queue_len,0}, {reductions,525}, {trap_exit,false}]}, {<6517.194.0>, [{registered_name,[]}, {status,waiting}, {initial_call,{net_kernel,ticker,2}}, {backtrace, ["Program counter: 0xb6c188fc (net_kernel:ticker_loop/2 + 28)", "CP: 0x00000000 (invalid)","arity = 0", "0xb6d88df0 Return addr 0x08229d74 ()", "y(0) 15000","y(1) <0.192.0>"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,0}]}, {heap_size,233}, {total_heap_size,233}, {links,[<6517.192.0>]}, {memory,1316}, {message_queue_len,0}, {reductions,43}, {trap_exit,false}]}, {<6517.252.0>, [{registered_name,[]}, {status,waiting}, {initial_call,{inet_tcp_dist,do_setup,6}}, {backtrace, ["Program counter: 0xb66ed4e4 (dist_util:con_loop/9 + 72)", "CP: 0x00000000 (invalid)","arity = 0", "0xb6d85498 Return addr 0x08229d74 ()", "y(0) []", "y(1) #Fun", "y(2) #Fun", "y(3) {tick,2315,1141,0,4}","y(4) normal", "y(5) 'ns_1@10.1.4.244'", "y(6) {net_address,{{10,1,4,243},21100},\"10.1.4.243\",tcp,inet}", "y(7) #Port<0.2559>", "y(8) 'ns_1@10.1.4.243'","y(9) <0.192.0>"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,5}]}, {heap_size,987}, {total_heap_size,1364}, {links,[<6517.192.0>,#Port<6517.2559>]}, {memory,5860}, {message_queue_len,0}, {reductions,1020}, {trap_exit,false}]}, {<6517.255.0>, [{registered_name,ns_server_sup}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb756951c (gen_server:loop/6 + 144)", "CP: 0x00000000 (invalid)","arity = 0", "0xb6df0a70 Return addr 0xb753b298 (proc_lib:init_p_do_apply/3 + 28)", "y(0) []","y(1) infinity", "y(2) supervisor", "y(3) {state,{local,ns_server_sup},one_for_one,[{child,<0.347.0>,ns_tick,{ns_tick,start", "y(4) ns_server_sup","y(5) <0.54.0>", "0xb6df0a8c Return addr 0x08229d74 ()", "y(0) Catch 0xb753b2a8 (proc_lib:init_p_do_apply/3 + 44)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,1}]}, {heap_size,28657}, {total_heap_size,225075}, {links, [<6517.285.0>,<6517.304.0>,<6517.344.0>,<6517.346.0>, <6517.347.0>,<6517.345.0>,<6517.310.0>,<6517.311.0>, <6517.305.0>,<6517.302.0>,<6517.303.0>,<6517.296.0>, <6517.258.0>,<6517.277.0>,<6517.278.0>,<6517.261.0>, <6517.256.0>,<6517.257.0>,<6517.54.0>]}, {memory,901104}, {message_queue_len,0}, {reductions,99106}, {trap_exit,true}]}, {<6517.256.0>, [{registered_name,ns_log}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb756951c (gen_server:loop/6 + 144)", "CP: 0x00000000 (invalid)","arity = 0", "0xb73fabc8 Return addr 0xb753b298 (proc_lib:init_p_do_apply/3 + 28)", "y(0) []","y(1) infinity","y(2) ns_log", "y(3) {state,[{log_entry,{1304,814867,278433},'ns_1@127.0.0.1',ns_node_disco,3,\"Initial", "y(4) ns_log","y(5) <0.255.0>", "0xb73fabe4 Return addr 0x08229d74 ()", "y(0) Catch 0xb753b2a8 (proc_lib:init_p_do_apply/3 + 44)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,2}]}, {heap_size,4181}, {total_heap_size,10946}, {links,[<6517.50.0>,<6517.255.0>]}, {memory,44248}, {message_queue_len,0}, {reductions,1738}, {trap_exit,false}]}, {<6517.257.0>, [{registered_name,ns_log_events}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0x08229d70 (unknown function)", "CP: 0x08229d74 ()", "arity = 3"," proc_lib"," wake_up", " [gen_event,wake_hib,[<0.255.0>,ns_log_events,[{handler,ns_mail_log,false,{state},false}"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,0}]}, {heap_size,34}, {total_heap_size,34}, {links,[<6517.255.0>]}, {memory,580}, {message_queue_len,0}, {reductions,283}, {trap_exit,true}]}, {<6517.258.0>, [{registered_name,ns_mail_sup}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb756951c (gen_server:loop/6 + 144)", "CP: 0x00000000 (invalid)","arity = 0", "0xb73fd2d8 Return addr 0xb753b298 (proc_lib:init_p_do_apply/3 + 28)", "y(0) []","y(1) infinity", "y(2) supervisor", "y(3) {state,{local,ns_mail_sup},one_for_all,[{child,undefined,ns_mail_log,{ns_mail_log", "y(4) ns_mail_sup","y(5) <0.255.0>", "0xb73fd2f4 Return addr 0x08229d74 ()", "y(0) Catch 0xb753b2a8 (proc_lib:init_p_do_apply/3 + 44)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,1}]}, {heap_size,233}, {total_heap_size,610}, {links,[<6517.259.0>,<6517.255.0>]}, {memory,2904}, {message_queue_len,0}, {reductions,652}, {trap_exit,true}]}, {<6517.259.0>, [{registered_name,ns_mail}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb756951c (gen_server:loop/6 + 144)", "CP: 0x00000000 (invalid)","arity = 0", "0xb7406c20 Return addr 0xb753b298 (proc_lib:init_p_do_apply/3 + 28)", "y(0) []","y(1) infinity","y(2) ns_mail", "y(3) empty_state","y(4) ns_mail", "y(5) <0.258.0>", "0xb7406c3c Return addr 0x08229d74 ()", "y(0) Catch 0xb753b2a8 (proc_lib:init_p_do_apply/3 + 44)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,0}]}, {heap_size,233}, {total_heap_size,233}, {links,[<6517.258.0>]}, {memory,1376}, {message_queue_len,0}, {reductions,27}, {trap_exit,true}]}, {<6517.261.0>, [{registered_name,ns_node_disco_sup}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb756951c (gen_server:loop/6 + 144)", "CP: 0x00000000 (invalid)","arity = 0", "0xb7399ff8 Return addr 0xb753b298 (proc_lib:init_p_do_apply/3 + 28)", "y(0) []","y(1) infinity", "y(2) supervisor", "y(3) {state,{local,ns_node_disco_sup},rest_for_one,[{child,<0.270.0>,ns_config_rep,{ns", "y(4) ns_node_disco_sup","y(5) <0.255.0>", "0xb739a014 Return addr 0x08229d74 ()", "y(0) Catch 0xb753b2a8 (proc_lib:init_p_do_apply/3 + 44)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,4}]}, {heap_size,233}, {total_heap_size,610}, {links, [<6517.264.0>,<6517.270.0>,<6517.263.0>,<6517.255.0>]}, {memory,2944}, {message_queue_len,0}, {reductions,802}, {trap_exit,true}]}, {<6517.263.0>, [{registered_name,ns_node_disco_events}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb753125c (gen_event:fetch_msg/5 + 44)", "CP: 0x00000000 (invalid)","arity = 0", "0xb73a3474 Return addr 0xb753b308 (proc_lib:wake_up/3 + 60)", "y(0) false","y(1) []", "y(2) [{handler,menelaus_event,ns_node_disco_events,{state,ns_node_disco_events,undefin", "y(3) ns_node_disco_events","y(4) <0.261.0>", "0xb73a348c Return addr 0x08229d74 ()", "y(0) Catch 0xb753b318 (proc_lib:wake_up/3 + 76)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,1}]}, {heap_size,233}, {total_heap_size,322}, {links,[<6517.261.0>]}, {memory,1772}, {message_queue_len,0}, {reductions,60}, {trap_exit,true}]}, {<6517.264.0>, [{registered_name,ns_node_disco}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb756951c (gen_server:loop/6 + 144)", "CP: 0x00000000 (invalid)","arity = 0", "0xb6d71768 Return addr 0xb753b298 (proc_lib:init_p_do_apply/3 + 28)", "y(0) []","y(1) infinity", "y(2) ns_node_disco", "y(3) {state,false,{interval,#Ref<0.0.0.2321>}}", "y(4) ns_node_disco","y(5) <0.261.0>", "0xb6d71784 Return addr 0x08229d74 ()", "y(0) Catch 0xb753b2a8 (proc_lib:init_p_do_apply/3 + 44)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,88}]}, {heap_size,17711}, {total_heap_size,46368}, {links,[<6517.261.0>,<6517.50.0>]}, {memory,185936}, {message_queue_len,0}, {reductions,3156}, {trap_exit,false}]}, {<6517.270.0>, [{registered_name,ns_config_rep}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb756951c (gen_server:loop/6 + 144)", "CP: 0x00000000 (invalid)","arity = 0", "0xb6308dbc Return addr 0xb753b298 (proc_lib:init_p_do_apply/3 + 28)", "y(0) []","y(1) infinity", "y(2) ns_config_rep","y(3) {state}", "y(4) ns_config_rep","y(5) <0.261.0>", "0xb6308dd8 Return addr 0x08229d74 ()", "y(0) Catch 0xb753b2a8 (proc_lib:init_p_do_apply/3 + 44)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,10}]}, {heap_size,46368}, {total_heap_size,75025}, {links,[<6517.261.0>]}, {memory,300544}, {message_queue_len,0}, {reductions,7867}, {trap_exit,false}]}, {<6517.277.0>, [{registered_name,ns_heart}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb756951c (gen_server:loop/6 + 144)", "CP: 0x00000000 (invalid)","arity = 0", "0xb6d78120 Return addr 0xb753b298 (proc_lib:init_p_do_apply/3 + 28)", "y(0) []","y(1) infinity", "y(2) ns_heart", "y(3) [{meminfo,<<1248 bytes>>},{replication,[{\"default\",1.000000e+00}]},{system_memory", "y(4) ns_heart","y(5) <0.255.0>", "0xb6d7813c Return addr 0x08229d74 ()", "y(0) Catch 0xb753b2a8 (proc_lib:init_p_do_apply/3 + 44)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,1}]}, {heap_size,6765}, {total_heap_size,17711}, {links,[<6517.255.0>,<6517.50.0>]}, {memory,71308}, {message_queue_len,0}, {reductions,557513}, {trap_exit,false}]}, {<6517.278.0>, [{registered_name,ns_doctor}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb756951c (gen_server:loop/6 + 144)", "CP: 0x00000000 (invalid)","arity = 0", "0xb6da4db0 Return addr 0xb753b298 (proc_lib:init_p_do_apply/3 + 28)", "y(0) []","y(1) infinity", "y(2) ns_doctor", "y(3) {state,{dict,2,16,16,8,80,48,{[],[],[],[],[],[],[],[],[],[],[],[],[],[],[],[]},{{", "y(4) ns_doctor","y(5) <0.255.0>", "0xb6da4dcc Return addr 0x08229d74 ()", "y(0) Catch 0xb753b2a8 (proc_lib:init_p_do_apply/3 + 44)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,0}]}, {heap_size,6765}, {total_heap_size,6765}, {links,[<6517.255.0>,<6517.50.0>]}, {memory,27524}, {message_queue_len,0}, {reductions,17439}, {trap_exit,false}]}, {<6517.285.0>, [{registered_name,menelaus_sup}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb756951c (gen_server:loop/6 + 144)", "CP: 0x00000000 (invalid)","arity = 0", "0xb6d5d0d0 Return addr 0xb753b298 (proc_lib:init_p_do_apply/3 + 28)", "y(0) []","y(1) infinity", "y(2) supervisor", "y(3) {state,{local,menelaus_sup},one_for_one,[{child,<0.418.0>,hot_keys_keeper,{hot_ke", "y(4) menelaus_sup","y(5) <0.255.0>", "0xb6d5d0ec Return addr 0x08229d74 ()", "y(0) Catch 0xb753b2a8 (proc_lib:init_p_do_apply/3 + 44)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,1}]}, {heap_size,4181}, {total_heap_size,32838}, {links,[<6517.288.0>,<6517.418.0>,<6517.255.0>]}, {memory,131836}, {message_queue_len,0}, {reductions,1875}, {trap_exit,true}]}, {<6517.288.0>, [{registered_name,menelaus_web}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb756951c (gen_server:loop/6 + 144)", "CP: 0x00000000 (invalid)","arity = 0", "0xb73be5b0 Return addr 0xb753b298 (proc_lib:init_p_do_apply/3 + 28)", "y(0) []","y(1) infinity", "y(2) mochiweb_socket_server", "y(3) {mochiweb_socket_server,8091,#Fun,{local,menelaus_web}", "y(4) menelaus_web","y(5) <0.285.0>", "0xb73be5cc Return addr 0x08229d74 ()", "y(0) Catch 0xb753b2a8 (proc_lib:init_p_do_apply/3 + 44)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,1}]}, {heap_size,377}, {total_heap_size,987}, {links, [<6517.285.0>,<6517.289.0>,<6517.340.0>, #Port<6517.2568>]}, {memory,4452}, {message_queue_len,0}, {reductions,331}, {trap_exit,true}]}, {<6517.289.0>, [{registered_name,[]}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb6a3de20 (menelaus_web:handle_streaming/4 + 196)", "CP: 0x00000000 (invalid)","arity = 0", "0xb6858450 Return addr 0xb6a3c178 (menelaus_web:loop/3 + 12136)", "y(0) {struct,[{buckets,[{struct,[{name,<<7 bytes>>},{nodeLocator,vbucket},{saslPasswor", "y(1) {struct,[{buckets,[{struct,[{name,<<7 bytes>>},{nodeLocator,vbucket},{saslPasswor", "y(2) {mochiweb_response,{mochiweb_request,#Port<0.2630>,'GET',\"/pools/default/saslBuck", "y(3) {mochiweb_request,#Port<0.2630>,'GET',\"/pools/default/saslBucketsStreaming\",{1,1}", "y(4) #Fun", "0xb6858468 Return addr 0xb6a48ce8 (mochiweb_http:headers/5 + 680)", "y(0) []","y(1) []","y(2) []", "y(3) []", "y(4) {mochiweb_request,#Port<0.2630>,'GET',\"/pools/default/saslBucketsStreaming\",{1,1}", "y(5) Catch 0xb6a3c188 (menelaus_web:loop/3 + 12152)", "0xb6858484 Return addr 0xb753b298 (proc_lib:init_p_do_apply/3 + 28)", "y(0) #Fun", "y(1) []","y(2) []", "y(3) {mochiweb_request,#Port<0.2630>,'GET',\"/pools/default/saslBucketsStreaming\",{1,1}", "0xb6858498 Return addr 0x08229d74 ()", "y(0) Catch 0xb753b2a8 (proc_lib:init_p_do_apply/3 + 44)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,2}]}, {heap_size,10946}, {total_heap_size,85971}, {links,[<6517.288.0>,#Port<6517.2630>]}, {memory,344428}, {message_queue_len,0}, {reductions,999032}, {trap_exit,false}]}, {<6517.296.0>, [{registered_name,ns_port_sup}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb756951c (gen_server:loop/6 + 144)", "CP: 0x00000000 (invalid)","arity = 0", "0xb68b9740 Return addr 0xb753b298 (proc_lib:init_p_do_apply/3 + 28)", "y(0) []","y(1) infinity", "y(2) supervisor", "y(3) {state,{local,ns_port_sup},one_for_one,[{child,<0.351.0>,{memcached,\"./bin/memcac", "y(4) ns_port_sup","y(5) <0.255.0>", "0xb68b975c Return addr 0x08229d74 ()", "y(0) Catch 0xb753b2a8 (proc_lib:init_p_do_apply/3 + 44)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,1}]}, {heap_size,28657}, {total_heap_size,75025}, {links,[<6517.298.0>,<6517.351.0>,<6517.255.0>]}, {memory,300584}, {message_queue_len,0}, {reductions,3024}, {trap_exit,true}]}, {<6517.298.0>, [{registered_name,[]}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb756951c (gen_server:loop/6 + 144)", "CP: 0x00000000 (invalid)","arity = 0", "0xb6dab018 Return addr 0xb753b298 (proc_lib:init_p_do_apply/3 + 28)", "y(0) []","y(1) infinity", "y(2) supervisor_cushion", "y(3) {state,moxi,5000,{1304,814959,13780},<0.299.0>}", "y(4) <0.298.0>","y(5) <0.296.0>", "0xb6dab034 Return addr 0x08229d74 ()", "y(0) Catch 0xb753b2a8 (proc_lib:init_p_do_apply/3 + 44)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,1}]}, {heap_size,987}, {total_heap_size,2584}, {links,[<6517.296.0>,<6517.299.0>]}, {memory,10800}, {message_queue_len,0}, {reductions,55}, {trap_exit,true}]}, {<6517.299.0>, [{registered_name,[]}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb756951c (gen_server:loop/6 + 144)", "CP: 0x00000000 (invalid)","arity = 0", "0xb693922c Return addr 0xb753b298 (proc_lib:init_p_do_apply/3 + 28)", "y(0) []","y(1) infinity", "y(2) ns_port_server", "y(3) {state,#Port<0.2569>,moxi,{[\"2011-05-07 17:35:59: (cproxy_config.c.325) env: MOXI", "y(4) <0.299.0>","y(5) <0.298.0>", "0xb6939248 Return addr 0x08229d74 ()", "y(0) Catch 0xb753b2a8 (proc_lib:init_p_do_apply/3 + 44)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,1}]}, {heap_size,2584}, {total_heap_size,5168}, {links,[<6517.298.0>,#Port<6517.2569>]}, {memory,21136}, {message_queue_len,0}, {reductions,305}, {trap_exit,true}]}, {<6517.302.0>, [{registered_name,ns_tick_event}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb753125c (gen_event:fetch_msg/5 + 44)", "CP: 0x00000000 (invalid)","arity = 0", "0xb6d18ad8 Return addr 0xb753b298 (proc_lib:init_p_do_apply/3 + 28)", "y(0) false","y(1) []", "y(2) [{handler,ns_pubsub,#Ref<0.0.0.3191>,{state,#Fun,ignored},<", "y(3) ns_tick_event","y(4) <0.255.0>", "0xb6d18af0 Return addr 0x08229d74 ()", "y(0) Catch 0xb753b2a8 (proc_lib:init_p_do_apply/3 + 44)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,18}]}, {heap_size,610}, {total_heap_size,987}, {links,[<6517.255.0>,<6517.368.0>]}, {memory,4412}, {message_queue_len,0}, {reductions,3367}, {trap_exit,true}]}, {<6517.303.0>, [{registered_name,ns_stats_event}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb753125c (gen_event:fetch_msg/5 + 44)", "CP: 0x00000000 (invalid)","arity = 0", "0xb73fcf34 Return addr 0xb753b298 (proc_lib:init_p_do_apply/3 + 28)", "y(0) false","y(1) []", "y(2) [{handler,ns_pubsub,#Ref<0.0.0.3449>,{state,#Fun,ignored},<", "y(3) ns_stats_event","y(4) <0.255.0>", "0xb73fcf4c Return addr 0x08229d74 ()", "y(0) Catch 0xb753b2a8 (proc_lib:init_p_do_apply/3 + 44)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,581}]}, {heap_size,377}, {total_heap_size,987}, {links,[<6517.255.0>,<6517.379.0>]}, {memory,4412}, {message_queue_len,0}, {reductions,6431}, {trap_exit,true}]}, {<6517.304.0>, [{registered_name,ns_good_bucket_worker}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb756951c (gen_server:loop/6 + 144)", "CP: 0x00000000 (invalid)","arity = 0", "0xb73cbe30 Return addr 0xb753b298 (proc_lib:init_p_do_apply/3 + 28)", "y(0) []","y(1) infinity", "y(2) work_queue","y(3) []", "y(4) ns_good_bucket_worker", "y(5) <0.255.0>", "0xb73cbe4c Return addr 0x08229d74 ()", "y(0) Catch 0xb753b2a8 (proc_lib:init_p_do_apply/3 + 44)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,0}]}, {heap_size,233}, {total_heap_size,233}, {links,[<6517.255.0>]}, {memory,1376}, {message_queue_len,0}, {reductions,95}, {trap_exit,false}]}, {<6517.305.0>, [{registered_name,ns_good_bucket_sup}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb756951c (gen_server:loop/6 + 144)", "CP: 0x00000000 (invalid)","arity = 0", "0xb688c2b8 Return addr 0xb753b298 (proc_lib:init_p_do_apply/3 + 28)", "y(0) []","y(1) infinity", "y(2) supervisor", "y(3) {state,{local,ns_good_bucket_sup},one_for_one,[{child,<0.374.0>,{ns_memcached,\"de", "y(4) ns_good_bucket_sup","y(5) <0.255.0>", "0xb688c2d4 Return addr 0x08229d74 ()", "y(0) Catch 0xb753b2a8 (proc_lib:init_p_do_apply/3 + 44)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,0}]}, {heap_size,17711}, {total_heap_size,17711}, {links, [<6517.255.0>,<6517.373.0>,<6517.374.0>,<6517.64.0>]}, {memory,71348}, {message_queue_len,0}, {reductions,214}, {trap_exit,true}]}, {<6517.310.0>, [{registered_name,[]}, {status,waiting}, {initial_call,{erlang,apply,2}}, {backtrace, ["Program counter: 0xb6cc655c (misc:wait_for_process/2 + 104)", "CP: 0x00000000 (invalid)","arity = 0", "0xb73936c0 Return addr 0xb6cc92e0 (misc:'-start_singleton/4-fun-0-'/2 + 56)", "y(0) []","y(1) #Ref<0.0.0.2704>", "y(2) infinity", "0xb73936d0 Return addr 0x08229d74 ()", "y(0) <10111.113.0>","y(1) ns_orchestrator"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,0}]}, {heap_size,233}, {total_heap_size,233}, {links,[<6517.255.0>]}, {memory,1356}, {message_queue_len,0}, {reductions,6}, {trap_exit,false}]}, {<6517.311.0>, [{registered_name,ns_mnesia}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb756951c (gen_server:loop/6 + 144)", "CP: 0x00000000 (invalid)","arity = 0", "0xb6d30f34 Return addr 0xb753b298 (proc_lib:init_p_do_apply/3 + 28)", "y(0) []","y(1) infinity", "y(2) ns_mnesia","y(3) {state}", "y(4) ns_mnesia","y(5) <0.255.0>", "0xb6d30f50 Return addr 0x08229d74 ()", "y(0) Catch 0xb753b2a8 (proc_lib:init_p_do_apply/3 + 44)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,6}]}, {heap_size,2584}, {total_heap_size,3194}, {links,[<6517.320.0>,<6517.255.0>]}, {memory,13240}, {message_queue_len,0}, {reductions,1411}, {trap_exit,true}]}, {<6517.312.0>, [{registered_name,[]}, {status,waiting}, {initial_call,{erlang,apply,2}}, {backtrace, ["Program counter: 0xb6cc9728 (misc:'-wait_for_process/2-fun-0-'/3 + 112)", "CP: 0x00000000 (invalid)","arity = 0", "0xb73d2454 Return addr 0x08229d74 ()", "y(0) []","y(1) <0.310.0>", "y(2) #Ref<0.0.0.2704>"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,0}]}, {heap_size,233}, {total_heap_size,233}, {links,[<0.113.0>]}, {memory,1372}, {message_queue_len,0}, {reductions,13}, {trap_exit,true}]}, {<6517.314.0>, [{registered_name,[]}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb759c3b0 (application_master:main_loop/2 + 32)", "CP: 0x00000000 (invalid)","arity = 0", "0xb73e474c Return addr 0xb753b298 (proc_lib:init_p_do_apply/3 + 28)", "y(0) []", "y(1) {state,<0.315.0>,{appl_data,mnesia,[mnesia_dumper_load_regulator,mnesia_event,mne", "y(2) <0.7.0>", "0xb73e475c Return addr 0x08229d74 ()", "y(0) Catch 0xb753b2a8 (proc_lib:init_p_do_apply/3 + 44)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,5}]}, {heap_size,233}, {total_heap_size,610}, {links,[<6517.7.0>,<6517.315.0>]}, {memory,2904}, {message_queue_len,0}, {reductions,45}, {trap_exit,true}]}, {<6517.315.0>, [{registered_name,[]}, {status,waiting}, {initial_call,{application_master,start_it,4}}, {backtrace, ["Program counter: 0xb759d6f4 (application_master:loop_it/4 + 40)", "CP: 0x00000000 (invalid)","arity = 0", "0xb73cab28 Return addr 0x08229d74 ()", "y(0) {normal,[]}","y(1) mnesia_sup", "y(2) <0.316.0>","y(3) <0.314.0>"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,0}]}, {heap_size,233}, {total_heap_size,233}, {links,[<6517.314.0>,<6517.316.0>]}, {memory,1336}, {message_queue_len,0}, {reductions,32}, {trap_exit,true}]}, {<6517.316.0>, [{registered_name,mnesia_sup}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb756951c (gen_server:loop/6 + 144)", "CP: 0x00000000 (invalid)","arity = 0", "0xb73eea00 Return addr 0xb753b298 (proc_lib:init_p_do_apply/3 + 28)", "y(0) []","y(1) infinity", "y(2) supervisor", "y(3) {state,{local,mnesia_sup},one_for_all,[{child,<0.318.0>,mnesia_kernel_sup,{mnesia", "y(4) mnesia_sup","y(5) <0.315.0>", "0xb73eea1c Return addr 0x08229d74 ()", "y(0) Catch 0xb753b2a8 (proc_lib:init_p_do_apply/3 + 44)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,1}]}, {heap_size,377}, {total_heap_size,754}, {links,[<6517.317.0>,<6517.318.0>,<6517.315.0>]}, {memory,3500}, {message_queue_len,0}, {reductions,191}, {trap_exit,true}]}, {<6517.317.0>, [{registered_name,mnesia_event}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb753125c (gen_event:fetch_msg/5 + 44)", "CP: 0x00000000 (invalid)","arity = 0", "0xb7390cf0 Return addr 0xb753b298 (proc_lib:init_p_do_apply/3 + 28)", "y(0) false","y(1) []", "y(2) [{handler,mnesia_event,false,{state,[],false,[]},false}]", "y(3) mnesia_event","y(4) <0.316.0>", "0xb7390d08 Return addr 0x08229d74 ()", "y(0) Catch 0xb753b2a8 (proc_lib:init_p_do_apply/3 + 44)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,7}]}, {heap_size,610}, {total_heap_size,987}, {links,[<6517.316.0>,<6517.320.0>]}, {memory,4412}, {message_queue_len,0}, {reductions,402}, {trap_exit,true}]}, {<6517.318.0>, [{registered_name,mnesia_kernel_sup}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb756951c (gen_server:loop/6 + 144)", "CP: 0x00000000 (invalid)","arity = 0", "0xb738e824 Return addr 0xb753b298 (proc_lib:init_p_do_apply/3 + 28)", "y(0) []","y(1) infinity", "y(2) supervisor", "y(3) {state,{local,mnesia_kernel_sup},one_for_all,[{child,<0.327.0>,mnesia_late_loader", "y(4) mnesia_kernel_sup","y(5) <0.316.0>", "0xb738e840 Return addr 0x08229d74 ()", "y(0) Catch 0xb753b2a8 (proc_lib:init_p_do_apply/3 + 44)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,4}]}, {heap_size,610}, {total_heap_size,987}, {links, [<6517.321.0>,<6517.325.0>,<6517.326.0>,<6517.327.0>, <6517.323.0>,<6517.324.0>,<6517.322.0>,<6517.319.0>, <6517.320.0>,<6517.316.0>]}, {memory,4572}, {message_queue_len,0}, {reductions,545}, {trap_exit,true}]}, {<6517.319.0>, [{registered_name,mnesia_monitor}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb756951c (gen_server:loop/6 + 144)", "CP: 0x00000000 (invalid)","arity = 0", "0xb6d2e6cc Return addr 0xb753b298 (proc_lib:init_p_do_apply/3 + 28)", "y(0) []","y(1) infinity", "y(2) mnesia_monitor", "y(3) {state,<0.318.0>,[],[],true,[],undefined,[]}", "y(4) mnesia_monitor","y(5) <0.318.0>", "0xb6d2e6e8 Return addr 0x08229d74 ()", "y(0) Catch 0xb753b2a8 (proc_lib:init_p_do_apply/3 + 44)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,11}]}, {heap_size,2584}, {total_heap_size,3194}, {links,[<6517.330.0>,<6517.318.0>]}, {memory,13240}, {message_queue_len,0}, {reductions,3394}, {trap_exit,true}]}, {<6517.320.0>, [{registered_name,mnesia_subscr}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb756951c (gen_server:loop/6 + 144)", "CP: 0x00000000 (invalid)","arity = 0", "0xb7398930 Return addr 0xb753b298 (proc_lib:init_p_do_apply/3 + 28)", "y(0) []","y(1) infinity", "y(2) mnesia_subscr", "y(3) {state,<0.318.0>,77850}", "y(4) mnesia_subscr","y(5) <0.318.0>", "0xb739894c Return addr 0x08229d74 ()", "y(0) Catch 0xb753b2a8 (proc_lib:init_p_do_apply/3 + 44)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,0}]}, {heap_size,233}, {total_heap_size,233}, {links,[<6517.317.0>,<6517.318.0>,<6517.311.0>]}, {memory,1416}, {message_queue_len,0}, {reductions,111}, {trap_exit,true}]}, {<6517.321.0>, [{registered_name,mnesia_locker}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb6ad7da8 (mnesia_locker:loop/1 + 20)", "CP: 0x00000000 (invalid)","arity = 0", "0xb6a1117c Return addr 0xb6acda38 (mnesia_sp:init_proc/4 + 132)", "y(0) []","y(1) []","y(2) []", "y(3) []","y(4) []", "y(5) {state,<0.318.0>}", "0xb6a11198 Return addr 0xb753b298 (proc_lib:init_p_do_apply/3 + 28)", "y(0) Catch 0xb6acda38 (mnesia_sp:init_proc/4 + 132)", "y(1) mnesia_locker","y(2) []","y(3) []", "y(4) [<0.318.0>]", "0xb6a111b0 Return addr 0x08229d74 ()", "y(0) Catch 0xb753b2a8 (proc_lib:init_p_do_apply/3 + 44)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,67}]}, {heap_size,2584}, {total_heap_size,2961}, {links,[<6517.318.0>]}, {memory,12288}, {message_queue_len,0}, {reductions,43568}, {trap_exit,true}]}, {<6517.322.0>, [{registered_name,mnesia_recover}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb756951c (gen_server:loop/6 + 144)", "CP: 0x00000000 (invalid)","arity = 0", "0xb73fc948 Return addr 0xb753b298 (proc_lib:init_p_do_apply/3 + 28)", "y(0) []","y(1) infinity", "y(2) mnesia_recover", "y(3) {state,<0.318.0>,undefined,undefined,undefined,0,true,[]}", "y(4) mnesia_recover","y(5) <0.318.0>", "0xb73fc964 Return addr 0x08229d74 ()", "y(0) Catch 0xb753b2a8 (proc_lib:init_p_do_apply/3 + 44)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,5}]}, {heap_size,377}, {total_heap_size,754}, {links,[<6517.318.0>,<6517.50.0>]}, {memory,3480}, {message_queue_len,0}, {reductions,970}, {trap_exit,true}]}, {<6517.323.0>, [{registered_name,mnesia_tm}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb6affdac (mnesia_tm:doit_loop/1 + 108)", "CP: 0x00000000 (invalid)","arity = 0", "0xb73ee3f4 Return addr 0xb6acda38 (mnesia_sp:init_proc/4 + 132)", "y(0) []","y(1) []", "y(2) {state,{0,nil},{0,nil},<0.318.0>,[],[],[]}", "y(3) []","y(4) []","y(5) <0.318.0>", "y(6) {0,nil}","y(7) {0,nil}", "0xb73ee418 Return addr 0xb753b298 (proc_lib:init_p_do_apply/3 + 28)", "y(0) Catch 0xb6acda38 (mnesia_sp:init_proc/4 + 132)", "y(1) mnesia_tm","y(2) []","y(3) []", "y(4) [<0.318.0>]", "0xb73ee430 Return addr 0x08229d74 ()", "y(0) Catch 0xb753b2a8 (proc_lib:init_p_do_apply/3 + 44)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,99}]}, {heap_size,610}, {total_heap_size,987}, {links,[<6517.318.0>]}, {memory,4392}, {message_queue_len,0}, {reductions,101693}, {trap_exit,true}]}, {<6517.324.0>, [{registered_name,mnesia_checkpoint_sup}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb756951c (gen_server:loop/6 + 144)", "CP: 0x00000000 (invalid)","arity = 0", "0xb73caec0 Return addr 0xb753b298 (proc_lib:init_p_do_apply/3 + 28)", "y(0) []","y(1) infinity", "y(2) supervisor", "y(3) {state,{local,mnesia_checkpoint_sup},simple_one_for_one,[{child,undefined,mnesia_", "y(4) mnesia_checkpoint_sup", "y(5) <0.318.0>", "0xb73caedc Return addr 0x08229d74 ()", "y(0) Catch 0xb753b2a8 (proc_lib:init_p_do_apply/3 + 44)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,0}]}, {heap_size,233}, {total_heap_size,233}, {links,[<6517.318.0>]}, {memory,1376}, {message_queue_len,0}, {reductions,61}, {trap_exit,true}]}, {<6517.325.0>, [{registered_name,mnesia_snmp_sup}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb756951c (gen_server:loop/6 + 144)", "CP: 0x00000000 (invalid)","arity = 0", "0xb73e5478 Return addr 0xb753b298 (proc_lib:init_p_do_apply/3 + 28)", "y(0) []","y(1) infinity", "y(2) supervisor", "y(3) {state,{local,mnesia_snmp_sup},simple_one_for_one,[{child,undefined,mnesia_snmp_s", "y(4) mnesia_snmp_sup","y(5) <0.318.0>", "0xb73e5494 Return addr 0x08229d74 ()", "y(0) Catch 0xb753b2a8 (proc_lib:init_p_do_apply/3 + 44)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,0}]}, {heap_size,233}, {total_heap_size,233}, {links,[<6517.318.0>]}, {memory,1376}, {message_queue_len,0}, {reductions,61}, {trap_exit,true}]}, {<6517.326.0>, [{registered_name,mnesia_controller}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb756951c (gen_server:loop/6 + 144)", "CP: 0x00000000 (invalid)","arity = 0", "0xb6d24d88 Return addr 0xb753b298 (proc_lib:init_p_do_apply/3 + 28)", "y(0) []","y(1) infinity", "y(2) mnesia_controller", "y(3) {state,<0.318.0>,true,[],[],{0,nil},[],[],{0,nil},undefined,[],[],{interval,#Ref<", "y(4) mnesia_controller","y(5) <0.318.0>", "0xb6d24da4 Return addr 0x08229d74 ()", "y(0) Catch 0xb753b2a8 (proc_lib:init_p_do_apply/3 + 44)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,3}]}, {heap_size,377}, {total_heap_size,754}, {links,[<6517.318.0>,<6517.50.0>]}, {memory,3480}, {message_queue_len,0}, {reductions,465}, {trap_exit,true}]}, {<6517.327.0>, [{registered_name,mnesia_late_loader}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb66963fc (mnesia_late_loader:loop/1 + 20)", "CP: 0x00000000 (invalid)","arity = 0", "0xb73e4adc Return addr 0xb6acda38 (mnesia_sp:init_proc/4 + 132)", "y(0) []","y(1) []", "y(2) {state,<0.318.0>}", "0xb73e4aec Return addr 0xb753b298 (proc_lib:init_p_do_apply/3 + 28)", "y(0) Catch 0xb6acda38 (mnesia_sp:init_proc/4 + 132)", "y(1) mnesia_late_loader","y(2) []", "y(3) []","y(4) [<0.318.0>]", "0xb73e4b04 Return addr 0x08229d74 ()", "y(0) Catch 0xb753b2a8 (proc_lib:init_p_do_apply/3 + 44)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,1}]}, {heap_size,233}, {total_heap_size,610}, {links,[<6517.318.0>]}, {memory,2884}, {message_queue_len,0}, {reductions,180}, {trap_exit,false}]}, {<6517.330.0>, [{registered_name,[]}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb66b23a8 (disk_log:loop/1 + 84)", "CP: 0x00000000 (invalid)","arity = 0", "0xb6d23254 Return addr 0xb753b298 (proc_lib:init_p_do_apply/3 + 28)", "y(0) {state,[],[],<0.127.0>,<0.128.0>,325,{arg,latest_log,undefined,\"/var/opt/membase/", "0xb6d2325c Return addr 0x08229d74 ()", "y(0) Catch 0xb753b2a8 (proc_lib:init_p_do_apply/3 + 44)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,1}]}, {heap_size,987}, {total_heap_size,2584}, {links, [<6517.128.0>,<6517.319.0>,<6517.127.0>, #Port<6517.2797>]}, {memory,10840}, {message_queue_len,0}, {reductions,33597}, {trap_exit,true}]}, {<6517.340.0>, [{registered_name,[]}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb750aa08 (prim_inet:accept0/2 + 92)", "CP: 0x00000000 (invalid)","arity = 0", "0xb6d23e08 Return addr 0xb6cd08ec (inet_tcp:accept/1 + 20)", "y(0) 42","y(1) #Port<0.2568>", "0xb6d23e14 Return addr 0xb6a5004c (mochiweb_socket_server:acceptor_loop/1 + 80)", "y(0) []", "0xb6d23e1c Return addr 0xb753b298 (proc_lib:init_p_do_apply/3 + 28)", "y(0) #Fun", "y(1) <0.288.0>", "y(2) Catch 0xb6a5004c (mochiweb_socket_server:acceptor_loop/1 + 80)", "0xb6d23e2c Return addr 0x08229d74 ()", "y(0) Catch 0xb753b2a8 (proc_lib:init_p_do_apply/3 + 44)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,1}]}, {heap_size,377}, {total_heap_size,754}, {links,[<6517.288.0>]}, {memory,3500}, {message_queue_len,0}, {reductions,39}, {trap_exit,false}]}, {<6517.344.0>, [{registered_name,ns_bad_bucket_worker}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb756951c (gen_server:loop/6 + 144)", "CP: 0x00000000 (invalid)","arity = 0", "0xb6d325f0 Return addr 0xb753b298 (proc_lib:init_p_do_apply/3 + 28)", "y(0) []","y(1) infinity", "y(2) work_queue","y(3) []", "y(4) ns_bad_bucket_worker","y(5) <0.255.0>", "0xb6d3260c Return addr 0x08229d74 ()", "y(0) Catch 0xb753b2a8 (proc_lib:init_p_do_apply/3 + 44)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,1}]}, {heap_size,233}, {total_heap_size,610}, {links,[<6517.255.0>]}, {memory,2884}, {message_queue_len,0}, {reductions,128}, {trap_exit,false}]}, {<6517.345.0>, [{registered_name,ns_bad_bucket_sup}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb756951c (gen_server:loop/6 + 144)", "CP: 0x00000000 (invalid)","arity = 0", "0xb69efd28 Return addr 0xb753b298 (proc_lib:init_p_do_apply/3 + 28)", "y(0) []","y(1) infinity", "y(2) supervisor", "y(3) {state,{local,ns_bad_bucket_sup},one_for_one,[{child,<0.416.0>,{stats_reader,\"def", "y(4) ns_bad_bucket_sup","y(5) <0.255.0>", "0xb69efd44 Return addr 0x08229d74 ()", "y(0) Catch 0xb753b2a8 (proc_lib:init_p_do_apply/3 + 44)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,0}]}, {heap_size,17711}, {total_heap_size,17711}, {links, [<6517.255.0>,<6517.379.0>,<6517.416.0>,<6517.368.0>, <6517.64.0>]}, {memory,71368}, {message_queue_len,0}, {reductions,301}, {trap_exit,true}]}, {<6517.346.0>, [{registered_name,ns_moxi_sup}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb756951c (gen_server:loop/6 + 144)", "CP: 0x00000000 (invalid)","arity = 0", "0xb6a011e8 Return addr 0xb753b298 (proc_lib:init_p_do_apply/3 + 28)", "y(0) []","y(1) infinity", "y(2) supervisor", "y(3) {state,{local,ns_moxi_sup},one_for_one,[],{dict,0,16,16,8,80,48,{[],[],[],[],[],[", "y(4) ns_moxi_sup","y(5) <0.255.0>", "0xb6a01204 Return addr 0x08229d74 ()", "y(0) Catch 0xb753b2a8 (proc_lib:init_p_do_apply/3 + 44)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,0}]}, {heap_size,17711}, {total_heap_size,17711}, {links,[<6517.255.0>,<6517.64.0>]}, {memory,71308}, {message_queue_len,0}, {reductions,649}, {trap_exit,true}]}, {<6517.347.0>, [{registered_name,[]}, {status,waiting}, {initial_call,{erlang,apply,2}}, {backtrace, ["Program counter: 0xb6cc655c (misc:wait_for_process/2 + 104)", "CP: 0x00000000 (invalid)","arity = 0", "0xb6d235f0 Return addr 0xb6cc92e0 (misc:'-start_singleton/4-fun-0-'/2 + 56)", "y(0) []","y(1) #Ref<0.0.0.2885>", "y(2) infinity", "0xb6d23600 Return addr 0x08229d74 ()", "y(0) <10111.177.0>","y(1) ns_tick"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,0}]}, {heap_size,233}, {total_heap_size,233}, {links,[<6517.255.0>]}, {memory,1356}, {message_queue_len,0}, {reductions,6}, {trap_exit,false}]}, {<6517.348.0>, [{registered_name,[]}, {status,waiting}, {initial_call,{erlang,apply,2}}, {backtrace, ["Program counter: 0xb6cc9728 (misc:'-wait_for_process/2-fun-0-'/3 + 112)", "CP: 0x00000000 (invalid)","arity = 0", "0xb73fe9b4 Return addr 0x08229d74 ()", "y(0) []","y(1) <0.347.0>", "y(2) #Ref<0.0.0.2885>"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,0}]}, {heap_size,233}, {total_heap_size,233}, {links,[<0.177.0>]}, {memory,1372}, {message_queue_len,0}, {reductions,13}, {trap_exit,true}]}, {<6517.351.0>, [{registered_name,[]}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb756951c (gen_server:loop/6 + 144)", "CP: 0x00000000 (invalid)","arity = 0", "0xb6d222d0 Return addr 0xb753b298 (proc_lib:init_p_do_apply/3 + 28)", "y(0) []","y(1) infinity", "y(2) supervisor_cushion", "y(3) {state,memcached,5000,{1304,814964,77393},<0.352.0>}", "y(4) <0.351.0>","y(5) <0.296.0>", "0xb6d222ec Return addr 0x08229d74 ()", "y(0) Catch 0xb753b2a8 (proc_lib:init_p_do_apply/3 + 44)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,0}]}, {heap_size,987}, {total_heap_size,987}, {links,[<6517.296.0>,<6517.352.0>]}, {memory,4412}, {message_queue_len,0}, {reductions,55}, {trap_exit,true}]}, {<6517.352.0>, [{registered_name,[]}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb756951c (gen_server:loop/6 + 144)", "CP: 0x00000000 (invalid)","arity = 0", "0xb6d2479c Return addr 0xb753b298 (proc_lib:init_p_do_apply/3 + 28)", "y(0) []","y(1) infinity", "y(2) ns_port_server", "y(3) {state,#Port<0.2637>,memcached,{[\"WARNING: Found duplicate entry for \\\"tap_keepal", "y(4) <0.352.0>","y(5) <0.351.0>", "0xb6d247b8 Return addr 0x08229d74 ()", "y(0) Catch 0xb753b2a8 (proc_lib:init_p_do_apply/3 + 44)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,0}]}, {heap_size,610}, {total_heap_size,610}, {links,[<6517.351.0>,#Port<6517.2637>]}, {memory,2904}, {message_queue_len,0}, {reductions,202}, {trap_exit,true}]}, {<6517.368.0>, [{registered_name,[]}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb756951c (gen_server:loop/6 + 144)", "CP: 0x00000000 (invalid)","arity = 0", "0xb6a0e930 Return addr 0xb753b298 (proc_lib:init_p_do_apply/3 + 28)", "y(0) []","y(1) infinity", "y(2) stats_collector", "y(3) {state,\"default\",[76118,2265909,0,0,0,0,0,0,0,0,0,0,0,0,0,0,200,0,0,0,0,0,0,0,0],", "y(4) <0.368.0>","y(5) <0.345.0>", "0xb6a0e94c Return addr 0x08229d74 ()", "y(0) Catch 0xb753b2a8 (proc_lib:init_p_do_apply/3 + 44)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,7}]}, {heap_size,6765}, {total_heap_size,17711}, {links,[<6517.345.0>,<6517.302.0>]}, {memory,71308}, {message_queue_len,0}, {reductions,2104422}, {trap_exit,false}]}, {<6517.373.0>, [{registered_name,'ns_vbm_sup-default'}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb756951c (gen_server:loop/6 + 144)", "CP: 0x00000000 (invalid)","arity = 0", "0xb693a8f0 Return addr 0xb753b298 (proc_lib:init_p_do_apply/3 + 28)", "y(0) []","y(1) infinity", "y(2) supervisor", "y(3) {state,{local,'ns_vbm_sup-default'},one_for_one,[],{dict,0,16,16,8,80,48,{[],[],[", "y(4) 'ns_vbm_sup-default'","y(5) <0.305.0>", "0xb693a90c Return addr 0x08229d74 ()", "y(0) Catch 0xb753b2a8 (proc_lib:init_p_do_apply/3 + 44)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,0}]}, {heap_size,233}, {total_heap_size,233}, {links,[<6517.305.0>]}, {memory,1376}, {message_queue_len,0}, {reductions,583}, {trap_exit,true}]}, {<6517.374.0>, [{registered_name,'ns_memcached-default'}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb756951c (gen_server:loop/6 + 144)", "CP: 0x00000000 (invalid)","arity = 0", "0xb73f6a70 Return addr 0xb753b298 (proc_lib:init_p_do_apply/3 + 28)", "y(0) []","y(1) infinity", "y(2) ns_memcached", "y(3) {state,\"default\",#Port<0.2655>}", "y(4) <0.374.0>","y(5) <0.305.0>", "0xb73f6a8c Return addr 0x08229d74 ()", "y(0) Catch 0xb753b2a8 (proc_lib:init_p_do_apply/3 + 44)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,16}]}, {heap_size,6765}, {total_heap_size,35422}, {links,[<6517.50.0>,<6517.305.0>,#Port<6517.2655>]}, {memory,142172}, {message_queue_len,0}, {reductions,1726668}, {trap_exit,true}]}, {<6517.379.0>, [{registered_name,'stats_archiver-default'}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb756951c (gen_server:loop/6 + 144)", "CP: 0x00000000 (invalid)","arity = 0", "0xb69dcf8c Return addr 0xb753b298 (proc_lib:init_p_do_apply/3 + 28)", "y(0) []","y(1) infinity", "y(2) stats_archiver", "y(3) {state,\"default\"}", "y(4) 'stats_archiver-default'", "y(5) <0.345.0>", "0xb69dcfa8 Return addr 0x08229d74 ()", "y(0) Catch 0xb753b2a8 (proc_lib:init_p_do_apply/3 + 44)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,55}]}, {heap_size,2584}, {total_heap_size,20295}, {links,[<6517.303.0>,<6517.345.0>,<6517.50.0>]}, {memory,81664}, {message_queue_len,0}, {reductions,861368}, {trap_exit,false}]}, {<6517.416.0>, [{registered_name,'stats_reader-default'}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb756951c (gen_server:loop/6 + 144)", "CP: 0x00000000 (invalid)","arity = 0", "0xb691eb50 Return addr 0xb753b298 (proc_lib:init_p_do_apply/3 + 28)", "y(0) []","y(1) infinity", "y(2) stats_reader", "y(3) {state,\"default\"}", "y(4) 'stats_reader-default'", "y(5) <0.345.0>", "0xb691eb6c Return addr 0x08229d74 ()", "y(0) Catch 0xb753b2a8 (proc_lib:init_p_do_apply/3 + 44)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,1}]}, {heap_size,28657}, {total_heap_size,103682}, {links,[<6517.345.0>]}, {memory,415172}, {message_queue_len,0}, {reductions,218495}, {trap_exit,false}]}, {<6517.418.0>, [{registered_name,hot_keys_keeper}, {status,waiting}, {initial_call,{proc_lib,init_p,5}}, {backtrace, ["Program counter: 0xb756951c (gen_server:loop/6 + 144)", "CP: 0x00000000 (invalid)","arity = 0", "0xb73a24f8 Return addr 0xb753b298 (proc_lib:init_p_do_apply/3 + 28)", "y(0) []","y(1) infinity", "y(2) hot_keys_keeper", "y(3) {state,[{\"default\",[]}],[{\"default\",[]}],<0.795.0>}", "y(4) hot_keys_keeper","y(5) <0.285.0>", "0xb73a2514 Return addr 0x08229d74 ()", "y(0) Catch 0xb753b2a8 (proc_lib:init_p_do_apply/3 + 44)"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,20}]}, {heap_size,377}, {total_heap_size,754}, {links,[<6517.285.0>,<6517.50.0>]}, {memory,3480}, {message_queue_len,0}, {reductions,1867}, {trap_exit,false}]}, {<6517.801.0>, [{registered_name,[]}, {status,running}, {initial_call,{erlang,apply,2}}, {backtrace, ["Program counter: 0xb749ada4 (unknown function)", "CP: 0xb6d0d6dc (diag_handler:grab_process_info/1 + 32)", "0xb62d4a14 Return addr 0xb6d0e98c (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 44)", "y(0) []", "0xb62d4a1c Return addr 0xb6d0e9ac (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) []","y(1) <0.801.0>", "0xb62d4a28 Return addr 0xb6d0e9ac (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,hot_keys_keeper},{status,waiting},{initial_call,{proc_lib,init_", "y(1) <0.418.0>", "0xb62d4a34 Return addr 0xb6d0e9ac (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,'stats_reader-default'},{status,waiting},{initial_call,{proc_li", "y(1) <0.416.0>", "0xb62d4a40 Return addr 0xb6d0e9ac (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,'stats_archiver-default'},{status,waiting},{initial_call,{proc_", "y(1) <0.379.0>", "0xb62d4a4c Return addr 0xb6d0e9ac (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,'ns_memcached-default'},{status,waiting},{initial_call,{proc_li", "y(1) <0.374.0>", "0xb62d4a58 Return addr 0xb6d0e9ac (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,'ns_vbm_sup-default'},{status,waiting},{initial_call,{proc_lib,", "y(1) <0.373.0>", "0xb62d4a64 Return addr 0xb6d0e9ac (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,[]},{status,waiting},{initial_call,{proc_lib,init_p,5}},{backtr", "y(1) <0.368.0>", "0xb62d4a70 Return addr 0xb6d0e9ac (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,[]},{status,waiting},{initial_call,{proc_lib,init_p,5}},{backtr", "y(1) <0.352.0>", "0xb62d4a7c Return addr 0xb6d0e9ac (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,[]},{status,waiting},{initial_call,{proc_lib,init_p,5}},{backtr", "y(1) <0.351.0>", "0xb62d4a88 Return addr 0xb6d0e9ac (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,[]},{status,waiting},{initial_call,{erlang,apply,2}},{backtrace", "y(1) <0.348.0>", "0xb62d4a94 Return addr 0xb6d0e9ac (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,[]},{status,waiting},{initial_call,{erlang,apply,2}},{backtrace", "y(1) <0.347.0>", "0xb62d4aa0 Return addr 0xb6d0e9ac (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,ns_moxi_sup},{status,waiting},{initial_call,{proc_lib,init_p,5}", "y(1) <0.346.0>", "0xb62d4aac Return addr 0xb6d0e9ac (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,ns_bad_bucket_sup},{status,waiting},{initial_call,{proc_lib,ini", "y(1) <0.345.0>", "0xb62d4ab8 Return addr 0xb6d0e9ac (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,ns_bad_bucket_worker},{status,waiting},{initial_call,{proc_lib,", "y(1) <0.344.0>", "0xb62d4ac4 Return addr 0xb6d0e9ac (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,[]},{status,waiting},{initial_call,{proc_lib,init_p,5}},{backtr", "y(1) <0.340.0>", "0xb62d4ad0 Return addr 0xb6d0e9ac (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,[]},{status,waiting},{initial_call,{proc_lib,init_p,5}},{backtr", "y(1) <0.330.0>", "0xb62d4adc Return addr 0xb6d0e9ac (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,mnesia_late_loader},{status,waiting},{initial_call,{proc_lib,in", "y(1) <0.327.0>", "0xb62d4ae8 Return addr 0xb6d0e9ac (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,mnesia_controller},{status,waiting},{initial_call,{proc_lib,ini", "y(1) <0.326.0>", "0xb62d4af4 Return addr 0xb6d0e9ac (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,mnesia_snmp_sup},{status,waiting},{initial_call,{proc_lib,init_", "y(1) <0.325.0>", "0xb62d4b00 Return addr 0xb6d0e9ac (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,mnesia_checkpoint_sup},{status,waiting},{initial_call,{proc_lib", "y(1) <0.324.0>", "0xb62d4b0c Return addr 0xb6d0e9ac (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,mnesia_tm},{status,waiting},{initial_call,{proc_lib,init_p,5}},", "y(1) <0.323.0>", "0xb62d4b18 Return addr 0xb6d0e9ac (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,mnesia_recover},{status,waiting},{initial_call,{proc_lib,init_p", "y(1) <0.322.0>", "0xb62d4b24 Return addr 0xb6d0e9ac (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,mnesia_locker},{status,waiting},{initial_call,{proc_lib,init_p,", "y(1) <0.321.0>", "0xb62d4b30 Return addr 0xb6d0e9ac (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,mnesia_subscr},{status,waiting},{initial_call,{proc_lib,init_p,", "y(1) <0.320.0>", "0xb62d4b3c Return addr 0xb6d0e9ac (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,mnesia_monitor},{status,waiting},{initial_call,{proc_lib,init_p", "y(1) <0.319.0>", "0xb62d4b48 Return addr 0xb6d0e9ac (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,mnesia_kernel_sup},{status,waiting},{initial_call,{proc_lib,ini", "y(1) <0.318.0>", "0xb62d4b54 Return addr 0xb6d0e9ac (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,mnesia_event},{status,waiting},{initial_call,{proc_lib,init_p,5", "y(1) <0.317.0>", "0xb62d4b60 Return addr 0xb6d0e9ac (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,mnesia_sup},{status,waiting},{initial_call,{proc_lib,init_p,5}}", "y(1) <0.316.0>", "0xb62d4b6c Return addr 0xb6d0e9ac (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,[]},{status,waiting},{initial_call,{application_master,start_it", "y(1) <0.315.0>", "0xb62d4b78 Return addr 0xb6d0e9ac (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,[]},{status,waiting},{initial_call,{proc_lib,init_p,5}},{backtr", "y(1) <0.314.0>", "0xb62d4b84 Return addr 0xb6d0e9ac (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,[]},{status,waiting},{initial_call,{erlang,apply,2}},{backtrace", "y(1) <0.312.0>", "0xb62d4b90 Return addr 0xb6d0e9ac (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,ns_mnesia},{status,waiting},{initial_call,{proc_lib,init_p,5}},", "y(1) <0.311.0>", "0xb62d4b9c Return addr 0xb6d0e9ac (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,[]},{status,waiting},{initial_call,{erlang,apply,2}},{backtrace", "y(1) <0.310.0>", "0xb62d4ba8 Return addr 0xb6d0e9ac (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,ns_good_bucket_sup},{status,waiting},{initial_call,{proc_lib,in", "y(1) <0.305.0>", "0xb62d4bb4 Return addr 0xb6d0e9ac (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,ns_good_bucket_worker},{status,waiting},{initial_call,{proc_lib", "y(1) <0.304.0>", "0xb62d4bc0 Return addr 0xb6d0e9ac (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,ns_stats_event},{status,waiting},{initial_call,{proc_lib,init_p", "y(1) <0.303.0>", "0xb62d4bcc Return addr 0xb6d0e9ac (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,ns_tick_event},{status,waiting},{initial_call,{proc_lib,init_p,", "y(1) <0.302.0>", "0xb62d4bd8 Return addr 0xb6d0e9ac (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,[]},{status,waiting},{initial_call,{proc_lib,init_p,5}},{backtr", "y(1) <0.299.0>", "0xb62d4be4 Return addr 0xb6d0e9ac (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,[]},{status,waiting},{initial_call,{proc_lib,init_p,5}},{backtr", "y(1) <0.298.0>", "0xb62d4bf0 Return addr 0xb6d0e9ac (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,ns_port_sup},{status,waiting},{initial_call,{proc_lib,init_p,5}", "y(1) <0.296.0>", "0xb62d4bfc Return addr 0xb6d0e9ac (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,[]},{status,waiting},{initial_call,{proc_lib,init_p,5}},{backtr", "y(1) <0.289.0>", "0xb62d4c08 Return addr 0xb6d0e9ac (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,menelaus_web},{status,waiting},{initial_call,{proc_lib,init_p,5", "y(1) <0.288.0>", "0xb62d4c14 Return addr 0xb6d0e9ac (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,menelaus_sup},{status,waiting},{initial_call,{proc_lib,init_p,5", "y(1) <0.285.0>", "0xb62d4c20 Return addr 0xb6d0e9ac (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,ns_doctor},{status,waiting},{initial_call,{proc_lib,init_p,5}},", "y(1) <0.278.0>", "0xb62d4c2c Return addr 0xb6d0e9ac (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,ns_heart},{status,waiting},{initial_call,{proc_lib,init_p,5}},{", "y(1) <0.277.0>", "0xb62d4c38 Return addr 0xb6d0e9ac (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,ns_config_rep},{status,waiting},{initial_call,{proc_lib,init_p,", "y(1) <0.270.0>", "0xb62d4c44 Return addr 0xb6d0e9ac (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,ns_node_disco},{status,waiting},{initial_call,{proc_lib,init_p,", "y(1) <0.264.0>", "0xb62d4c50 Return addr 0xb6d0e9ac (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,ns_node_disco_events},{status,waiting},{initial_call,{proc_lib,", "y(1) <0.263.0>", "0xb62d4c5c Return addr 0xb6d0e9ac (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,ns_node_disco_sup},{status,waiting},{initial_call,{proc_lib,ini", "y(1) <0.261.0>", "0xb62d4c68 Return addr 0xb6d0e9ac (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,ns_mail},{status,waiting},{initial_call,{proc_lib,init_p,5}},{b", "y(1) <0.259.0>", "0xb62d4c74 Return addr 0xb6d0e9ac (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,ns_mail_sup},{status,waiting},{initial_call,{proc_lib,init_p,5}", "y(1) <0.258.0>", "0xb62d4c80 Return addr 0xb6d0e9ac (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,ns_log_events},{status,waiting},{initial_call,{proc_lib,init_p,", "y(1) <0.257.0>", "0xb62d4c8c Return addr 0xb6d0e9ac (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,ns_log},{status,waiting},{initial_call,{proc_lib,init_p,5}},{ba", "y(1) <0.256.0>", "0xb62d4c98 Return addr 0xb6d0e9ac (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,ns_server_sup},{status,waiting},{initial_call,{proc_lib,init_p,", "y(1) <0.255.0>", "0xb62d4ca4 Return addr 0xb6d0e9ac (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,[]},{status,waiting},{initial_call,{inet_tcp_dist,do_setup,6}},", "y(1) <0.252.0>", "0xb62d4cb0 Return addr 0xb6d0e9ac (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,[]},{status,waiting},{initial_call,{net_kernel,ticker,2}},{back", "y(1) <0.194.0>", "0xb62d4cbc Return addr 0xb6d0e9ac (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,[]},{status,waiting},{initial_call,{inet_tcp_dist,accept_loop,2", "y(1) <0.193.0>", "0xb62d4cc8 Return addr 0xb6d0e9ac (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,net_kernel},{status,waiting},{initial_call,{proc_lib,init_p,5}}", "y(1) <0.192.0>", "0xb62d4cd4 Return addr 0xb6d0e9ac (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,auth},{status,waiting},{initial_call,{proc_lib,init_p,5}},{back", "y(1) <0.191.0>", "0xb62d4ce0 Return addr 0xb6d0e9ac (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,erl_epmd},{status,waiting},{initial_call,{proc_lib,init_p,5}},{", "y(1) <0.190.0>", "0xb62d4cec Return addr 0xb6d0e9ac (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,net_sup},{status,waiting},{initial_call,{proc_lib,init_p,5}},{b", "y(1) <0.189.0>", "0xb62d4cf8 Return addr 0xb6d0e9ac (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,inet_gethost_native},{status,waiting},{initial_call,{inet_getho", "y(1) <0.181.0>", "0xb62d4d04 Return addr 0xb6d0e9ac (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,inet_gethost_native_sup},{status,waiting},{initial_call,{proc_l", "y(1) <0.180.0>", "0xb62d4d10 Return addr 0xb6d0e9ac (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,disk_log_server},{status,waiting},{initial_call,{proc_lib,init_", "y(1) <0.128.0>", "0xb62d4d1c Return addr 0xb6d0e9ac (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,disk_log_sup},{status,waiting},{initial_call,{proc_lib,init_p,5", "y(1) <0.127.0>", "0xb62d4d28 Return addr 0xb6d0e9ac (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,dets},{status,waiting},{initial_call,{proc_lib,init_p,5}},{back", "y(1) <0.123.0>", "0xb62d4d34 Return addr 0xb6d0e9ac (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,dets_sup},{status,waiting},{initial_call,{proc_lib,init_p,5}},{", "y(1) <0.122.0>", "0xb62d4d40 Return addr 0xb6d0e9ac (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,ns_config},{status,waiting},{initial_call,{proc_lib,init_p,5}},", "y(1) <0.65.0>", "0xb62d4d4c Return addr 0xb6d0e9ac (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,ns_config_events},{status,waiting},{initial_call,{proc_lib,init", "y(1) <0.64.0>", "0xb62d4d58 Return addr 0xb6d0e9ac (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,ns_config_sup},{status,waiting},{initial_call,{proc_lib,init_p,", "y(1) <0.63.0>", "0xb62d4d64 Return addr 0xb6d0e9ac (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,ns_cluster},{status,waiting},{initial_call,{proc_lib,init_p,5}}", "y(1) <0.62.0>", "0xb62d4d70 Return addr 0xb6d0e9ac (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,dist_manager},{status,waiting},{initial_call,{proc_lib,init_p,5", "y(1) <0.55.0>", "0xb62d4d7c Return addr 0xb6d0e9ac (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,ns_server_cluster_sup},{status,waiting},{initial_call,{proc_lib", "y(1) <0.54.0>", "0xb62d4d88 Return addr 0xb6d0e9ac (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,[]},{status,waiting},{initial_call,{application_master,start_it", "y(1) <0.53.0>", "0xb62d4d94 Return addr 0xb6d0e9ac (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,[]},{status,waiting},{initial_call,{proc_lib,init_p,5}},{backtr", "y(1) <0.52.0>", "0xb62d4da0 Return addr 0xb6d0e9ac (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,timer_server},{status,waiting},{initial_call,{proc_lib,init_p,5", "y(1) <0.50.0>", "0xb62d4dac Return addr 0xb6d0e9ac (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,[]},{status,waiting},{initial_call,{erlang,apply,2}},{backtrace", "y(1) <0.49.0>", "0xb62d4db8 Return addr 0xb6d0e9ac (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,[]},{status,waiting},{initial_call,{erlang,apply,2}},{backtrace", "y(1) <0.48.0>", "0xb62d4dc4 Return addr 0xb6d0e9ac (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,cpu_sup},{status,waiting},{initial_call,{proc_lib,init_p,5}},{b", "y(1) <0.47.0>", "0xb62d4dd0 Return addr 0xb6d0e9ac (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,[]},{status,waiting},{initial_call,{erlang,apply,2}},{backtrace", "y(1) <0.46.0>", "0xb62d4ddc Return addr 0xb6d0e9ac (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,memsup},{status,waiting},{initial_call,{proc_lib,init_p,5}},{ba", "y(1) <0.45.0>", "0xb62d4de8 Return addr 0xb6d0e9ac (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,disksup},{status,waiting},{initial_call,{proc_lib,init_p,5}},{b", "y(1) <0.44.0>", "0xb62d4df4 Return addr 0xb6d0e9ac (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,os_mon_sup},{status,waiting},{initial_call,{proc_lib,init_p,5}}", "y(1) <0.43.0>", "0xb62d4e00 Return addr 0xb6d0e9ac (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,[]},{status,waiting},{initial_call,{application_master,start_it", "y(1) <0.42.0>", "0xb62d4e0c Return addr 0xb6d0e9ac (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,[]},{status,waiting},{initial_call,{proc_lib,init_p,5}},{backtr", "y(1) <0.41.0>", "0xb62d4e18 Return addr 0xb6d0e9ac (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,release_handler},{status,waiting},{initial_call,{proc_lib,init_", "y(1) <0.38.0>", "0xb62d4e24 Return addr 0xb6d0e9ac (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,overload},{status,waiting},{initial_call,{proc_lib,init_p,5}},{", "y(1) <0.37.0>", "0xb62d4e30 Return addr 0xb6d0e9ac (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,alarm_handler},{status,waiting},{initial_call,{proc_lib,init_p,", "y(1) <0.36.0>", "0xb62d4e3c Return addr 0xb6d0e9ac (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,sasl_safe_sup},{status,waiting},{initial_call,{proc_lib,init_p,", "y(1) <0.35.0>", "0xb62d4e48 Return addr 0xb6d0e9ac (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,sasl_sup},{status,waiting},{initial_call,{proc_lib,init_p,5}},{", "y(1) <0.34.0>", "0xb62d4e54 Return addr 0xb6d0e9ac (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,[]},{status,waiting},{initial_call,{application_master,start_it", "y(1) <0.33.0>", "0xb62d4e60 Return addr 0xb6d0e9ac (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,[]},{status,waiting},{initial_call,{proc_lib,init_p,5}},{backtr", "y(1) <0.32.0>", "0xb62d4e6c Return addr 0xb6d0e9ac (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,kernel_safe_sup},{status,waiting},{initial_call,{proc_lib,init_", "y(1) <0.27.0>", "0xb62d4e78 Return addr 0xb6d0e9ac (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,[]},{status,waiting},{initial_call,{proc_lib,init_p,5}},{backtr", "y(1) <0.26.0>", "0xb62d4e84 Return addr 0xb6d0e9ac (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,user},{status,waiting},{initial_call,{user,server,2}},{backtrac", "y(1) <0.25.0>", "0xb62d4e90 Return addr 0xb6d0e9ac (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,[]},{status,waiting},{initial_call,{proc_lib,init_p,5}},{backtr", "y(1) <0.24.0>", "0xb62d4e9c Return addr 0xb6d0e9ac (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,standard_error},{status,waiting},{initial_call,{standard_error,", "y(1) <0.23.0>", "0xb62d4ea8 Return addr 0xb6d0e9ac (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,standard_error_sup},{status,waiting},{initial_call,{proc_lib,in", "y(1) <0.22.0>", "0xb62d4eb4 Return addr 0xb6d0e9ac (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,code_server},{status,waiting},{initial_call,{erlang,apply,2}},{", "y(1) <0.21.0>", "0xb62d4ec0 Return addr 0xb6d0e9ac (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,file_server_2},{status,waiting},{initial_call,{proc_lib,init_p,", "y(1) <0.20.0>", "0xb62d4ecc Return addr 0xb6d0e9ac (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,global_group},{status,waiting},{initial_call,{proc_lib,init_p,5", "y(1) <0.19.0>", "0xb62d4ed8 Return addr 0xb6d0e9ac (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,inet_db},{status,waiting},{initial_call,{proc_lib,init_p,5}},{b", "y(1) <0.17.0>", "0xb62d4ee4 Return addr 0xb6d0e9ac (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,[]},{status,waiting},{initial_call,{erlang,apply,2}},{backtrace", "y(1) <0.16.0>", "0xb62d4ef0 Return addr 0xb6d0e9ac (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,[]},{status,waiting},{initial_call,{erlang,apply,2}},{backtrace", "y(1) <0.15.0>", "0xb62d4efc Return addr 0xb6d0e9ac (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,[]},{status,waiting},{initial_call,{erlang,apply,2}},{backtrace", "y(1) <0.14.0>", "0xb62d4f08 Return addr 0xb6d0e9ac (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,global_name_server},{status,waiting},{initial_call,{proc_lib,in", "y(1) <0.13.0>", "0xb62d4f14 Return addr 0xb6d0e9ac (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,rex},{status,waiting},{initial_call,{proc_lib,init_p,5}},{backt", "y(1) <0.12.0>", "0xb62d4f20 Return addr 0xb6d0e9ac (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,kernel_sup},{status,waiting},{initial_call,{proc_lib,init_p,5}}", "y(1) <0.11.0>", "0xb62d4f2c Return addr 0xb6d0e9ac (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,[]},{status,waiting},{initial_call,{application_master,start_it", "y(1) <0.10.0>", "0xb62d4f38 Return addr 0xb6d0e9ac (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,[]},{status,waiting},{initial_call,{proc_lib,init_p,5}},{backtr", "y(1) <0.9.0>", "0xb62d4f44 Return addr 0xb6d0e9ac (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,application_controller},{status,waiting},{initial_call,{erlang,", "y(1) <0.7.0>", "0xb62d4f50 Return addr 0xb6d0e9ac (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,error_logger},{status,waiting},{initial_call,{proc_lib,init_p,5", "y(1) <0.6.0>", "0xb62d4f5c Return addr 0xb6d0e9ac (diag_handler:'-do_diag_per_node/0-lc$^0/1-0-'/1 + 76)", "y(0) [{registered_name,erl_prim_loader},{status,waiting},{initial_call,{erlang,apply,2", "y(1) <0.3.0>", "0xb62d4f68 Return addr 0xb6d0d7e4 (diag_handler:do_diag_per_node/0 + 112)", "y(0) [{registered_name,init},{status,waiting},{initial_call,{otp_ring0,start,2}},{back", "y(1) <0.0.0>", "0xb62d4f74 Return addr 0xb75b24f4 (rpc:'-handle_call_call/6-fun-0-'/5 + 104)", "y(0) []","y(1) []", "y(2) [{version,[{kernel,\"2.13.4\"},{mnesia,\"4.4.12\"},{stdlib,\"1.16.4\"},{os_mon,\"2.2.4\"}", "y(3) [{{node,'ns_1@127.0.0.1',ns_log},[{filename,\"/var/opt/membase/1.6.5.4r/data/ns_1/", "y(4) [\"bucket_engine_1.6.5.3_1_g5ba5371-Linux.i686.tar.gz\",\"curl-7.21.1-w64_patched.ta", "y(5) [{kernel,\"2.13.4\"},{mnesia,\"4.4.12\"},{stdlib,\"1.16.4\"},{os_mon,\"2.2.4\"},{ns_serve", "0xb62d4f90 Return addr 0x08229d74 ()", "y(0) Catch 0xb75b24f4 (rpc:'-handle_call_call/6-fun-0-'/5 + 104)", "y(1) []","y(2) []","y(3) []", "y(4) <0.12.0>"]}, {error_handler,error_handler}, {garbage_collection, [{fullsweep_after,65535},{minor_gcs,27}]}, {heap_size,46368}, {total_heap_size,364179}, {links,[]}, {memory,1457120}, {message_queue_len,0}, {reductions,593381}, {trap_exit,false}]}]}, {memory,{1050726400,412069888,{<6517.255.0>,901104}}}, {disk, [{"/",7583436,16}, {"/dev",508460,1}, {"/dev/shm",513048,0}, {"/var/run",513048,1}, {"/var/lock",513048,0}, {"/lib/init/rw",513048,0}, {"/boot",233191,8}]}]}] nodes_info = [{struct,[{uptime,<<"321">>}, {memoryTotal,1050726400}, {memoryFree,261578752}, {mcdMemoryReserved,801}, {mcdMemoryAllocated,801}, {otpNode,<<"ns_1@10.1.4.243">>}, {otpCookie,<<"elyrdmyctqykhepq">>}, {clusterMembership,<<"active">>}, {status,<<"healthy">>}, {hostname,<<"10.1.4.243:8091">>}, {clusterCompatibility,1}, {version,<<"1.6.5.4r">>}, {os,<<"i686-pc-linux-gnu">>}, {ports,{struct,[{proxy,11211},{direct,11210}]}}]}, {struct,[{uptime,<<"373">>}, {memoryTotal,1050726400}, {memoryFree,638656512}, {mcdMemoryReserved,801}, {mcdMemoryAllocated,801}, {otpNode,<<"ns_1@10.1.4.244">>}, {otpCookie,<<"elyrdmyctqykhepq">>}, {clusterMembership,<<"active">>}, {status,<<"healthy">>}, {hostname,<<"10.1.4.244:8091">>}, {clusterCompatibility,1}, {version,<<"1.6.5.4r">>}, {os,<<"i686-pc-linux-gnu">>}, {ports,{struct,[{proxy,11211},{direct,11210}]}}]}] buckets = [{"default", [{num_replicas,1}, {ram_quota,839909376}, {auth_type,sasl}, {sasl_password,[]}, {type,membase}, {num_vbuckets,1024}, {db_shards,4}, {ht_size,3079}, {tap_keepalive,0}, {tap_noop_interval,20}, {max_txn_size,1000}, {ht_locks,5}, {servers,['ns_1@10.1.4.243','ns_1@10.1.4.244']}, {map,[['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined]]}]}] logs: ------------------------------- 2011-05-07 17:05:02.949 ns_node_disco:3:info:cookie update - Initial otp cookie generated: elyrdmyctqykhepq 2011-05-07 17:05:03.162 menelaus_app:1:info:web start ok - Membase Server has started on web port 8091 on node 'ns_1@127.0.0.1'. 2011-05-07 17:06:57.368 menelaus_web:12:info:message - Created bucket "default" of type: membase 2011-05-07 17:06:59.419 ns_memcached:1:info:message - Bucket "default" loaded on node 'ns_1@127.0.0.1' in 1 seconds. 2011-05-07 17:27:43.247 ns_node_disco:5:warning:node down - Node nonode@nohost saw that node 'ns_1@127.0.0.1' went down. 2011-05-07 17:27:43.252 ns_node_disco:4:info:node up - Node 'ns_1@10.1.4.243' saw that node 'ns_1@10.1.4.243' came up. 2011-05-07 17:27:43.955 ns_node_disco:4:info:node up - Node 'ns_1@10.1.4.243' saw that node 'ns_1@10.1.4.244' came up. 2011-05-07 17:27:44.590 menelaus_app:1:info:web start ok - Membase Server has started on web port 8091 on node 'ns_1@10.1.4.244'. 2011-05-07 17:27:45.321 ns_cluster:3:info:message - Node ns_1@10.1.4.244 joined cluster 2011-05-07 17:27:59.528 ns_orchestrator:4:info:message - Starting rebalance, KeepNodes = ['ns_1@10.1.4.243','ns_1@10.1.4.244'], EjectNodes = [] 2011-05-07 17:28:11.010 ns_memcached:1:info:message - Bucket "default" loaded on node 'ns_1@10.1.4.244' in 0 seconds. 2011-05-07 17:28:11.873 ns_orchestrator:2:info:message - Rebalance exited with reason wait_for_memcached_failed 2011-05-07 17:29:31.181 ns_orchestrator:4:info:message - Starting rebalance, KeepNodes = ['ns_1@10.1.4.243'], EjectNodes = ['ns_1@10.1.4.244'] 2011-05-07 17:29:31.275 ns_memcached:2:info:message - Shutting down bucket "default" on 'ns_1@10.1.4.244' 2011-05-07 17:29:31.276 ns_orchestrator:1:info:message - Rebalance completed successfully. 2011-05-07 17:29:31.276 ns_cluster:1:info:message - Node 'ns_1@10.1.4.244' is leaving cluster. 2011-05-07 17:29:31.325 ns_node_disco:5:warning:node down - Node 'ns_1@10.1.4.243' saw that node 'ns_1@10.1.4.244' went down. 2011-05-07 17:30:33.161 menelaus_app:1:info:web start ok - Membase Server has started on web port 8091 on node 'ns_1@10.1.4.244'. 2011-05-07 17:30:33.194 ns_port_server:0:info:message - Port server memcached on node 'ns_1@10.1.4.244' exited with status 71. Restarting. Messages: failed to listen on TCP port 11210: Address already in use 2011-05-07 17:30:33.194 supervisor_cushion:1:warning:port exited too soon after restart - Service memcached exited on node 'ns_1@10.1.4.244' in 0.03s 2011-05-07 17:30:33.245 ns_cluster:3:info:message - Node ns_1@10.1.4.244 joined cluster 2011-05-07 17:30:40.831 ns_memcached:1:info:message - Bucket "default" loaded on node 'ns_1@10.1.4.244' in 1 seconds. 2011-05-07 17:31:41.168 ns_orchestrator:2:info:message - Rebalance exited with reason stopped 2011-05-07 17:31:57.858 ns_cluster:1:info:message - Node 'ns_1@10.1.4.244' is leaving cluster. 2011-05-07 17:31:57.874 ns_memcached:2:info:message - Shutting down bucket "default" on 'ns_1@10.1.4.244' 2011-05-07 17:33:02.944 ns_node_disco:4:info:node up - Node 'ns_1@10.1.4.243' saw that node 'ns_1@10.1.4.244' came up. (repeated 1 times) 2011-05-07 17:33:02.944 ns_orchestrator:4:info:message - Starting rebalance, KeepNodes = ['ns_1@10.1.4.243','ns_1@10.1.4.244'], EjectNodes = [] (repeated 1 times) 2011-05-07 17:35:02.943 ns_node_disco:5:warning:node down - Node 'ns_1@10.1.4.243' saw that node 'ns_1@10.1.4.244' went down. (repeated 1 times) 2011-05-07 17:35:02.943 ns_orchestrator:1:info:message - Rebalance completed successfully. (repeated 1 times) 2011-05-07 17:35:02.943 ns_orchestrator:4:info:message - Starting rebalance, KeepNodes = ['ns_1@10.1.4.243'], EjectNodes = ['ns_1@10.1.4.244'] (repeated 2 times) 2011-05-07 17:35:45.855 ns_node_disco:2:info:cookie update - Node 'ns_1@10.1.4.243' synchronized otp cookie elyrdmyctqykhepq from cluster 2011-05-07 17:35:45.951 menelaus_app:1:info:web start ok - Membase Server has started on web port 8091 on node 'ns_1@10.1.4.243'. 2011-05-07 17:35:48.436 ns_memcached:1:info:message - Bucket "default" loaded on node 'ns_1@10.1.4.243' in 1 seconds. 2011-05-07 17:35:58.716 ns_node_disco:4:info:node up - Node 'ns_1@10.1.4.243' saw that node 'ns_1@10.1.4.244' came up. 2011-05-07 17:35:59.011 menelaus_app:1:info:web start ok - Membase Server has started on web port 8091 on node 'ns_1@10.1.4.244'. 2011-05-07 17:35:59.076 ns_port_server:0:info:message - Port server memcached on node 'ns_1@10.1.4.244' exited with status 71. Restarting. Messages: failed to listen on TCP port 11210: Address already in use 2011-05-07 17:35:59.076 supervisor_cushion:1:warning:port exited too soon after restart - Service memcached exited on node 'ns_1@10.1.4.244' in 0.06s 2011-05-07 17:35:59.145 ns_cluster:3:info:message - Node ns_1@10.1.4.244 joined cluster 2011-05-07 17:36:12.620 ns_orchestrator:4:info:message - Starting rebalance, KeepNodes = ['ns_1@10.1.4.243','ns_1@10.1.4.244'], EjectNodes = [] 2011-05-07 17:36:15.575 ns_memcached:1:info:message - Bucket "default" loaded on node 'ns_1@10.1.4.244' in 2 seconds. logs_node: ------------------------------- INFO REPORT <3.54.0> 2011-05-07 17:05:02 =============================================================================== nonode@nohost:log_os_info:25: OS type: {unix,linux} Version: {2,6,32} Runtime info: [{otp_release,"R13B03"}, {erl_version,"5.7.4"}, {erl_version_long,"Erlang R13B03 (erts-5.7.4) [source] [rq:1] [async-threads:16] [hipe] [kernel-poll:false]\n"}, {system_arch_raw,"i686-pc-linux-gnu"}, {system_arch,"i686-pc-linux-gnu"}, {localtime,{{2011,5,7},{17,5,2}}}, {memory,[{total,7669032}, {processes,649420}, {processes_used,642092}, {system,7019612}, {atom,255161}, {atom_used,252399}, {binary,3648}, {code,1775796}, {ets,165068}]}, {loaded,[ns_info,log_os_info,misc,ns_log_mf_h, ns_server_cluster_sup,ns_server,timer,io_lib_fread, cpu_sup,memsup,disksup,os_mon,io_lib_pretty,unicode, io_lib_format,io_lib,io,sasl_report,release_handler, calendar,overload,alarm_handler,sasl_report_tty_h, sasl,ns_bootstrap,file_io_server,orddict,erl_eval, file,c,error_logger_tty_h,queue,kernel_config,user, user_sup,supervisor_bridge,standard_error,gb_sets, hipe_unified_loader,packages,code_server,code, file_server,net_kernel,global_group,erl_distribution, string,ets,filename,inet_gethost_native,inet_parse, inet,inet_udp,os,inet_config,inet_db,global,gb_trees, rpc,dict,supervisor,kernel,application_master,sys, application,gen_server,erl_parse,proplists,erl_scan, lists,application_controller,proc_lib,gen,gen_event, error_logger,heart,error_handler,erlang, erl_prim_loader,prim_zip,zlib,prim_file,prim_inet, init,otp_ring0]}, {applications,[{kernel,"ERTS CXC 138 10","2.13.4"}, {stdlib,"ERTS CXC 138 10","1.16.4"}, {os_mon,"CPO CXC 138 46","2.2.4"}, {ns_server,"Membase server","1.6.5.4r"}, {menelaus,"Membase menelaus","1.6.5.4r"}, {sasl,"SASL CXC 138 11","2.1.8"}]}, {pre_loaded,[erlang,erl_prim_loader,prim_zip,zlib,prim_file, prim_inet,init,otp_ring0]}, {process_count,44}, {node,nonode@nohost}, {nodes,[]}, {registered,[standard_error_sup,rex,user,init,kernel_sup, inet_db,global_name_server,release_handler, code_server,file_server_2,error_logger,overload, ns_server_cluster_sup,application_controller, kernel_safe_sup,standard_error,alarm_handler, global_group,timer_server,sasl_sup,os_mon_sup, cpu_sup,erl_prim_loader,memsup,sasl_safe_sup, disksup]}, {cookie,nocookie}, {wordsize,4}, {wall_clock,0}] INFO REPORT <3.55.0> 2011-05-07 17:05:02 =============================================================================== reading ip config from "/etc/opt/membase/1.6.5.4r/ip" PROGRESS REPORT <3.56.0> 2011-05-07 17:05:02 =============================================================================== supervisor {local,net_sup} started [{pid,<3.57.0>}, {name,erl_epmd}, {mfa,{erl_epmd,start_link,[]}}, {restart_type,permanent}, {shutdown,2000}, {child_type,worker}] PROGRESS REPORT <3.56.0> 2011-05-07 17:05:02 =============================================================================== supervisor {local,net_sup} started [{pid,<3.58.0>}, {name,auth}, {mfa,{auth,start_link,[]}}, {restart_type,permanent}, {shutdown,2000}, {child_type,worker}] PROGRESS REPORT <6516.56.0> 2011-05-07 17:05:02 =============================================================================== supervisor {local,net_sup} started [{pid,<6516.59.0>}, {name,net_kernel}, {mfa,{net_kernel,start_link,[['ns_1@127.0.0.1',longnames]]}}, {restart_type,permanent}, {shutdown,2000}, {child_type,worker}] PROGRESS REPORT <6516.11.0> 2011-05-07 17:05:02 =============================================================================== supervisor {local,kernel_sup} started [{pid,<6516.56.0>}, {name,net_sup_dynamic}, {mfa,{erl_distribution,start_link,[['ns_1@127.0.0.1',longnames]]}}, {restart_type,permanent}, {shutdown,1000}, {child_type,supervisor}] PROGRESS REPORT <6516.54.0> 2011-05-07 17:05:02 =============================================================================== supervisor {local,ns_server_cluster_sup} started [{pid,<6516.55.0>}, {name,dist_manager}, {mfa,{dist_manager,start_link,[]}}, {restart_type,permanent}, {shutdown,10}, {child_type,worker}] PROGRESS REPORT <6516.54.0> 2011-05-07 17:05:02 =============================================================================== supervisor {local,ns_server_cluster_sup} started [{pid,<6516.62.0>}, {name,ns_cluster}, {mfa,{ns_cluster,start_link,[]}}, {restart_type,permanent}, {shutdown,5000}, {child_type,worker}] INFO REPORT <6516.63.0> 2011-05-07 17:05:02 =============================================================================== loading config from "/etc/opt/membase/1.6.5.4r/config" PROGRESS REPORT <6516.63.0> 2011-05-07 17:05:02 =============================================================================== supervisor {local,ns_config_sup} started [{pid,<6516.64.0>}, {name,ns_config_events}, {mfa,{gen_event,start_link,[{local,ns_config_events}]}}, {restart_type,permanent}, {shutdown,10}, {child_type,worker}] PROGRESS REPORT <6516.63.0> 2011-05-07 17:05:02 =============================================================================== supervisor {local,ns_config_sup} started [{pid,<6516.65.0>}, {name,ns_config}, {mfa,{ns_config,start_link, ["/etc/opt/membase/1.6.5.4r/config", ns_config_default]}}, {restart_type,permanent}, {shutdown,10}, {child_type,worker}] PROGRESS REPORT <6516.63.0> 2011-05-07 17:05:02 =============================================================================== supervisor {local,ns_config_sup} started [{pid,<6516.67.0>}, {name,ns_config_isasl_sync}, {mfa,{ns_config_isasl_sync,start_link,[]}}, {restart_type,transient}, {shutdown,10}, {child_type,worker}] INFO REPORT <6516.64.0> 2011-05-07 17:05:02 =============================================================================== isasl_sync init: ["/var/opt/membase/1.6.5.4r/data/ns_1/isasl.pw","_admin", "_admin"] INFO REPORT <6516.64.0> 2011-05-07 17:05:02 =============================================================================== isasl_sync init buckets: [] INFO REPORT <6516.64.0> 2011-05-07 17:05:02 =============================================================================== Writing isasl passwd file: "/var/opt/membase/1.6.5.4r/data/ns_1/isasl.pw" PROGRESS REPORT <6516.63.0> 2011-05-07 17:05:02 =============================================================================== supervisor {local,ns_config_sup} started [{pid,<6516.69.0>}, {name,ns_config_log}, {mfa,{ns_config_log,start_link,[]}}, {restart_type,transient}, {shutdown,10}, {child_type,worker}] PROGRESS REPORT <6516.54.0> 2011-05-07 17:05:02 =============================================================================== supervisor {local,ns_server_cluster_sup} started [{pid,<6516.63.0>}, {name,ns_config_sup}, {mfa,{ns_config_sup,start_link,[]}}, {restart_type,permanent}, {shutdown,infinity}, {child_type,supervisor}] INFO REPORT <6516.71.0> 2011-05-07 17:05:02 =============================================================================== ns_log:init(): Couldn't load logs from "/var/opt/membase/1.6.5.4r/data/ns_1/ns_log": {error, enoent} PROGRESS REPORT <6516.70.0> 2011-05-07 17:05:02 =============================================================================== supervisor {local,ns_server_sup} started [{pid,<6516.71.0>}, {name,ns_log}, {mfa,{ns_log,start_link,[]}}, {restart_type,permanent}, {shutdown,10}, {child_type,worker}] PROGRESS REPORT <6516.70.0> 2011-05-07 17:05:02 =============================================================================== supervisor {local,ns_server_sup} started [{pid,<6516.72.0>}, {name,ns_log_events}, {mfa,{gen_event,start_link,[{local,ns_log_events}]}}, {restart_type,permanent}, {shutdown,10}, {child_type,worker}] PROGRESS REPORT <6516.73.0> 2011-05-07 17:05:02 =============================================================================== supervisor {local,ns_mail_sup} started [{pid,<6516.74.0>}, {name,ns_mail}, {mfa,{ns_mail,start_link,[]}}, {restart_type,permanent}, {shutdown,10}, {child_type,worker}] PROGRESS REPORT <6516.73.0> 2011-05-07 17:05:02 =============================================================================== supervisor {local,ns_mail_sup} started [{pid,<6516.75.0>}, {name,ns_mail_log}, {mfa,{ns_mail_log,start_link,[]}}, {restart_type,transient}, {shutdown,10}, {child_type,worker}] PROGRESS REPORT <6516.70.0> 2011-05-07 17:05:02 =============================================================================== supervisor {local,ns_server_sup} started [{pid,<6516.73.0>}, {name,ns_mail_sup}, {mfa,{ns_mail_sup,start_link,[]}}, {restart_type,permanent}, {shutdown,infinity}, {child_type,supervisor}] PROGRESS REPORT <6516.76.0> 2011-05-07 17:05:02 =============================================================================== supervisor {local,ns_node_disco_sup} started [{pid,<6516.77.0>}, {name,ns_node_disco_events}, {mfa,{gen_event,start_link,[{local,ns_node_disco_events}]}}, {restart_type,permanent}, {shutdown,10}, {child_type,worker}] INFO REPORT <6516.78.0> 2011-05-07 17:05:02 =============================================================================== Initting ns_node_disco with [] INFO REPORT <6516.79.0> 2011-05-07 17:05:02 =============================================================================== ns_node_disco cookie_sync INFO REPORT <6516.79.0> 2011-05-07 17:05:02 =============================================================================== ns_log: logging ns_node_disco:3:Initial otp cookie generated: elyrdmyctqykhepq INFO REPORT <6516.64.0> 2011-05-07 17:05:02 =============================================================================== config change: otp -> [{cookie,elyrdmyctqykhepq}] INFO REPORT <6516.79.0> 2011-05-07 17:05:02 =============================================================================== ns_node_disco: nodes_wanted updated: ['ns_1@127.0.0.1'], with cookie: elyrdmyctqykhepq INFO REPORT <6516.79.0> 2011-05-07 17:05:02 =============================================================================== ns_node_disco: nodes_wanted pong: ['ns_1@127.0.0.1'], with cookie: elyrdmyctqykhepq PROGRESS REPORT <6516.76.0> 2011-05-07 17:05:03 =============================================================================== supervisor {local,ns_node_disco_sup} started [{pid,<6516.78.0>}, {name,ns_node_disco}, {mfa,{ns_node_disco,start_link,[]}}, {restart_type,permanent}, {shutdown,10}, {child_type,worker}] PROGRESS REPORT <6516.76.0> 2011-05-07 17:05:03 =============================================================================== supervisor {local,ns_node_disco_sup} started [{pid,<6516.83.0>}, {name,ns_node_disco_log}, {mfa,{ns_node_disco_log,start_link,[]}}, {restart_type,transient}, {shutdown,10}, {child_type,worker}] PROGRESS REPORT <6516.76.0> 2011-05-07 17:05:03 =============================================================================== supervisor {local,ns_node_disco_sup} started [{pid,<6516.84.0>}, {name,ns_node_disco_conf_events}, {mfa,{ns_node_disco_conf_events,start_link,[]}}, {restart_type,transient}, {shutdown,10}, {child_type,worker}] INFO REPORT <6516.85.0> 2011-05-07 17:05:03 =============================================================================== ns_config_rep init pulling INFO REPORT <6516.85.0> 2011-05-07 17:05:03 =============================================================================== ns_config_rep init pushing INFO REPORT <6516.85.0> 2011-05-07 17:05:03 =============================================================================== ns_config_rep init reannouncing INFO REPORT <6516.64.0> 2011-05-07 17:05:03 =============================================================================== ns_node_disco_conf_events config on otp INFO REPORT <6516.64.0> 2011-05-07 17:05:03 =============================================================================== config change: otp -> [{cookie,elyrdmyctqykhepq}] INFO REPORT <6516.64.0> 2011-05-07 17:05:03 =============================================================================== config change: alerts -> [{email,[]}, {email_alerts,false}, [{user,undefined}, {pass,"********"}, {addr,undefined}, {port,undefined}, {encrypt,false}], {alerts,[server_down,server_unresponsive,server_up,server_joined,server_left, bucket_created,bucket_deleted,bucket_auth_failed]}] INFO REPORT <6516.64.0> 2011-05-07 17:05:03 =============================================================================== config change: buckets -> [{configs,[]}] INFO REPORT <6516.64.0> 2011-05-07 17:05:03 =============================================================================== config change: memory_quota -> 801 INFO REPORT <6516.87.0> 2011-05-07 17:05:03 =============================================================================== ns_node_disco cookie_sync INFO REPORT <6516.64.0> 2011-05-07 17:05:03 =============================================================================== config change: moxi -> [{port,11211},{verbosity,[]}] INFO REPORT <6516.64.0> 2011-05-07 17:05:03 =============================================================================== ns_node_disco_conf_events config on nodes_wanted INFO REPORT <6516.64.0> 2011-05-07 17:05:03 =============================================================================== config change: nodes_wanted -> ['ns_1@127.0.0.1'] INFO REPORT <6516.64.0> 2011-05-07 17:05:03 =============================================================================== config change: port_servers -> [{moxi,"./bin/moxi/moxi", ["-Z", {"port_listen=~B,default_bucket_name=default,downstream_max=1024,downstream_conn_max=4,connect_max_errors=5,connect_retry_interval=30000,connect_timeout=400,auth_timeout=100,cycle=200,downstream_conn_queue_timeout=200,downstream_timeout=5000,wait_queue_timeout=200", [port]}, "-z", {"url=http://127.0.0.1:~B/pools/default/saslBucketsStreaming", [{rest,port}]}, "-p","0","-Y","y","-O","stderr", {"~s",[verbosity]}], [{env,[{"EVENT_NOSELECT","1"}, {"MOXI_SASL_PLAIN_USR",{"~s",[{ns_moxi_sup,rest_user,[]}]}}, {"MOXI_SASL_PLAIN_PWD",{"~s",[{ns_moxi_sup,rest_pass,[]}]}}]}, use_stdio,stderr_to_stdout,stream]}, {memcached,"./bin/memcached/memcached", ["-X","./bin/memcached/stdin_term_handler.so","-p", {"~B",[port]}, "-E","./bin/bucket_engine/bucket_engine.so","-B","binary","-r", "-c","10000","-e", {"admin=~s;default_bucket_name=default;auto_create=false", [admin_user]}, {"~s",[verbosity]}], [{env,[{"EVENT_NOSELECT","1"}, {"MEMCACHED_TOP_KEYS","100"}, {"ISASL_PWFILE",{"~s",[{isasl,path}]}}, {"ISASL_DB_CHECK_TIME","1"}]}, use_stdio,stderr_to_stdout,stream]}] INFO REPORT <6516.88.0> 2011-05-07 17:05:03 =============================================================================== ns_node_disco cookie_sync INFO REPORT <6516.64.0> 2011-05-07 17:05:03 =============================================================================== config change: replication -> [{enabled,true}] INFO REPORT <6516.87.0> 2011-05-07 17:05:03 =============================================================================== ns_node_disco: nodes_wanted updated: ['ns_1@127.0.0.1'], with cookie: elyrdmyctqykhepq INFO REPORT <6516.64.0> 2011-05-07 17:05:03 =============================================================================== config change: rest -> [{port,8091}] INFO REPORT <6516.64.0> 2011-05-07 17:05:03 =============================================================================== config change: rest_creds -> ******** INFO REPORT <6516.64.0> 2011-05-07 17:05:03 =============================================================================== config change: {node,'ns_1@127.0.0.1',isasl} -> [{path,"/var/opt/membase/1.6.5.4r/data/ns_1/isasl.pw"}] INFO REPORT <6516.64.0> 2011-05-07 17:05:03 =============================================================================== config change: {node,'ns_1@127.0.0.1',membership} -> active INFO REPORT <6516.64.0> 2011-05-07 17:05:03 =============================================================================== config change: {node,'ns_1@127.0.0.1',memcached} -> [{port,11210}, {dbdir,"/var/opt/membase/1.6.5.4r/data/ns_1"}, {admin_user,"_admin"}, {admin_pass,"_admin"}, {bucket_engine,"./bin/bucket_engine/bucket_engine.so"}, {engines,[{membase,[{engine,"bin/ep_engine/ep.so"}, {initfile,"/etc/opt/membase/1.6.5.4r/init.sql"}]}, {memcached,[{engine,"bin/memcached/default_engine.so"}]}]}, {verbosity,[]}] INFO REPORT <6516.87.0> 2011-05-07 17:05:03 =============================================================================== ns_node_disco: nodes_wanted pong: ['ns_1@127.0.0.1'], with cookie: elyrdmyctqykhepq INFO REPORT <6516.88.0> 2011-05-07 17:05:03 =============================================================================== ns_node_disco: nodes_wanted updated: ['ns_1@127.0.0.1'], with cookie: elyrdmyctqykhepq INFO REPORT <6516.64.0> 2011-05-07 17:05:03 =============================================================================== config change: {node,'ns_1@127.0.0.1',ns_log} -> [{filename,"/var/opt/membase/1.6.5.4r/data/ns_1/ns_log"}] INFO REPORT <6516.64.0> 2011-05-07 17:05:03 =============================================================================== ns_node_disco_conf_events config all INFO REPORT <6516.88.0> 2011-05-07 17:05:03 =============================================================================== ns_node_disco: nodes_wanted pong: ['ns_1@127.0.0.1'], with cookie: elyrdmyctqykhepq INFO REPORT <6516.85.0> 2011-05-07 17:05:03 =============================================================================== Pushing config PROGRESS REPORT <6516.76.0> 2011-05-07 17:05:03 =============================================================================== supervisor {local,ns_node_disco_sup} started [{pid,<6516.85.0>}, {name,ns_config_rep}, {mfa,{ns_config_rep,start_link,[]}}, {restart_type,permanent}, {shutdown,10}, {child_type,worker}] PROGRESS REPORT <6516.70.0> 2011-05-07 17:05:03 =============================================================================== supervisor {local,ns_server_sup} started [{pid,<6516.76.0>}, {name,ns_node_disco_sup}, {mfa,{ns_node_disco_sup,start_link,[]}}, {restart_type,permanent}, {shutdown,infinity}, {child_type,supervisor}] INFO REPORT <6516.85.0> 2011-05-07 17:05:03 =============================================================================== Pushing config done PROGRESS REPORT <6516.70.0> 2011-05-07 17:05:03 =============================================================================== supervisor {local,ns_server_sup} started [{pid,<6516.90.0>}, {name,ns_heart}, {mfa,{ns_heart,start_link,[]}}, {restart_type,permanent}, {shutdown,10}, {child_type,worker}] PROGRESS REPORT <6516.70.0> 2011-05-07 17:05:03 =============================================================================== supervisor {local,ns_server_sup} started [{pid,<6516.92.0>}, {name,ns_doctor}, {mfa,{ns_doctor,start_link,[]}}, {restart_type,permanent}, {shutdown,10}, {child_type,worker}] INFO REPORT <6516.92.0> 2011-05-07 17:05:03 =============================================================================== ns_1@127.0.0.1:ns_doctor:78: Got initial status [{'ns_1@127.0.0.1', [{last_heard, {1304,813103,73861}}, {active_buckets,[]}, {memory, [{total,8609216}, {processes,1104760}, {processes_used,1098760}, {system,7504456}, {atom,312841}, {atom_used,291414}, {binary,35328}, {code,2132418}, {ets,190532}]}, {cluster_compatibility_version, 1}, {version, [{kernel,"2.13.4"}, {stdlib,"1.16.4"}, {os_mon,"2.2.4"}, {ns_server,"1.6.5.4r"}, {menelaus,"1.6.5.4r"}, {sasl,"2.1.8"}]}, {system_arch, "i686-pc-linux-gnu"}, {wall_clock,0}, {memory_data, {1050726400,642637824, {<6516.7.0>,142252}}}, {disk_data, [{"/",7583436,17}, {"/dev",508460,1}, {"/dev/shm",513048,0}, {"/var/run",513048,1}, {"/var/lock",513048,0}, {"/lib/init/rw",513048,0}, {"/boot",233191,8}]}, {meminfo, <<"MemTotal: 1026100 kB\nMemFree: 397648 kB\nBuffers: 152648 kB\nCached: 413804 kB\nSwapCached: 0 kB\nActive: 333000 kB\nInactive: 259168 kB\nActive(anon): 21580 kB\nInactive(anon): 4412 kB\nActive(file): 311420 kB\nInactive(file): 254756 kB\nUnevictable: 0 kB\nMlocked: 0 kB\nHighTotal: 137160 kB\nHighFree: 252 kB\nLowTotal: 888940 kB\nLowFree: 397396 kB\nSwapTotal: 397304 kB\nSwapFree: 397304 kB\nDirty: 3200 kB\nWriteback: 0 kB\nAnonPages: 25728 kB\nMapped: 8068 kB\nShmem: 224 kB\nSlab: 27304 kB\nSReclaimable: 22576 kB\nSUnreclaim: 4728 kB\nKernelStack: 792 kB\nPageTables: 648 kB\nNFS_Unstable: 0 kB\nBounce: 0 kB\nWritebackTmp: 0 kB\nCommitLimit: 910352 kB\nCommitted_AS: 71204 kB\nVmallocTotal: 122880 kB\nVmallocUsed: 3540 kB\nVmallocChunk: 114424 kB\nHardwareCorrupted: 0 kB\nHugePages_Total: 0\nHugePages_Free: 0\nHugePages_Rsvd: 0\nHugePages_Surp: 0\nHugepagesize: 2048 kB\nDirectMap4k: 10232 kB\nDirectMap2M: 901120 kB\n">>}, {replication,[]}, {system_memory_data, [{system_total_memory, 1050726400}, {free_swap,406839296}, {total_swap,406839296}, {cached_memory,423735296}, {buffered_memory, 156311552}, {free_memory,407191552}, {total_memory, 1050726400}]}, {statistics, [{wall_clock,{412,155}}, {context_switches, {4716,0}}, {garbage_collection, {1057,2024145,0}}, {io, {{input,2189920}, {output,112828}}}, {reductions, {909617,909617}}, {run_queue,0}, {runtime,{230,230}}]}]}] PROGRESS REPORT <6516.94.0> 2011-05-07 17:05:03 =============================================================================== supervisor {local,menelaus_sup} started [{pid,<6516.95.0>}, {name,menelaus_web}, {mfa,{menelaus_web,start_link,[]}}, {restart_type,permanent}, {shutdown,5000}, {child_type,worker}] PROGRESS REPORT <6516.94.0> 2011-05-07 17:05:03 =============================================================================== supervisor {local,menelaus_sup} started [{pid,<6516.97.0>}, {name,menelaus_event}, {mfa,{menelaus_event,start_link,[]}}, {restart_type,transient}, {shutdown,5000}, {child_type,worker}] PROGRESS REPORT <6516.94.0> 2011-05-07 17:05:03 =============================================================================== supervisor {local,menelaus_sup} started [{pid,<6516.98.0>}, {name,hot_keys_keeper}, {mfa,{hot_keys_keeper,start_link,[]}}, {restart_type,permanent}, {shutdown,5000}, {child_type,worker}] INFO REPORT <6516.70.0> 2011-05-07 17:05:03 =============================================================================== ns_log: logging menelaus_app:1:Membase Server has started on web port 8091 on node 'ns_1@127.0.0.1'. PROGRESS REPORT <6516.70.0> 2011-05-07 17:05:03 =============================================================================== supervisor {local,ns_server_sup} started [{pid,<6516.94.0>}, {name,menelaus}, {mfa,{menelaus_app,start_subapp,[]}}, {restart_type,permanent}, {shutdown,infinity}, {child_type,supervisor}] PROGRESS REPORT <6516.99.0> 2011-05-07 17:05:03 =============================================================================== supervisor {local,ns_port_sup} started [{pid,<6516.100.0>}, {name,ns_port_init}, {mfa,{ns_port_init,start_link,[]}}, {restart_type,transient}, {shutdown,10}, {child_type,worker}] INFO REPORT <6516.101.0> 2011-05-07 17:05:03 =============================================================================== starting ns_port_server with delay of 5000 PROGRESS REPORT <6516.99.0> 2011-05-07 17:05:03 =============================================================================== supervisor {local,ns_port_sup} started [{pid,<6516.101.0>}, {name, {moxi,"./bin/moxi/moxi", ["-Z", "port_listen=11211,default_bucket_name=default,downstream_max=1024,downstream_conn_max=4,connect_max_errors=5,connect_retry_interval=30000,connect_timeout=400,auth_timeout=100,cycle=200,downstream_conn_queue_timeout=200,downstream_timeout=5000,wait_queue_timeout=200", "-z", "url=http://127.0.0.1:8091/pools/default/saslBucketsStreaming", "-p","0","-Y","y","-O","stderr",[]], [{env, [{"EVENT_NOSELECT","1"}, {"MOXI_SASL_PLAIN_USR",[]}, {"MOXI_SASL_PLAIN_PWD",[]}]}, use_stdio,stderr_to_stdout,stream]}}, {mfa, {supervisor_cushion,start_link, [moxi,5000,ns_port_server,start_link, [moxi,"./bin/moxi/moxi", ["-Z", "port_listen=11211,default_bucket_name=default,downstream_max=1024,downstream_conn_max=4,connect_max_errors=5,connect_retry_interval=30000,connect_timeout=400,auth_timeout=100,cycle=200,downstream_conn_queue_timeout=200,downstream_timeout=5000,wait_queue_timeout=200", "-z", "url=http://127.0.0.1:8091/pools/default/saslBucketsStreaming", "-p","0","-Y","y","-O","stderr",[]], [{env, [{"EVENT_NOSELECT","1"}, {"MOXI_SASL_PLAIN_USR",[]}, {"MOXI_SASL_PLAIN_PWD",[]}]}, use_stdio,stderr_to_stdout,stream]]]}}, {restart_type,permanent}, {shutdown,10}, {child_type,worker}] INFO REPORT <6516.103.0> 2011-05-07 17:05:03 =============================================================================== starting ns_port_server with delay of 5000 PROGRESS REPORT <6516.99.0> 2011-05-07 17:05:03 =============================================================================== supervisor {local,ns_port_sup} started [{pid,<6516.103.0>}, {name, {memcached,"./bin/memcached/memcached", ["-X","./bin/memcached/stdin_term_handler.so","-p","11210", "-E","./bin/bucket_engine/bucket_engine.so","-B","binary", "-r","-c","10000","-e", "admin=_admin;default_bucket_name=default;auto_create=false", []], [{env, [{"EVENT_NOSELECT","1"}, {"MEMCACHED_TOP_KEYS","100"}, {"ISASL_PWFILE", "/var/opt/membase/1.6.5.4r/data/ns_1/isasl.pw"}, {"ISASL_DB_CHECK_TIME","1"}]}, use_stdio,stderr_to_stdout,stream]}}, {mfa, {supervisor_cushion,start_link, [memcached,5000,ns_port_server,start_link, [memcached,"./bin/memcached/memcached", ["-X","./bin/memcached/stdin_term_handler.so","-p", "11210","-E","./bin/bucket_engine/bucket_engine.so","-B", "binary","-r","-c","10000","-e", "admin=_admin;default_bucket_name=default;auto_create=false", []], [{env, [{"EVENT_NOSELECT","1"}, {"MEMCACHED_TOP_KEYS","100"}, {"ISASL_PWFILE", "/var/opt/membase/1.6.5.4r/data/ns_1/isasl.pw"}, {"ISASL_DB_CHECK_TIME","1"}]}, use_stdio,stderr_to_stdout,stream]]]}}, {restart_type,permanent}, {shutdown,10}, {child_type,worker}] PROGRESS REPORT <6516.70.0> 2011-05-07 17:05:03 =============================================================================== supervisor {local,ns_server_sup} started [{pid,<6516.99.0>}, {name,ns_port_sup}, {mfa,{ns_port_sup,start_link,[]}}, {restart_type,permanent}, {shutdown,10}, {child_type,worker}] PROGRESS REPORT <6516.70.0> 2011-05-07 17:05:03 =============================================================================== supervisor {local,ns_server_sup} started [{pid,<6516.105.0>}, {name,ns_tick_event}, {mfa,{gen_event,start_link,[{local,ns_tick_event}]}}, {restart_type,permanent}, {shutdown,10}, {child_type,worker}] PROGRESS REPORT <6516.70.0> 2011-05-07 17:05:03 =============================================================================== supervisor {local,ns_server_sup} started [{pid,<6516.106.0>}, {name,ns_stats_event}, {mfa,{gen_event,start_link,[{local,ns_stats_event}]}}, {restart_type,permanent}, {shutdown,10}, {child_type,worker}] PROGRESS REPORT <6516.70.0> 2011-05-07 17:05:03 =============================================================================== supervisor {local,ns_server_sup} started [{pid,<6516.107.0>}, {name,ns_good_bucket_worker}, {mfa,{work_queue,start_link,[ns_good_bucket_worker]}}, {restart_type,permanent}, {shutdown,10}, {child_type,worker}] PROGRESS REPORT <6516.70.0> 2011-05-07 17:05:03 =============================================================================== supervisor {local,ns_server_sup} started [{pid,<6516.108.0>}, {name,ns_good_bucket_sup}, {mfa,{ns_bucket_sup,start_link, [ns_good_bucket_sup, #Fun, ns_good_bucket_worker]}}, {restart_type,permanent}, {shutdown,infinity}, {child_type,supervisor}] INFO REPORT <6516.70.0> 2011-05-07 17:05:03 =============================================================================== ns_1@127.0.0.1:misc:725: start_singleton(gen_fsm, ns_orchestrator, [], []): started as <6516.109.0> on 'ns_1@127.0.0.1' PROGRESS REPORT <6516.70.0> 2011-05-07 17:05:03 =============================================================================== supervisor {local,ns_server_sup} started [{pid,<6516.109.0>}, {name,ns_orchestrator}, {mfa,{ns_orchestrator,start_link,[]}}, {restart_type,permanent}, {shutdown,20}, {child_type,worker}] PROGRESS REPORT <6516.115.0> 2011-05-07 17:05:03 =============================================================================== supervisor {local,mnesia_sup} started [{pid,<6516.116.0>}, {name,mnesia_event}, {mfa,{mnesia_sup,start_event,[]}}, {restart_type,permanent}, {shutdown,30000}, {child_type,worker}] PROGRESS REPORT <6516.117.0> 2011-05-07 17:05:03 =============================================================================== supervisor {local,mnesia_kernel_sup} started [{pid,<6516.118.0>}, {name,mnesia_monitor}, {mfa,{mnesia_monitor,start,[]}}, {restart_type,permanent}, {shutdown,3000}, {child_type,worker}] PROGRESS REPORT <6516.117.0> 2011-05-07 17:05:03 =============================================================================== supervisor {local,mnesia_kernel_sup} started [{pid,<6516.119.0>}, {name,mnesia_subscr}, {mfa,{mnesia_subscr,start,[]}}, {restart_type,permanent}, {shutdown,3000}, {child_type,worker}] PROGRESS REPORT <6516.117.0> 2011-05-07 17:05:03 =============================================================================== supervisor {local,mnesia_kernel_sup} started [{pid,<6516.120.0>}, {name,mnesia_locker}, {mfa,{mnesia_locker,start,[]}}, {restart_type,permanent}, {shutdown,3000}, {child_type,worker}] PROGRESS REPORT <6516.117.0> 2011-05-07 17:05:03 =============================================================================== supervisor {local,mnesia_kernel_sup} started [{pid,<6516.121.0>}, {name,mnesia_recover}, {mfa,{mnesia_recover,start,[]}}, {restart_type,permanent}, {shutdown,180000}, {child_type,worker}] PROGRESS REPORT <6516.117.0> 2011-05-07 17:05:03 =============================================================================== supervisor {local,mnesia_kernel_sup} started [{pid,<6516.122.0>}, {name,mnesia_tm}, {mfa,{mnesia_tm,start,[]}}, {restart_type,permanent}, {shutdown,30000}, {child_type,worker}] PROGRESS REPORT <6516.117.0> 2011-05-07 17:05:03 =============================================================================== supervisor {local,mnesia_kernel_sup} started [{pid,<6516.123.0>}, {name,mnesia_checkpoint_sup}, {mfa,{mnesia_checkpoint_sup,start,[]}}, {restart_type,permanent}, {shutdown,infinity}, {child_type,supervisor}] PROGRESS REPORT <6516.117.0> 2011-05-07 17:05:03 =============================================================================== supervisor {local,mnesia_kernel_sup} started [{pid,<6516.124.0>}, {name,mnesia_snmp_sup}, {mfa,{mnesia_snmp_sup,start,[]}}, {restart_type,permanent}, {shutdown,infinity}, {child_type,supervisor}] PROGRESS REPORT <6516.117.0> 2011-05-07 17:05:03 =============================================================================== supervisor {local,mnesia_kernel_sup} started [{pid,<6516.125.0>}, {name,mnesia_controller}, {mfa,{mnesia_controller,start,[]}}, {restart_type,permanent}, {shutdown,3000}, {child_type,worker}] PROGRESS REPORT <6516.117.0> 2011-05-07 17:05:03 =============================================================================== supervisor {local,mnesia_kernel_sup} started [{pid,<6516.126.0>}, {name,mnesia_late_loader}, {mfa,{mnesia_late_loader,start,[]}}, {restart_type,permanent}, {shutdown,3000}, {child_type,worker}] PROGRESS REPORT <6516.115.0> 2011-05-07 17:05:03 =============================================================================== supervisor {local,mnesia_sup} started [{pid,<6516.117.0>}, {name,mnesia_kernel_sup}, {mfa,{mnesia_kernel_sup,start,[]}}, {restart_type,permanent}, {shutdown,infinity}, {child_type,supervisor}] PROGRESS REPORT <6516.7.0> 2011-05-07 17:05:03 =============================================================================== application mnesia started_at 'ns_1@127.0.0.1' PROGRESS REPORT <6516.27.0> 2011-05-07 17:05:03 =============================================================================== supervisor {local,kernel_safe_sup} started [{pid,<6516.129.0>}, {name,disk_log_sup}, {mfa,{disk_log_sup,start_link,[]}}, {restart_type,permanent}, {shutdown,1000}, {child_type,supervisor}] PROGRESS REPORT <6516.27.0> 2011-05-07 17:05:03 =============================================================================== supervisor {local,kernel_safe_sup} started [{pid,<6516.130.0>}, {name,disk_log_server}, {mfa,{disk_log_server,start_link,[]}}, {restart_type,permanent}, {shutdown,2000}, {child_type,worker}] PROGRESS REPORT <6516.27.0> 2011-05-07 17:05:03 =============================================================================== supervisor {local,kernel_safe_sup} started [{pid,<6516.133.0>}, {name,dets_sup}, {mfa,{dets_sup,start_link,[]}}, {restart_type,permanent}, {shutdown,1000}, {child_type,supervisor}] PROGRESS REPORT <6516.27.0> 2011-05-07 17:05:03 =============================================================================== supervisor {local,kernel_safe_sup} started [{pid,<6516.134.0>}, {name,dets}, {mfa,{dets_server,start_link,[]}}, {restart_type,permanent}, {shutdown,2000}, {child_type,worker}] INFO REPORT <6516.110.0> 2011-05-07 17:05:03 =============================================================================== ns_1@127.0.0.1:ns_mnesia:250: Committed schema to disk. INFO REPORT <6516.110.0> 2011-05-07 17:05:03 =============================================================================== ns_1@127.0.0.1:ns_mnesia:196: Current config: [{access_module,mnesia}, {auto_repair,true}, {backup_module,mnesia_backup}, {checkpoints,[]}, {db_nodes,['ns_1@127.0.0.1']}, {debug,verbose}, {directory, "/var/opt/membase/1.6.5.4r/mnesia"}, {dump_log_load_regulation, false}, {dump_log_time_threshold, 180000}, {dump_log_update_in_place,true}, {dump_log_write_threshold,1000}, {embedded_mnemosyne,false}, {event_module,mnesia_event}, {extra_db_nodes,[]}, {fallback_activated,false}, {held_locks,[]}, {ignore_fallback_at_startup, false}, {fallback_error_function, {mnesia,lkill}}, {is_running,yes}, {local_tables,[schema]}, {lock_queue,[]}, {log_version,"4.3"}, {master_node_tables,[]}, {max_wait_for_decision,10000}, {protocol_version,{7,6}}, {running_db_nodes, ['ns_1@127.0.0.1']}, {schema_location,opt_disc}, {schema_version,{3,0}}, {subscribers, [<6516.116.0>,<6516.110.0>]}, {tables,[schema]}, {transaction_commits,3}, {transaction_failures,0}, {transaction_log_writes,1}, {transaction_restarts,0}, {transactions,[]}, {use_dir,true}, {core_dir,false}, {no_table_loaders,2}, {dc_dump_limit,4}, {version,"4.4.12"}] INFO REPORT <6516.110.0> 2011-05-07 17:05:03 =============================================================================== ns_1@127.0.0.1:ns_mnesia:170: Mnesia table event: {write,schema, {schema,schema, [{name,schema}, {type,set}, {ram_copies,[]}, {disc_copies,['ns_1@127.0.0.1']}, {disc_only_copies,[]}, {load_order,0}, {access_mode,read_write}, {index,[]}, {snmp,[]}, {local_content,false}, {record_name,schema}, {attributes,[table,cstruct]}, {user_properties,[]}, {frag_properties,[]}, {cookie,{{1304,813103,390489},'ns_1@127.0.0.1'}}, {version,{{3,0},{'ns_1@127.0.0.1',{1304,813103,427023}}}}]}, [{schema,schema, [{name,schema}, {type,set}, {ram_copies,[]}, {disc_copies,[]}, {disc_only_copies,[]}, {load_order,0}, {access_mode,read_write}, {index,[]}, {snmp,[]}, {local_content,false}, {record_name,schema}, {attributes,[table,cstruct]}, {user_properties,[]}, {frag_properties,[]}, {cookie,{{1304,813103,390489},'ns_1@127.0.0.1'}}, {version,{{2,0},[]}}]}], {tid,3,<6516.128.0>}} PROGRESS REPORT <6516.70.0> 2011-05-07 17:05:03 =============================================================================== supervisor {local,ns_server_sup} started [{pid,<6516.110.0>}, {name,ns_mnesia}, {mfa,{ns_mnesia,start_link,[]}}, {restart_type,permanent}, {shutdown,10000}, {child_type,worker}] PROGRESS REPORT <6516.70.0> 2011-05-07 17:05:03 =============================================================================== supervisor {local,ns_server_sup} started [{pid,<6516.146.0>}, {name,ns_bad_bucket_worker}, {mfa,{work_queue,start_link,[ns_bad_bucket_worker]}}, {restart_type,permanent}, {shutdown,10}, {child_type,worker}] PROGRESS REPORT <6516.70.0> 2011-05-07 17:05:03 =============================================================================== supervisor {local,ns_server_sup} started [{pid,<6516.147.0>}, {name,ns_bad_bucket_sup}, {mfa,{ns_bucket_sup,start_link, [ns_bad_bucket_sup, #Fun, ns_bad_bucket_worker]}}, {restart_type,permanent}, {shutdown,infinity}, {child_type,supervisor}] PROGRESS REPORT <6516.70.0> 2011-05-07 17:05:03 =============================================================================== supervisor {local,ns_server_sup} started [{pid,<6516.148.0>}, {name,ns_moxi_sup}, {mfa,{ns_moxi_sup,start_link,[]}}, {restart_type,permanent}, {shutdown,infinity}, {child_type,supervisor}] INFO REPORT <6516.70.0> 2011-05-07 17:05:03 =============================================================================== ns_1@127.0.0.1:misc:725: start_singleton(gen_server, ns_tick, [], []): started as <6516.149.0> on 'ns_1@127.0.0.1' PROGRESS REPORT <6516.70.0> 2011-05-07 17:05:03 =============================================================================== supervisor {local,ns_server_sup} started [{pid,<6516.149.0>}, {name,ns_tick}, {mfa,{ns_tick,start_link,[]}}, {restart_type,permanent}, {shutdown,10}, {child_type,worker}] PROGRESS REPORT <6516.54.0> 2011-05-07 17:05:03 =============================================================================== supervisor {local,ns_server_cluster_sup} started [{pid,<6516.70.0>}, {name,ns_server_sup}, {mfa,{ns_server_sup,start_link,[]}}, {restart_type,permanent}, {shutdown,infinity}, {child_type,supervisor}] PROGRESS REPORT <6516.7.0> 2011-05-07 17:05:03 =============================================================================== application ns_server started_at 'ns_1@127.0.0.1' INFO REPORT <6516.92.0> 2011-05-07 17:06:03 =============================================================================== ns_1@127.0.0.1:ns_doctor:82: Current node statuses: [{'ns_1@127.0.0.1', [{last_heard,{1304,813163,60291}}, {active_buckets,[]}, {memory, [{total,11716944}, {processes,2261296}, {processes_used,2255744}, {system,9455648}, {atom,450225}, {atom_used,445169}, {binary,156560}, {code,3706907}, {ets,264012}]}, {cluster_compatibility_version,1}, {version, [{kernel,"2.13.4"}, {mnesia,"4.4.12"}, {stdlib,"1.16.4"}, {os_mon,"2.2.4"}, {ns_server,"1.6.5.4r"}, {menelaus,"1.6.5.4r"}, {sasl,"2.1.8"}]}, {system_arch,"i686-pc-linux-gnu"}, {wall_clock,60}, {memory_data,{1050726400,637935616,{<6516.70.0>,486376}}}, {disk_data, [{"/",7583436,17}, {"/dev",508460,1}, {"/dev/shm",513048,0}, {"/var/run",513048,1}, {"/var/lock",513048,0}, {"/lib/init/rw",513048,0}, {"/boot",233191,8}]}, {meminfo, <<"MemTotal: 1026100 kB\nMemFree: 403116 kB\nBuffers: 152648 kB\nCached: 413796 kB\nSwapCached: 0 kB\nActive: 329240 kB\nInactive: 257068 kB\nActive(anon): 15680 kB\nInactive(anon): 4412 kB\nActive(file): 313560 kB\nInactive(file): 252656 kB\nUnevictable: 0 kB\nMlocked: 0 kB\nHighTotal: 137160 kB\nHighFree: 8064 kB\nLowTotal: 888940 kB\nLowFree: 395052 kB\nSwapTotal: 397304 kB\nSwapFree: 397304 kB\nDirty: 28 kB\nWriteback: 0 kB\nAnonPages: 19876 kB\nMapped: 8600 kB\nShmem: 228 kB\nSlab: 27268 kB\nSReclaimable: 22536 kB\nSUnreclaim: 4732 kB\nKernelStack: 880 kB\nPageTables: 656 kB\nNFS_Unstable: 0 kB\nBounce: 0 kB\nWritebackTmp: 0 kB\nCommitLimit: 910352 kB\nCommitted_AS: 180992 kB\nVmallocTotal: 122880 kB\nVmallocUsed: 3540 kB\nVmallocChunk: 114424 kB\nHardwareCorrupted: 0 kB\nHugePages_Total: 0\nHugePages_Free: 0\nHugePages_Rsvd: 0\nHugePages_Surp: 0\nHugepagesize: 2048 kB\nDirectMap4k: 10232 kB\nDirectMap2M: 901120 kB\n">>}, {replication,[]}, {system_memory_data, [{system_total_memory,1050726400}, {free_swap,406839296}, {total_swap,406839296}, {cached_memory,423727104}, {buffered_memory,156311552}, {free_memory,412790784}, {total_memory,1050726400}]}, {statistics, [{wall_clock,{50409,0}}, {context_switches,{12552,0}}, {garbage_collection,{2553,4960540,0}}, {io,{{input,4818240},{output,1141750}}}, {reductions,{2211385,358365}}, {run_queue,0}, {runtime,{530,40}}]}]}] INFO REPORT <6516.64.0> 2011-05-07 17:06:55 =============================================================================== config change: {node,'ns_1@127.0.0.1',memcached} -> [{dbdir,"/var/opt/membase/1.6.5.4r/data/ns_1"}, {port,11210}, {admin_user,"_admin"}, {admin_pass,"_admin"}, {bucket_engine,"./bin/bucket_engine/bucket_engine.so"}, {engines,[{membase,[{engine,"bin/ep_engine/ep.so"}, {initfile,"/etc/opt/membase/1.6.5.4r/init.sql"}]}, {memcached,[{engine,"bin/memcached/default_engine.so"}]}]}, {verbosity,[]}] INFO REPORT <6516.64.0> 2011-05-07 17:06:55 =============================================================================== ns_node_disco_conf_events config all INFO REPORT <6516.85.0> 2011-05-07 17:06:55 =============================================================================== Pushing config INFO REPORT <6516.85.0> 2011-05-07 17:06:55 =============================================================================== Pushing config done INFO REPORT <6516.64.0> 2011-05-07 17:06:55 =============================================================================== config change: memory_quota -> 801 INFO REPORT <6516.64.0> 2011-05-07 17:06:55 =============================================================================== ns_node_disco_conf_events config all INFO REPORT <6516.85.0> 2011-05-07 17:06:55 =============================================================================== Pushing config INFO REPORT <6516.85.0> 2011-05-07 17:06:55 =============================================================================== Pushing config done INFO REPORT <6516.217.0> 2011-05-07 17:06:57 =============================================================================== ns_1@127.0.0.1:ns_storage_conf:279: Result of deleting file "/var/opt/membase/1.6.5.4r/data/ns_1/default": {error, enoent} INFO REPORT <6516.217.0> 2011-05-07 17:06:57 =============================================================================== ns_1@127.0.0.1:ns_storage_conf:279: Result of deleting file "/var/opt/membase/1.6.5.4r/data/ns_1/default-0.mb": {error, enoent} INFO REPORT <6516.217.0> 2011-05-07 17:06:57 =============================================================================== ns_1@127.0.0.1:ns_storage_conf:279: Result of deleting file "/var/opt/membase/1.6.5.4r/data/ns_1/default-1.mb": {error, enoent} INFO REPORT <6516.217.0> 2011-05-07 17:06:57 =============================================================================== ns_1@127.0.0.1:ns_storage_conf:279: Result of deleting file "/var/opt/membase/1.6.5.4r/data/ns_1/default-2.mb": {error, enoent} INFO REPORT <6516.217.0> 2011-05-07 17:06:57 =============================================================================== ns_1@127.0.0.1:ns_storage_conf:279: Result of deleting file "/var/opt/membase/1.6.5.4r/data/ns_1/default-3.mb": {error, enoent} INFO REPORT <6516.177.0> 2011-05-07 17:06:57 =============================================================================== ns_log: logging menelaus_web:12:Created bucket "default" of type: membase INFO REPORT <6516.64.0> 2011-05-07 17:06:57 =============================================================================== config change: buckets -> [{configs,[{"default", [{num_replicas,1}, {ram_quota,839909376}, {auth_type,sasl}, {sasl_password,[]}, {type,membase}, {num_vbuckets,1024}, {db_shards,4}, {ht_size,3079}, {tap_keepalive,0}, {tap_noop_interval,20}, {max_txn_size,1000}, {ht_locks,5}, {servers,[]}, {map,undefined}]}]}] INFO REPORT <6516.64.0> 2011-05-07 17:06:57 =============================================================================== Writing isasl passwd file: "/var/opt/membase/1.6.5.4r/data/ns_1/isasl.pw" INFO REPORT <6516.64.0> 2011-05-07 17:06:57 =============================================================================== ns_node_disco_conf_events config all INFO REPORT <6516.85.0> 2011-05-07 17:06:57 =============================================================================== Pushing config INFO REPORT <6516.85.0> 2011-05-07 17:06:57 =============================================================================== Pushing config done INFO REPORT <6516.146.0> 2011-05-07 17:06:57 =============================================================================== ns_1@127.0.0.1:ns_bucket_sup:75: ns_bad_bucket_sup: Starting new child: {{stats_collector, "default"}, {stats_collector, start_link, ["default"]}, permanent, 10, worker, [stats_collector]} INFO REPORT <6516.107.0> 2011-05-07 17:06:57 =============================================================================== ns_1@127.0.0.1:ns_bucket_sup:75: ns_good_bucket_sup: Starting new child: {{ns_vbm_sup, "default"}, {ns_vbm_sup, start_link, ["default"]}, permanent, 1000, worker, [ns_vbm_sup]} INFO REPORT <6516.64.0> 2011-05-07 17:06:57 =============================================================================== config change: buckets -> [{configs,[{"default", [{num_replicas,1}, {ram_quota,839909376}, {auth_type,sasl}, {sasl_password,[]}, {type,membase}, {num_vbuckets,1024}, {db_shards,4}, {ht_size,3079}, {tap_keepalive,0}, {tap_noop_interval,20}, {max_txn_size,1000}, {ht_locks,5}, {servers,['ns_1@127.0.0.1']}, {map,undefined}]}]}] INFO REPORT <6516.64.0> 2011-05-07 17:06:57 =============================================================================== ns_node_disco_conf_events config all INFO REPORT <6516.85.0> 2011-05-07 17:06:57 =============================================================================== Pushing config INFO REPORT <6516.85.0> 2011-05-07 17:06:57 =============================================================================== Pushing config done PROGRESS REPORT <6516.147.0> 2011-05-07 17:06:57 =============================================================================== supervisor {local,ns_bad_bucket_sup} started [{pid,<6516.224.0>}, {name,{stats_collector,"default"}}, {mfa,{stats_collector,start_link,["default"]}}, {restart_type,permanent}, {shutdown,10}, {child_type,worker}] INFO REPORT <6516.146.0> 2011-05-07 17:06:57 =============================================================================== ns_1@127.0.0.1:ns_bucket_sup:75: ns_bad_bucket_sup: Starting new child: {{stats_archiver, "default"}, {stats_archiver, start_link, ["default"]}, permanent, 10, worker, [stats_archiver]} INFO REPORT <6516.64.0> 2011-05-07 17:06:57 =============================================================================== config change: buckets -> [{configs,[{"default", [{num_replicas,1}, {ram_quota,839909376}, {auth_type,sasl}, {sasl_password,[]}, {type,membase}, {num_vbuckets,1024}, {db_shards,4}, {ht_size,3079}, {tap_keepalive,0}, {tap_noop_interval,20}, {max_txn_size,1000}, {ht_locks,5}, {servers,['ns_1@127.0.0.1']}, {map,[['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1'|...], [...]|...]}]}]}] PROGRESS REPORT <6516.108.0> 2011-05-07 17:06:57 =============================================================================== supervisor {local,ns_good_bucket_sup} started [{pid,<6516.226.0>}, {name,{ns_vbm_sup,"default"}}, {mfa,{ns_vbm_sup,start_link,["default"]}}, {restart_type,permanent}, {shutdown,1000}, {child_type,worker}] INFO REPORT <6516.107.0> 2011-05-07 17:06:57 =============================================================================== ns_1@127.0.0.1:ns_bucket_sup:75: ns_good_bucket_sup: Starting new child: {{ns_memcached, "default"}, {ns_memcached, start_link, ["default"]}, permanent, 86400000, worker, [ns_memcached]} PROGRESS REPORT <6516.108.0> 2011-05-07 17:06:57 =============================================================================== supervisor {local,ns_good_bucket_sup} started [{pid,<6516.227.0>}, {name,{ns_memcached,"default"}}, {mfa,{ns_memcached,start_link,["default"]}}, {restart_type,permanent}, {shutdown,86400000}, {child_type,worker}] PROGRESS REPORT <6516.228.0> 2011-05-07 17:06:57 =============================================================================== supervisor {local,inet_gethost_native_sup} started [{pid,<6516.229.0>},{mfa,{inet_gethost_native,init,[[]]}}] INFO REPORT <6516.64.0> 2011-05-07 17:06:57 =============================================================================== ns_node_disco_conf_events config all PROGRESS REPORT <6516.27.0> 2011-05-07 17:06:57 =============================================================================== supervisor {local,kernel_safe_sup} started [{pid,<6516.228.0>}, {name,inet_gethost_native_sup}, {mfa,{inet_gethost_native,start_link,[]}}, {restart_type,temporary}, {shutdown,1000}, {child_type,worker}] INFO REPORT <6516.85.0> 2011-05-07 17:06:57 =============================================================================== Pushing config INFO REPORT <6516.85.0> 2011-05-07 17:06:57 =============================================================================== Pushing config done INFO REPORT <6516.110.0> 2011-05-07 17:06:57 =============================================================================== ns_1@127.0.0.1:ns_mnesia:120: Created table 'stats_archiver-default-minute' INFO REPORT <6516.110.0> 2011-05-07 17:06:57 =============================================================================== ns_1@127.0.0.1:ns_mnesia:170: Mnesia table event: {write,schema, {schema,'stats_archiver-default-minute', [{name,'stats_archiver-default-minute'}, {type,ordered_set}, {ram_copies,[]}, {disc_copies,['ns_1@127.0.0.1']}, {disc_only_copies,[]}, {load_order,0}, {access_mode,read_write}, {index,[]}, {snmp,[]}, {local_content,true}, {record_name,stat_entry}, {attributes,[timestamp,values]}, {user_properties,[]}, {frag_properties,[]}, {cookie,{{1304,813217,421213},'ns_1@127.0.0.1'}}, {version,{{2,0},[]}}]}, [], {tid,4,<6516.232.0>}} INFO REPORT <6516.110.0> 2011-05-07 17:06:57 =============================================================================== ns_1@127.0.0.1:ns_mnesia:170: Mnesia table event: {write,schema, {schema,'stats_archiver-default-minute', [{name,'stats_archiver-default-minute'}, {type,ordered_set}, {ram_copies,[]}, {disc_copies,['ns_1@127.0.0.1']}, {disc_only_copies,[]}, {load_order,0}, {access_mode,read_write}, {index,[]}, {snmp,[]}, {local_content,true}, {record_name,stat_entry}, {attributes,[timestamp,values]}, {user_properties,[]}, {frag_properties,[]}, {cookie,{{1304,813217,421213},'ns_1@127.0.0.1'}}, {version,{{2,0},[]}}]}, [{schema,'stats_archiver-default-minute', [{name,'stats_archiver-default-minute'}, {type,ordered_set}, {ram_copies,[]}, {disc_copies,['ns_1@127.0.0.1']}, {disc_only_copies,[]}, {load_order,0}, {access_mode,read_write}, {index,[]}, {snmp,[]}, {local_content,true}, {record_name,stat_entry}, {attributes,[timestamp,values]}, {user_properties,[]}, {frag_properties,[]}, {cookie,{{1304,813217,421213},'ns_1@127.0.0.1'}}, {version,{{2,0},[]}}]}], {tid,4,<6516.232.0>}} INFO REPORT <6516.110.0> 2011-05-07 17:06:57 =============================================================================== ns_1@127.0.0.1:ns_mnesia:120: Created table 'stats_archiver-default-hour' INFO REPORT <6516.110.0> 2011-05-07 17:06:57 =============================================================================== ns_1@127.0.0.1:ns_mnesia:170: Mnesia table event: {write,schema, {schema,'stats_archiver-default-hour', [{name,'stats_archiver-default-hour'}, {type,ordered_set}, {ram_copies,[]}, {disc_copies,['ns_1@127.0.0.1']}, {disc_only_copies,[]}, {load_order,0}, {access_mode,read_write}, {index,[]}, {snmp,[]}, {local_content,true}, {record_name,stat_entry}, {attributes,[timestamp,values]}, {user_properties,[]}, {frag_properties,[]}, {cookie,{{1304,813217,443046},'ns_1@127.0.0.1'}}, {version,{{2,0},[]}}]}, [], {tid,5,<6516.238.0>}} INFO REPORT <6516.110.0> 2011-05-07 17:06:57 =============================================================================== ns_1@127.0.0.1:ns_mnesia:170: Mnesia table event: {write,schema, {schema,'stats_archiver-default-hour', [{name,'stats_archiver-default-hour'}, {type,ordered_set}, {ram_copies,[]}, {disc_copies,['ns_1@127.0.0.1']}, {disc_only_copies,[]}, {load_order,0}, {access_mode,read_write}, {index,[]}, {snmp,[]}, {local_content,true}, {record_name,stat_entry}, {attributes,[timestamp,values]}, {user_properties,[]}, {frag_properties,[]}, {cookie,{{1304,813217,443046},'ns_1@127.0.0.1'}}, {version,{{2,0},[]}}]}, [{schema,'stats_archiver-default-hour', [{name,'stats_archiver-default-hour'}, {type,ordered_set}, {ram_copies,[]}, {disc_copies,['ns_1@127.0.0.1']}, {disc_only_copies,[]}, {load_order,0}, {access_mode,read_write}, {index,[]}, {snmp,[]}, {local_content,true}, {record_name,stat_entry}, {attributes,[timestamp,values]}, {user_properties,[]}, {frag_properties,[]}, {cookie,{{1304,813217,443046},'ns_1@127.0.0.1'}}, {version,{{2,0},[]}}]}], {tid,5,<6516.238.0>}} INFO REPORT <6516.110.0> 2011-05-07 17:06:57 =============================================================================== ns_1@127.0.0.1:ns_mnesia:120: Created table 'stats_archiver-default-day' INFO REPORT <6516.110.0> 2011-05-07 17:06:57 =============================================================================== ns_1@127.0.0.1:ns_mnesia:170: Mnesia table event: {write,schema, {schema,'stats_archiver-default-day', [{name,'stats_archiver-default-day'}, {type,ordered_set}, {ram_copies,[]}, {disc_copies,['ns_1@127.0.0.1']}, {disc_only_copies,[]}, {load_order,0}, {access_mode,read_write}, {index,[]}, {snmp,[]}, {local_content,true}, {record_name,stat_entry}, {attributes,[timestamp,values]}, {user_properties,[]}, {frag_properties,[]}, {cookie,{{1304,813217,455872},'ns_1@127.0.0.1'}}, {version,{{2,0},[]}}]}, [], {tid,6,<6516.244.0>}} INFO REPORT <6516.110.0> 2011-05-07 17:06:57 =============================================================================== ns_1@127.0.0.1:ns_mnesia:170: Mnesia table event: {write,schema, {schema,'stats_archiver-default-day', [{name,'stats_archiver-default-day'}, {type,ordered_set}, {ram_copies,[]}, {disc_copies,['ns_1@127.0.0.1']}, {disc_only_copies,[]}, {load_order,0}, {access_mode,read_write}, {index,[]}, {snmp,[]}, {local_content,true}, {record_name,stat_entry}, {attributes,[timestamp,values]}, {user_properties,[]}, {frag_properties,[]}, {cookie,{{1304,813217,455872},'ns_1@127.0.0.1'}}, {version,{{2,0},[]}}]}, [{schema,'stats_archiver-default-day', [{name,'stats_archiver-default-day'}, {type,ordered_set}, {ram_copies,[]}, {disc_copies,['ns_1@127.0.0.1']}, {disc_only_copies,[]}, {load_order,0}, {access_mode,read_write}, {index,[]}, {snmp,[]}, {local_content,true}, {record_name,stat_entry}, {attributes,[timestamp,values]}, {user_properties,[]}, {frag_properties,[]}, {cookie,{{1304,813217,455872},'ns_1@127.0.0.1'}}, {version,{{2,0},[]}}]}], {tid,6,<6516.244.0>}} INFO REPORT <6516.110.0> 2011-05-07 17:06:57 =============================================================================== ns_1@127.0.0.1:ns_mnesia:120: Created table 'stats_archiver-default-week' INFO REPORT <6516.110.0> 2011-05-07 17:06:57 =============================================================================== ns_1@127.0.0.1:ns_mnesia:170: Mnesia table event: {write,schema, {schema,'stats_archiver-default-week', [{name,'stats_archiver-default-week'}, {type,ordered_set}, {ram_copies,[]}, {disc_copies,['ns_1@127.0.0.1']}, {disc_only_copies,[]}, {load_order,0}, {access_mode,read_write}, {index,[]}, {snmp,[]}, {local_content,true}, {record_name,stat_entry}, {attributes,[timestamp,values]}, {user_properties,[]}, {frag_properties,[]}, {cookie,{{1304,813217,463265},'ns_1@127.0.0.1'}}, {version,{{2,0},[]}}]}, [], {tid,7,<6516.250.0>}} INFO REPORT <6516.110.0> 2011-05-07 17:06:57 =============================================================================== ns_1@127.0.0.1:ns_mnesia:170: Mnesia table event: {write,schema, {schema,'stats_archiver-default-week', [{name,'stats_archiver-default-week'}, {type,ordered_set}, {ram_copies,[]}, {disc_copies,['ns_1@127.0.0.1']}, {disc_only_copies,[]}, {load_order,0}, {access_mode,read_write}, {index,[]}, {snmp,[]}, {local_content,true}, {record_name,stat_entry}, {attributes,[timestamp,values]}, {user_properties,[]}, {frag_properties,[]}, {cookie,{{1304,813217,463265},'ns_1@127.0.0.1'}}, {version,{{2,0},[]}}]}, [{schema,'stats_archiver-default-week', [{name,'stats_archiver-default-week'}, {type,ordered_set}, {ram_copies,[]}, {disc_copies,['ns_1@127.0.0.1']}, {disc_only_copies,[]}, {load_order,0}, {access_mode,read_write}, {index,[]}, {snmp,[]}, {local_content,true}, {record_name,stat_entry}, {attributes,[timestamp,values]}, {user_properties,[]}, {frag_properties,[]}, {cookie,{{1304,813217,463265},'ns_1@127.0.0.1'}}, {version,{{2,0},[]}}]}], {tid,7,<6516.250.0>}} INFO REPORT <6516.110.0> 2011-05-07 17:06:57 =============================================================================== ns_1@127.0.0.1:ns_mnesia:120: Created table 'stats_archiver-default-month' INFO REPORT <6516.110.0> 2011-05-07 17:06:57 =============================================================================== ns_1@127.0.0.1:ns_mnesia:170: Mnesia table event: {write,schema, {schema,'stats_archiver-default-month', [{name,'stats_archiver-default-month'}, {type,ordered_set}, {ram_copies,[]}, {disc_copies,['ns_1@127.0.0.1']}, {disc_only_copies,[]}, {load_order,0}, {access_mode,read_write}, {index,[]}, {snmp,[]}, {local_content,true}, {record_name,stat_entry}, {attributes,[timestamp,values]}, {user_properties,[]}, {frag_properties,[]}, {cookie,{{1304,813217,469867},'ns_1@127.0.0.1'}}, {version,{{2,0},[]}}]}, [], {tid,8,<6516.256.0>}} INFO REPORT <6516.110.0> 2011-05-07 17:06:57 =============================================================================== ns_1@127.0.0.1:ns_mnesia:170: Mnesia table event: {write,schema, {schema,'stats_archiver-default-month', [{name,'stats_archiver-default-month'}, {type,ordered_set}, {ram_copies,[]}, {disc_copies,['ns_1@127.0.0.1']}, {disc_only_copies,[]}, {load_order,0}, {access_mode,read_write}, {index,[]}, {snmp,[]}, {local_content,true}, {record_name,stat_entry}, {attributes,[timestamp,values]}, {user_properties,[]}, {frag_properties,[]}, {cookie,{{1304,813217,469867},'ns_1@127.0.0.1'}}, {version,{{2,0},[]}}]}, [{schema,'stats_archiver-default-month', [{name,'stats_archiver-default-month'}, {type,ordered_set}, {ram_copies,[]}, {disc_copies,['ns_1@127.0.0.1']}, {disc_only_copies,[]}, {load_order,0}, {access_mode,read_write}, {index,[]}, {snmp,[]}, {local_content,true}, {record_name,stat_entry}, {attributes,[timestamp,values]}, {user_properties,[]}, {frag_properties,[]}, {cookie,{{1304,813217,469867},'ns_1@127.0.0.1'}}, {version,{{2,0},[]}}]}], {tid,8,<6516.256.0>}} INFO REPORT <6516.110.0> 2011-05-07 17:06:57 =============================================================================== ns_1@127.0.0.1:ns_mnesia:120: Created table 'stats_archiver-default-year' INFO REPORT <6516.110.0> 2011-05-07 17:06:57 =============================================================================== ns_1@127.0.0.1:ns_mnesia:170: Mnesia table event: {write,schema, {schema,'stats_archiver-default-year', [{name,'stats_archiver-default-year'}, {type,ordered_set}, {ram_copies,[]}, {disc_copies,['ns_1@127.0.0.1']}, {disc_only_copies,[]}, {load_order,0}, {access_mode,read_write}, {index,[]}, {snmp,[]}, {local_content,true}, {record_name,stat_entry}, {attributes,[timestamp,values]}, {user_properties,[]}, {frag_properties,[]}, {cookie,{{1304,813217,476239},'ns_1@127.0.0.1'}}, {version,{{2,0},[]}}]}, [], {tid,9,<6516.262.0>}} INFO REPORT <6516.110.0> 2011-05-07 17:06:57 =============================================================================== ns_1@127.0.0.1:ns_mnesia:170: Mnesia table event: {write,schema, {schema,'stats_archiver-default-year', [{name,'stats_archiver-default-year'}, {type,ordered_set}, {ram_copies,[]}, {disc_copies,['ns_1@127.0.0.1']}, {disc_only_copies,[]}, {load_order,0}, {access_mode,read_write}, {index,[]}, {snmp,[]}, {local_content,true}, {record_name,stat_entry}, {attributes,[timestamp,values]}, {user_properties,[]}, {frag_properties,[]}, {cookie,{{1304,813217,476239},'ns_1@127.0.0.1'}}, {version,{{2,0},[]}}]}, [{schema,'stats_archiver-default-year', [{name,'stats_archiver-default-year'}, {type,ordered_set}, {ram_copies,[]}, {disc_copies,['ns_1@127.0.0.1']}, {disc_only_copies,[]}, {load_order,0}, {access_mode,read_write}, {index,[]}, {snmp,[]}, {local_content,true}, {record_name,stat_entry}, {attributes,[timestamp,values]}, {user_properties,[]}, {frag_properties,[]}, {cookie,{{1304,813217,476239},'ns_1@127.0.0.1'}}, {version,{{2,0},[]}}]}], {tid,9,<6516.262.0>}} PROGRESS REPORT <6516.147.0> 2011-05-07 17:06:57 =============================================================================== supervisor {local,ns_bad_bucket_sup} started [{pid,<6516.231.0>}, {name,{stats_archiver,"default"}}, {mfa,{stats_archiver,start_link,["default"]}}, {restart_type,permanent}, {shutdown,10}, {child_type,worker}] INFO REPORT <6516.146.0> 2011-05-07 17:06:57 =============================================================================== ns_1@127.0.0.1:ns_bucket_sup:75: ns_bad_bucket_sup: Starting new child: {{stats_reader, "default"}, {stats_reader, start_link, ["default"]}, permanent, 10, worker, [stats_reader]} PROGRESS REPORT <6516.147.0> 2011-05-07 17:06:57 =============================================================================== supervisor {local,ns_bad_bucket_sup} started [{pid,<6516.268.0>}, {name,{stats_reader,"default"}}, {mfa,{stats_reader,start_link,["default"]}}, {restart_type,permanent}, {shutdown,10}, {child_type,worker}] SUPERVISOR REPORT <6516.94.0> 2011-05-07 17:06:58 =============================================================================== Reporting supervisor {local,menelaus_sup} Child process errorContext child_terminated reason {noproc,{gen_server,call,['ns_memcached-default',topkeys,30000]}} pid <6516.98.0> name hot_keys_keeper start_function {hot_keys_keeper,start_link,[]} restart_type permanent shutdown 5000 child_type worker PROGRESS REPORT <6516.94.0> 2011-05-07 17:06:58 =============================================================================== supervisor {local,menelaus_sup} started [{pid,<6516.270.0>}, {name,hot_keys_keeper}, {mfa,{hot_keys_keeper,start_link,[]}}, {restart_type,permanent}, {shutdown,5000}, {child_type,worker}] INFO REPORT <6516.102.0> 2011-05-07 17:06:58 =============================================================================== moxi<0.102.0>: 2011-05-07 17:08:53: (agent_config.c.650) ERROR: bad JSON configuration: Empty serverList ({ moxi<0.102.0>: "name": "default", moxi<0.102.0>: "nodeLocator": "vbucket", moxi<0.102.0>: "saslPassword": "", moxi<0.102.0>: "nodes": [], moxi<0.102.0>: "vBucketServerMap": { moxi<0.102.0>: "hashAlgorithm": "CRC", moxi<0.102.0>: "numReplicas": 1, moxi<0.102.0>: "serverList": [], moxi<0.102.0>: "vBucketMap": [[-1, -1], [-1, -1], [-1, -1], [-1, -1], [-1, -1], [-1, -1], [-1, -1], [-1, -1], [-1, -1], [-1, -1], [-1, -1], [-1, -1], [-1, -1], [-1, -1], [-1, -1], [-1, -1], [-1, -1], [-1, -1], [-1, -1], [-1, -1], [-1, -1], [-1, -1], [-1, -1], [-1, -1], [-1, -1], [-1, -1], [-1, -1], [-1, -1], [-1, -1], [-1, -1], [-1, -1], [-1, -1], [-1, -1], [-1, -1], [-1, -1], [-1, -1], [-1, -1], [-1, -1], [-1, -1], [-1, -1], [-1, -1], [-1, -1], [-1, -1], [-1, -1], [-1, -1], [-1, -1], [-1, -1], [-1, -1], [-1, -1], [-1, -1], [-1, -1], [-1, -1], [-1, -1], [-1, -1], [-1, -1], [-1, -1], [-1, -1], [-1, -1], [-1, -1], [-1, -1], [-1, -1], [-1, -1], [-1, -1], [-1, -1], [-1, -1], [-1, -1], [-1, -1], [-1, -1], [-1, -1], [-1, -1], [-1, -1], [-1, -1 INFO REPORT <6516.219.0> 2011-05-07 17:06:58 =============================================================================== ns_1@127.0.0.1:ns_janitor:270: Waiting for "default" on ['ns_1@127.0.0.1'] INFO REPORT <6516.227.0> 2011-05-07 17:06:58 =============================================================================== ns_1@127.0.0.1:ns_memcached:390: Created bucket "default" with config string "vb0=false;waitforwarmup=false;ht_size=3079;ht_locks=5;failpartialwarmup=false;db_shards=4;shardpattern=%d/%b-%i.mb;db_strategy=multiMTVBDB;tap_keepalive=0;tap_noop_interval=20;max_txn_size=1000;max_size=839909376;initfile=/etc/opt/membase/1.6.5.4r/init.sql;tap_keepalive=300;dbname=/var/opt/membase/1.6.5.4r/data/ns_1/default;" INFO REPORT <6516.104.0> 2011-05-07 17:06:58 =============================================================================== memcached<0.104.0>: WARNING: Found duplicate entry for "tap_keepalive" INFO REPORT <6516.219.0> 2011-05-07 17:06:59 =============================================================================== ns_1@127.0.0.1:ns_janitor:270: Waiting for "default" on ['ns_1@127.0.0.1'] INFO REPORT <6516.227.0> 2011-05-07 17:06:59 =============================================================================== ns_log: logging ns_memcached:1:Bucket "default" loaded on node 'ns_1@127.0.0.1' in 1 seconds. INFO REPORT <6516.224.0> 2011-05-07 17:06:59 =============================================================================== ns_1@127.0.0.1:stats_collector:71: Stats for bucket "default": auth_cmds 0 auth_errors 0 bucket_conns 1 bytes_read 72 bytes_written 9312 cas_badval 0 cas_hits 0 cas_misses 0 cmd_flush 0 cmd_get 0 cmd_set 0 conn_yields 0 connection_structures 11 curr_connections 11 curr_items 0 curr_items_tot 0 daemon_connections 10 decr_hits 0 decr_misses 0 delete_hits 0 delete_misses 0 ep_bg_fetched 0 ep_commit_num 0 ep_commit_time 0 ep_commit_time_total 0 ep_data_age 0 ep_data_age_highwat 0 ep_db_cleaner_status complete ep_db_strategy multiMTVBDB ep_dbinit 1 ep_dbname /var/opt/membase/1.6.5.4r/data/ns_1/default ep_dbshards 4 ep_expired 0 ep_flush_duration 0 ep_flush_duration_highwat 0 ep_flush_duration_total 0 ep_flush_preempts 0 ep_flusher_state running ep_flusher_todo 0 ep_io_num_read 0 ep_io_num_write 0 ep_io_read_bytes 0 ep_io_write_bytes 0 ep_item_begin_failed 0 ep_item_commit_failed 0 ep_item_flush_expired 0 ep_item_flush_failed 0 ep_kv_size 0 ep_max_data_size 839909376 ep_max_txn_size 1000 ep_mem_high_wat 629932032 ep_mem_low_wat 503945625 ep_min_data_age 0 ep_num_active_non_resident 0 ep_num_eject_failures 0 ep_num_eject_replicas 0 ep_num_expiry_pager_runs 0 ep_num_non_resident 0 ep_num_not_my_vbuckets 0 ep_num_pager_runs 0 ep_num_value_ejects 0 ep_oom_errors 0 ep_overhead 288 ep_pending_ops 0 ep_pending_ops_max 0 ep_pending_ops_max_duration 0 ep_pending_ops_total 0 ep_queue_age_cap 900 ep_queue_size 0 ep_storage_age 0 ep_storage_age_highwat 0 ep_storage_type featured ep_store_max_concurrency 10 ep_store_max_readers 9 ep_store_max_readwrite 1 ep_tap_bg_fetch_requeued 0 ep_tap_bg_fetched 0 ep_tap_keepalive 300 ep_tmp_oom_errors 0 ep_too_old 0 ep_too_young 0 ep_total_cache_size 0 ep_total_del_items 0 ep_total_enqueued 0 ep_total_new_items 0 ep_total_persisted 0 ep_vbucket_del 0 ep_vbucket_del_fail 0 ep_version 1.6.5.3.1_8_gce31baa ep_warmed_up 0 ep_warmup true ep_warmup_dups 0 ep_warmup_oom 0 ep_warmup_thread complete ep_warmup_time 29657 get_hits 0 get_misses 0 incr_hits 0 incr_misses 0 libevent 2.0.7-rc limit_maxbytes 67108864 mem_used 288 pid 5769 pointer_size 32 rejected_conns 0 rusage_system 0.268016 rusage_user 0.636039 threads 4 time 1304813219 total_connections 11 uptime 118 version 1.4.4_382_g9df3289 INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:270: Waiting for "default" on ['ns_1@127.0.0.1'] INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 0 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 1 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 2 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 3 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 4 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 5 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 6 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 7 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 8 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 9 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 10 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 11 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 12 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 13 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 14 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 15 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 16 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 17 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 18 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 19 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 20 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 21 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 22 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 23 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 24 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 25 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 26 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 27 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 28 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 29 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 30 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 31 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 32 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 33 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 34 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 35 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 36 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 37 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 38 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 39 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 40 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 41 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 42 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 43 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 44 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 45 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 46 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 47 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 48 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 49 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 50 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 51 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 52 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 53 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 54 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 55 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 56 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 57 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 58 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 59 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 60 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 61 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 62 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 63 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 64 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 65 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 66 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 67 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 68 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 69 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 70 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 71 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 72 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 73 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 74 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 75 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 76 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 77 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 78 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 79 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 80 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 81 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 82 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 83 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 84 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 85 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 86 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 87 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 88 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 89 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 90 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 91 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 92 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 93 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 94 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 95 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 96 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 97 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 98 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 99 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 100 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 101 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 102 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 103 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 104 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 105 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 106 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 107 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 108 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 109 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 110 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 111 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 112 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 113 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 114 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 115 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 116 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 117 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 118 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 119 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 120 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 121 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 122 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 123 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 124 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 125 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 126 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 127 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 128 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 129 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 130 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 131 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 132 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 133 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 134 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 135 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 136 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 137 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 138 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 139 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 140 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 141 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 142 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 143 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 144 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 145 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 146 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 147 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 148 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 149 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 150 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 151 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 152 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 153 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 154 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 155 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 156 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 157 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 158 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 159 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 160 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 161 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 162 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 163 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 164 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 165 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 166 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 167 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 168 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 169 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 170 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 171 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 172 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 173 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 174 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 175 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 176 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 177 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 178 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 179 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 180 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 181 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 182 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 183 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 184 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 185 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 186 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 187 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 188 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 189 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 190 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 191 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 192 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 193 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 194 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 195 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 196 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 197 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 198 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 199 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 200 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 201 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 202 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 203 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 204 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 205 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 206 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 207 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 208 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 209 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 210 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 211 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 212 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 213 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 214 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 215 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 216 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 217 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 218 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 219 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 220 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 221 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 222 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 223 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 224 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 225 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 226 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 227 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 228 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 229 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 230 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 231 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 232 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 233 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 234 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 235 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 236 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 237 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 238 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 239 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 240 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 241 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 242 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 243 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 244 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 245 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 246 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 247 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 248 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 249 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 250 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 251 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 252 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 253 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 254 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 255 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 256 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 257 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 258 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 259 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 260 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 261 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 262 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 263 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 264 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 265 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 266 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 267 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 268 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 269 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 270 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 271 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 272 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 273 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 274 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 275 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 276 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 277 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 278 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 279 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 280 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 281 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 282 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 283 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 284 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 285 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 286 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 287 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 288 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 289 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 290 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 291 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 292 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 293 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 294 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 295 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 296 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 297 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 298 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 299 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 300 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 301 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 302 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 303 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 304 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 305 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 306 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 307 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 308 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 309 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 310 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 311 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 312 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 313 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 314 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 315 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 316 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 317 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 318 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 319 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 320 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 321 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 322 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 323 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 324 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 325 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 326 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 327 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 328 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 329 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 330 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 331 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 332 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 333 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 334 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 335 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 336 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 337 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 338 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 339 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 340 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 341 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 342 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 343 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 344 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 345 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 346 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 347 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 348 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 349 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 350 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 351 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 352 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 353 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 354 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 355 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 356 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 357 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 358 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 359 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 360 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 361 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 362 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 363 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 364 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 365 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 366 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 367 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 368 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 369 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 370 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 371 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 372 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 373 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 374 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 375 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 376 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 377 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 378 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 379 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 380 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 381 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 382 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 383 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 384 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 385 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 386 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 387 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 388 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 389 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 390 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 391 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 392 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 393 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 394 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 395 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 396 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 397 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 398 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 399 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 400 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 401 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 402 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 403 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 404 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 405 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 406 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 407 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 408 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 409 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 410 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 411 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 412 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 413 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 414 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 415 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 416 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 417 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 418 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 419 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 420 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 421 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 422 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 423 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 424 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 425 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 426 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 427 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 428 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 429 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 430 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 431 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 432 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 433 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 434 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 435 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 436 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 437 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 438 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 439 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 440 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 441 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 442 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 443 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 444 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 445 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 446 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 447 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 448 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 449 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 450 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 451 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 452 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 453 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 454 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 455 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 456 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 457 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 458 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 459 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 460 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 461 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 462 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 463 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 464 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 465 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 466 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 467 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 468 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 469 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 470 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 471 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 472 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 473 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 474 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 475 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 476 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 477 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 478 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 479 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 480 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 481 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 482 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 483 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 484 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 485 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 486 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 487 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 488 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 489 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 490 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 491 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 492 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 493 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 494 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 495 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 496 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 497 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 498 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 499 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 500 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 501 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 502 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 503 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 504 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 505 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 506 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 507 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 508 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 509 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 510 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 511 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 512 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 513 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 514 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 515 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 516 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 517 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 518 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 519 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 520 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 521 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 522 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 523 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 524 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 525 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 526 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 527 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 528 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 529 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 530 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 531 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 532 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 533 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 534 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 535 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 536 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 537 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 538 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 539 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 540 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 541 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 542 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 543 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 544 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 545 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 546 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 547 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 548 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 549 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 550 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 551 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 552 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 553 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 554 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 555 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 556 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 557 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 558 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 559 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 560 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 561 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 562 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 563 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 564 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 565 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 566 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 567 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 568 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 569 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 570 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 571 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 572 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 573 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 574 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 575 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 576 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 577 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 578 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 579 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 580 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 581 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 582 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 583 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 584 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 585 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 586 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 587 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 588 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 589 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 590 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 591 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 592 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 593 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 594 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 595 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 596 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 597 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 598 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 599 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 600 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 601 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 602 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 603 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 604 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 605 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 606 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 607 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 608 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 609 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 610 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 611 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 612 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 613 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 614 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 615 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 616 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 617 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 618 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 619 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 620 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 621 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 622 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 623 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 624 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 625 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 626 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 627 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 628 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 629 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 630 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 631 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 632 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 633 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 634 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 635 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 636 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 637 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 638 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 639 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 640 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 641 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 642 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 643 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 644 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 645 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 646 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 647 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 648 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 649 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 650 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 651 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 652 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 653 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 654 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 655 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 656 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 657 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 658 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 659 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 660 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 661 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 662 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 663 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 664 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 665 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 666 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 667 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 668 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 669 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 670 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 671 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 672 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 673 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 674 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 675 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 676 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 677 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 678 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 679 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 680 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 681 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 682 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 683 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 684 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 685 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 686 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 687 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 688 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 689 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 690 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 691 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 692 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 693 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 694 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 695 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 696 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 697 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 698 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 699 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 700 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 701 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 702 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 703 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 704 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 705 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 706 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 707 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 708 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 709 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 710 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 711 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 712 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 713 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 714 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 715 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 716 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 717 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 718 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 719 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 720 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 721 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 722 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 723 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 724 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 725 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 726 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 727 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 728 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 729 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 730 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 731 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 732 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 733 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 734 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 735 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 736 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 737 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 738 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 739 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 740 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 741 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 742 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 743 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 744 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 745 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 746 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 747 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 748 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 749 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 750 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 751 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 752 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 753 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 754 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 755 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 756 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 757 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 758 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 759 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 760 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 761 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 762 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 763 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 764 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 765 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 766 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 767 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 768 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 769 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 770 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 771 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 772 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 773 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 774 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 775 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 776 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 777 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 778 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 779 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 780 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 781 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 782 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 783 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 784 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 785 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 786 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 787 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 788 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 789 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 790 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 791 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 792 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 793 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 794 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 795 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 796 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 797 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 798 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 799 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 800 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 801 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 802 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 803 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 804 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 805 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 806 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 807 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 808 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 809 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 810 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 811 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 812 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 813 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 814 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 815 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 816 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 817 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 818 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 819 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 820 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 821 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 822 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 823 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 824 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 825 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 826 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 827 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 828 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 829 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 830 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 831 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 832 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 833 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 834 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 835 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 836 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 837 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 838 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 839 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 840 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 841 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 842 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 843 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 844 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 845 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 846 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 847 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 848 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 849 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 850 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 851 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 852 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 853 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 854 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 855 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 856 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 857 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 858 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 859 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 860 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 861 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 862 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 863 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 864 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 865 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 866 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 867 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 868 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 869 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 870 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 871 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 872 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 873 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 874 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 875 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 876 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 877 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 878 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 879 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 880 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 881 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 882 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 883 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 884 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 885 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 886 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 887 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 888 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 889 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 890 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 891 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 892 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 893 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 894 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 895 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 896 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 897 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 898 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 899 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 900 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 901 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 902 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 903 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 904 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 905 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 906 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 907 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 908 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 909 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 910 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 911 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 912 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 913 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 914 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 915 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 916 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 917 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 918 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 919 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 920 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 921 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 922 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 923 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 924 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 925 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 926 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 927 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 928 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 929 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 930 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 931 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 932 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 933 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 934 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 935 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 936 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 937 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 938 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 939 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 940 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 941 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 942 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 943 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 944 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 945 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 946 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 947 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 948 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 949 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 950 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 951 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 952 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 953 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 954 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 955 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 956 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 957 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 958 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 959 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 960 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 961 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 962 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 963 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 964 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 965 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 966 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 967 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 968 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 969 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 970 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 971 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 972 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 973 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 974 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 975 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 976 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 977 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 978 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 979 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 980 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 981 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 982 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 983 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 984 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 985 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 986 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 987 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 988 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 989 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 990 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 991 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 992 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 993 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 994 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 995 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 996 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 997 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 998 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 999 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 1000 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 1001 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 1002 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 1003 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 1004 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 1005 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 1006 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 1007 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 1008 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 1009 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 1010 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 1011 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 1012 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 1013 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 1014 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 1015 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 1016 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 1017 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 1018 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 1019 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 1020 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 1021 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 1022 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.219.0> 2011-05-07 17:07:00 =============================================================================== ns_1@127.0.0.1:ns_janitor:152: Setting vbucket 1023 in "default" on 'ns_1@127.0.0.1' from missing to active. INFO REPORT <6516.64.0> 2011-05-07 17:07:02 =============================================================================== unsupervising port: {moxi,"./bin/moxi/moxi", ["-Z", "port_listen=11211,default_bucket_name=default,downstream_max=1024,downstream_conn_max=4,connect_max_errors=5,connect_retry_interval=30000,connect_timeout=400,auth_timeout=100,cycle=200,downstream_conn_queue_timeout=200,downstream_timeout=5000,wait_queue_timeout=200", "-z", "url=http://127.0.0.1:8091/pools/default/saslBucketsStreaming", "-p","0","-Y","y","-O","stderr",[]], [{env,[{"EVENT_NOSELECT","1"}, {"MOXI_SASL_PLAIN_USR",[]}, {"MOXI_SASL_PLAIN_PWD",[]}]}, use_stdio,stderr_to_stdout,stream]} INFO REPORT <6516.64.0> 2011-05-07 17:07:02 =============================================================================== supervising port: {moxi,"./bin/moxi/moxi", ["-Z", "port_listen=11211,default_bucket_name=default,downstream_max=1024,downstream_conn_max=4,connect_max_errors=5,connect_retry_interval=30000,connect_timeout=400,auth_timeout=100,cycle=200,downstream_conn_queue_timeout=200,downstream_timeout=5000,wait_queue_timeout=200", "-z", "url=http://127.0.0.1:8091/pools/default/saslBucketsStreaming", "-p","0","-Y","y","-O","stderr",[]], [{env,[{"EVENT_NOSELECT","1"}, {"MOXI_SASL_PLAIN_USR","Administrator"}, {"MOXI_SASL_PLAIN_PWD","111111"}]}, use_stdio,stderr_to_stdout,stream]} INFO REPORT <6516.273.0> 2011-05-07 17:07:02 =============================================================================== starting ns_port_server with delay of 5000 PROGRESS REPORT <6516.99.0> 2011-05-07 17:07:02 =============================================================================== supervisor {local,ns_port_sup} started [{pid,<6516.273.0>}, {name, {moxi,"./bin/moxi/moxi", ["-Z", "port_listen=11211,default_bucket_name=default,downstream_max=1024,downstream_conn_max=4,connect_max_errors=5,connect_retry_interval=30000,connect_timeout=400,auth_timeout=100,cycle=200,downstream_conn_queue_timeout=200,downstream_timeout=5000,wait_queue_timeout=200", "-z", "url=http://127.0.0.1:8091/pools/default/saslBucketsStreaming", "-p","0","-Y","y","-O","stderr",[]], [{env, [{"EVENT_NOSELECT","1"}, {"MOXI_SASL_PLAIN_USR","Administrator"}, {"MOXI_SASL_PLAIN_PWD","111111"}]}, use_stdio,stderr_to_stdout,stream]}}, {mfa, {supervisor_cushion,start_link, [moxi,5000,ns_port_server,start_link, [moxi,"./bin/moxi/moxi", ["-Z", "port_listen=11211,default_bucket_name=default,downstream_max=1024,downstream_conn_max=4,connect_max_errors=5,connect_retry_interval=30000,connect_timeout=400,auth_timeout=100,cycle=200,downstream_conn_queue_timeout=200,downstream_timeout=5000,wait_queue_timeout=200", "-z", "url=http://127.0.0.1:8091/pools/default/saslBucketsStreaming", "-p","0","-Y","y","-O","stderr",[]], [{env, [{"EVENT_NOSELECT","1"}, {"MOXI_SASL_PLAIN_USR","Administrator"}, {"MOXI_SASL_PLAIN_PWD","111111"}]}, use_stdio,stderr_to_stdout,stream]]]}}, {restart_type,permanent}, {shutdown,10}, {child_type,worker}] INFO REPORT <6516.96.0> 2011-05-07 17:07:02 =============================================================================== menelaus_web streaming socket closed by client INFO REPORT <6516.64.0> 2011-05-07 17:07:02 =============================================================================== config change: rest -> [{port,8091}] INFO REPORT <6516.64.0> 2011-05-07 17:07:02 =============================================================================== ns_node_disco_conf_events config all INFO REPORT <6516.85.0> 2011-05-07 17:07:02 =============================================================================== Pushing config INFO REPORT <6516.85.0> 2011-05-07 17:07:02 =============================================================================== Pushing config done INFO REPORT <6516.64.0> 2011-05-07 17:07:02 =============================================================================== config change: rest_creds -> ******** INFO REPORT <6516.64.0> 2011-05-07 17:07:02 =============================================================================== ns_node_disco_conf_events config all INFO REPORT <6516.85.0> 2011-05-07 17:07:02 =============================================================================== Pushing config INFO REPORT <6516.85.0> 2011-05-07 17:07:02 =============================================================================== Pushing config done INFO REPORT <6516.92.0> 2011-05-07 17:07:03 =============================================================================== ns_1@127.0.0.1:ns_doctor:82: Current node statuses: [{'ns_1@127.0.0.1', [{last_heard,{1304,813223,60222}}, {active_buckets,["default"]}, {memory, [{total,13453944}, {processes,3531860}, {processes_used,3528404}, {system,9922084}, {atom,516697}, {atom_used,493918}, {binary,216528}, {code,3998738}, {ets,305588}]}, {cluster_compatibility_version,1}, {version, [{kernel,"2.13.4"}, {mnesia,"4.4.12"}, {stdlib,"1.16.4"}, {os_mon,"2.2.4"}, {ns_server,"1.6.5.4r"}, {menelaus,"1.6.5.4r"}, {sasl,"2.1.8"}]}, {system_arch,"i686-pc-linux-gnu"}, {wall_clock,120}, {memory_data,{1050726400,689426432,{<6516.70.0>,486376}}}, {disk_data, [{"/",7583436,17}, {"/dev",508460,1}, {"/dev/shm",513048,0}, {"/var/run",513048,1}, {"/var/lock",513048,0}, {"/lib/init/rw",513048,0}, {"/boot",233191,8}]}, {meminfo, <<"MemTotal: 1026100 kB\nMemFree: 402940 kB\nBuffers: 152648 kB\nCached: 413800 kB\nSwapCached: 0 kB\nActive: 329256 kB\nInactive: 257068 kB\nActive(anon): 15692 kB\nInactive(anon): 4412 kB\nActive(file): 313564 kB\nInactive(file): 252656 kB\nUnevictable: 0 kB\nMlocked: 0 kB\nHighTotal: 137160 kB\nHighFree: 7896 kB\nLowTotal: 888940 kB\nLowFree: 395044 kB\nSwapTotal: 397304 kB\nSwapFree: 397304 kB\nDirty: 0 kB\nWriteback: 0 kB\nAnonPages: 19876 kB\nMapped: 8600 kB\nShmem: 228 kB\nSlab: 27280 kB\nSReclaimable: 22536 kB\nSUnreclaim: 4744 kB\nKernelStack: 872 kB\nPageTables: 656 kB\nNFS_Unstable: 0 kB\nBounce: 0 kB\nWritebackTmp: 0 kB\nCommitLimit: 910352 kB\nCommitted_AS: 180992 kB\nVmallocTotal: 122880 kB\nVmallocUsed: 3540 kB\nVmallocChunk: 114424 kB\nHardwareCorrupted: 0 kB\nHugePages_Total: 0\nHugePages_Free: 0\nHugePages_Rsvd: 0\nHugePages_Surp: 0\nHugepagesize: 2048 kB\nDirectMap4k: 10232 kB\nDirectMap2M: 901120 kB\n">>}, {replication,[]}, {system_memory_data, [{system_total_memory,1050726400}, {free_swap,406839296}, {total_swap,406839296}, {cached_memory,423731200}, {buffered_memory,156311552}, {free_memory,412610560}, {total_memory,1050726400}]}, {statistics, [{wall_clock,{110410,1}}, {context_switches,{15120,0}}, {garbage_collection,{3022,5696983,0}}, {io,{{input,4826892},{output,1144670}}}, {reductions,{2299597,16467}}, {run_queue,0}, {runtime,{550,0}}]}]}] INFO REPORT <6516.274.0> 2011-05-07 17:07:03 =============================================================================== moxi<0.274.0>: 2011-05-07 17:07:02: (cproxy_config.c.316) env: MOXI_SASL_PLAIN_USR (13) moxi<0.274.0>: 2011-05-07 17:07:02: (cproxy_config.c.325) env: MOXI_SASL_PLAIN_PWD (6) INFO REPORT <6516.92.0> 2011-05-07 17:08:03 =============================================================================== ns_1@127.0.0.1:ns_doctor:82: Current node statuses: [{'ns_1@127.0.0.1', [{last_heard,{1304,813283,60289}}, {active_buckets,["default"]}, {memory, [{total,13908800}, {processes,3999588}, {processes_used,3995668}, {system,9909212}, {atom,518633}, {atom_used,496964}, {binary,98304}, {code,4014832}, {ets,393828}]}, {cluster_compatibility_version,1}, {version, [{kernel,"2.13.4"}, {mnesia,"4.4.12"}, {stdlib,"1.16.4"}, {os_mon,"2.2.4"}, {ns_server,"1.6.5.4r"}, {menelaus,"1.6.5.4r"}, {sasl,"2.1.8"}]}, {system_arch,"i686-pc-linux-gnu"}, {wall_clock,180}, {memory_data,{1050726400,703123456,{<6516.70.0>,486376}}}, {disk_data, [{"/",7583436,17}, {"/dev",508460,1}, {"/dev/shm",513048,0}, {"/var/run",513048,1}, {"/var/lock",513048,0}, {"/lib/init/rw",513048,0}, {"/boot",233191,8}]}, {meminfo, <<"MemTotal: 1026100 kB\nMemFree: 339704 kB\nBuffers: 152672 kB\nCached: 422736 kB\nSwapCached: 0 kB\nActive: 389480 kB\nInactive: 260228 kB\nActive(anon): 70116 kB\nInactive(anon): 4412 kB\nActive(file): 319364 kB\nInactive(file): 255816 kB\nUnevictable: 0 kB\nMlocked: 0 kB\nHighTotal: 137160 kB\nHighFree: 252 kB\nLowTotal: 888940 kB\nLowFree: 339452 kB\nSwapTotal: 397304 kB\nSwapFree: 397304 kB\nDirty: 240 kB\nWriteback: 0 kB\nAnonPages: 74300 kB\nMapped: 10068 kB\nShmem: 228 kB\nSlab: 27440 kB\nSReclaimable: 22704 kB\nSUnreclaim: 4736 kB\nKernelStack: 920 kB\nPageTables: 824 kB\nNFS_Unstable: 0 kB\nBounce: 0 kB\nWritebackTmp: 0 kB\nCommitLimit: 910352 kB\nCommitted_AS: 269984 kB\nVmallocTotal: 122880 kB\nVmallocUsed: 3540 kB\nVmallocChunk: 114424 kB\nHardwareCorrupted: 0 kB\nHugePages_Total: 0\nHugePages_Free: 0\nHugePages_Rsvd: 0\nHugePages_Surp: 0\nHugepagesize: 2048 kB\nDirectMap4k: 10232 kB\nDirectMap2M: 901120 kB\n">>}, {replication,[{"default",1.0}]}, {system_memory_data, [{system_total_memory,1050726400}, {free_swap,406839296}, {total_swap,406839296}, {cached_memory,432881664}, {buffered_memory,156336128}, {free_memory,347856896}, {total_memory,1050726400}]}, {statistics, [{wall_clock,{170410,1}}, {context_switches,{32767,0}}, {garbage_collection,{6825,22142611,0}}, {io,{{input,5214796},{output,1967744}}}, {reductions,{10997778,1549378}}, {run_queue,0}, {runtime,{1580,170}}]}]}] INFO REPORT <6516.224.0> 2011-05-07 17:08:39 =============================================================================== ns_1@127.0.0.1:stats_collector:71: Stats for bucket "default": auth_cmds 2 auth_errors 0 bucket_conns 3 bytes_read 14354893 bytes_written 3493142 cas_badval 0 cas_hits 0 cas_misses 0 cmd_flush 0 cmd_get 0 cmd_set 98097 conn_yields 0 connection_structures 13 curr_connections 13 curr_items 93697 curr_items_tot 93697 daemon_connections 10 decr_hits 0 decr_misses 0 delete_hits 0 delete_misses 0 ep_bg_fetched 0 ep_commit_num 204 ep_commit_time 0 ep_commit_time_total 3 ep_data_age 1 ep_data_age_highwat 5 ep_db_cleaner_status complete ep_db_strategy multiMTVBDB ep_dbinit 1 ep_dbname /var/opt/membase/1.6.5.4r/data/ns_1/default ep_dbshards 4 ep_expired 0 ep_flush_duration 0 ep_flush_duration_highwat 1 ep_flush_duration_total 3 ep_flush_preempts 0 ep_flusher_state running ep_flusher_todo 0 ep_io_num_read 0 ep_io_num_write 95714 ep_io_read_bytes 0 ep_io_write_bytes 10911396 ep_item_begin_failed 0 ep_item_commit_failed 0 ep_item_flush_expired 0 ep_item_flush_failed 0 ep_kv_size 16396975 ep_max_data_size 839909376 ep_max_txn_size 1000 ep_mem_high_wat 629932032 ep_mem_low_wat 503945625 ep_min_data_age 0 ep_num_active_non_resident 0 ep_num_eject_failures 0 ep_num_eject_replicas 0 ep_num_expiry_pager_runs 0 ep_num_non_resident 0 ep_num_not_my_vbuckets 0 ep_num_pager_runs 0 ep_num_value_ejects 0 ep_oom_errors 0 ep_overhead 13050586 ep_pending_ops 0 ep_pending_ops_max 0 ep_pending_ops_max_duration 0 ep_pending_ops_total 0 ep_queue_age_cap 900 ep_queue_size 2383 ep_storage_age 0 ep_storage_age_highwat 2 ep_storage_type featured ep_store_max_concurrency 10 ep_store_max_readers 9 ep_store_max_readwrite 1 ep_tap_bg_fetch_requeued 0 ep_tap_bg_fetched 0 ep_tap_keepalive 300 ep_tmp_oom_errors 0 ep_too_old 0 ep_too_young 0 ep_total_cache_size 16396975 ep_total_del_items 0 ep_total_enqueued 98187 ep_total_new_items 91434 ep_total_persisted 95714 ep_vbucket_del 0 ep_vbucket_del_fail 0 ep_version 1.6.5.3.1_8_gce31baa ep_warmed_up 0 ep_warmup true ep_warmup_dups 0 ep_warmup_oom 0 ep_warmup_thread complete ep_warmup_time 29657 get_hits 0 get_misses 0 incr_hits 0 incr_misses 0 libevent 2.0.7-rc limit_maxbytes 67108864 mem_used 29447561 pid 5769 pointer_size 32 rejected_conns 0 rusage_system 1.196074 rusage_user 2.036127 threads 4 time 1304813318 total_connections 13 uptime 217 version 1.4.4_382_g9df3289 INFO REPORT <6516.92.0> 2011-05-07 17:09:03 =============================================================================== ns_1@127.0.0.1:ns_doctor:82: Current node statuses: [{'ns_1@127.0.0.1', [{last_heard,{1304,813343,60258}}, {active_buckets,["default"]}, {memory, [{total,13922720}, {processes,3967892}, {processes_used,3963012}, {system,9954828}, {atom,518633}, {atom_used,496964}, {binary,104320}, {code,4014832}, {ets,432764}]}, {cluster_compatibility_version,1}, {version, [{kernel,"2.13.4"}, {mnesia,"4.4.12"}, {stdlib,"1.16.4"}, {os_mon,"2.2.4"}, {ns_server,"1.6.5.4r"}, {menelaus,"1.6.5.4r"}, {sasl,"2.1.8"}]}, {system_arch,"i686-pc-linux-gnu"}, {wall_clock,240}, {memory_data,{1050726400,764452864,{<6516.70.0>,486376}}}, {disk_data, [{"/",7583436,18}, {"/dev",508460,1}, {"/dev/shm",513048,0}, {"/var/run",513048,1}, {"/var/lock",513048,0}, {"/lib/init/rw",513048,0}, {"/boot",233191,8}]}, {meminfo, <<"MemTotal: 1026100 kB\nMemFree: 292460 kB\nBuffers: 152708 kB\nCached: 438016 kB\nSwapCached: 0 kB\nActive: 424224 kB\nInactive: 272268 kB\nActive(anon): 101584 kB\nInactive(anon): 4412 kB\nActive(file): 322640 kB\nInactive(file): 267856 kB\nUnevictable: 0 kB\nMlocked: 0 kB\nHighTotal: 137160 kB\nHighFree: 252 kB\nLowTotal: 888940 kB\nLowFree: 292208 kB\nSwapTotal: 397304 kB\nSwapFree: 397304 kB\nDirty: 260 kB\nWriteback: 0 kB\nAnonPages: 105820 kB\nMapped: 10072 kB\nShmem: 228 kB\nSlab: 27700 kB\nSReclaimable: 22968 kB\nSUnreclaim: 4732 kB\nKernelStack: 920 kB\nPageTables: 884 kB\nNFS_Unstable: 0 kB\nBounce: 0 kB\nWritebackTmp: 0 kB\nCommitLimit: 910352 kB\nCommitted_AS: 300832 kB\nVmallocTotal: 122880 kB\nVmallocUsed: 3540 kB\nVmallocChunk: 114424 kB\nHardwareCorrupted: 0 kB\nHugePages_Total: 0\nHugePages_Free: 0\nHugePages_Rsvd: 0\nHugePages_Surp: 0\nHugepagesize: 2048 kB\nDirectMap4k: 10232 kB\nDirectMap2M: 901120 kB\n">>}, {replication,[{"default",1.0}]}, {system_memory_data, [{system_total_memory,1050726400}, {free_swap,406839296}, {total_swap,406839296}, {cached_memory,448528384}, {buffered_memory,156372992}, {free_memory,299479040}, {total_memory,1050726400}]}, {statistics, [{wall_clock,{230409,0}}, {context_switches,{44661,0}}, {garbage_collection,{9601,40881350,0}}, {io,{{input,5486857},{output,2582253}}}, {reductions,{21902013,1848062}}, {run_queue,0}, {runtime,{2740,190}}]}]}] INFO REPORT <6516.92.0> 2011-05-07 17:10:03 =============================================================================== ns_1@127.0.0.1:ns_doctor:82: Current node statuses: [{'ns_1@127.0.0.1', [{last_heard,{1304,813403,60263}}, {active_buckets,["default"]}, {memory, [{total,13951712}, {processes,3971844}, {processes_used,3966964}, {system,9979868}, {atom,518633}, {atom_used,496964}, {binary,96184}, {code,4014832}, {ets,465748}]}, {cluster_compatibility_version,1}, {version, [{kernel,"2.13.4"}, {mnesia,"4.4.12"}, {stdlib,"1.16.4"}, {os_mon,"2.2.4"}, {ns_server,"1.6.5.4r"}, {menelaus,"1.6.5.4r"}, {sasl,"2.1.8"}]}, {system_arch,"i686-pc-linux-gnu"}, {wall_clock,300}, {memory_data,{1050726400,777142272,{<6516.70.0>,486376}}}, {disk_data, [{"/",7583436,18}, {"/dev",508460,1}, {"/dev/shm",513048,0}, {"/var/run",513048,1}, {"/var/lock",513048,0}, {"/lib/init/rw",513048,0}, {"/boot",233191,8}]}, {meminfo, <<"MemTotal: 1026100 kB\nMemFree: 267288 kB\nBuffers: 152724 kB\nCached: 448228 kB\nSwapCached: 0 kB\nActive: 441040 kB\nInactive: 280348 kB\nActive(anon): 116252 kB\nInactive(anon): 4412 kB\nActive(file): 324788 kB\nInactive(file): 275936 kB\nUnevictable: 0 kB\nMlocked: 0 kB\nHighTotal: 137160 kB\nHighFree: 252 kB\nLowTotal: 888940 kB\nLowFree: 267036 kB\nSwapTotal: 397304 kB\nSwapFree: 397304 kB\nDirty: 88 kB\nWriteback: 0 kB\nAnonPages: 120436 kB\nMapped: 10080 kB\nShmem: 228 kB\nSlab: 27860 kB\nSReclaimable: 23128 kB\nSUnreclaim: 4732 kB\nKernelStack: 920 kB\nPageTables: 912 kB\nNFS_Unstable: 0 kB\nBounce: 0 kB\nWritebackTmp: 0 kB\nCommitLimit: 910352 kB\nCommitted_AS: 316320 kB\nVmallocTotal: 122880 kB\nVmallocUsed: 3540 kB\nVmallocChunk: 114424 kB\nHardwareCorrupted: 0 kB\nHugePages_Total: 0\nHugePages_Free: 0\nHugePages_Rsvd: 0\nHugePages_Surp: 0\nHugepagesize: 2048 kB\nDirectMap4k: 10232 kB\nDirectMap2M: 901120 kB\n">>}, {replication,[{"default",1.0}]}, {system_memory_data, [{system_total_memory,1050726400}, {free_swap,406839296}, {total_swap,406839296}, {cached_memory,458985472}, {buffered_memory,156389376}, {free_memory,273702912}, {total_memory,1050726400}]}, {statistics, [{wall_clock,{290410,1}}, {context_switches,{56293,0}}, {garbage_collection,{12206,59551648,0}}, {io,{{input,5499013},{output,2955635}}}, {reductions,{32765621,1849458}}, {run_queue,0}, {runtime,{3910,210}}]}]}] INFO REPORT <6516.224.0> 2011-05-07 17:10:19 =============================================================================== ns_1@127.0.0.1:stats_collector:71: Stats for bucket "default": auth_cmds 3 auth_errors 0 bucket_conns 4 bytes_read 37726923 bytes_written 31148317 cas_badval 0 cas_hits 0 cas_misses 0 cmd_flush 0 cmd_get 165643 cmd_set 215040 conn_yields 0 connection_structures 14 curr_connections 14 curr_items 204800 curr_items_tot 204800 daemon_connections 10 decr_hits 0 decr_misses 0 delete_hits 0 delete_misses 0 ep_bg_fetched 0 ep_commit_num 457 ep_commit_time 0 ep_commit_time_total 5 ep_data_age 1 ep_data_age_highwat 5 ep_db_cleaner_status complete ep_db_strategy multiMTVBDB ep_dbinit 1 ep_dbname /var/opt/membase/1.6.5.4r/data/ns_1/default ep_dbshards 4 ep_expired 0 ep_flush_duration 0 ep_flush_duration_highwat 1 ep_flush_duration_total 6 ep_flush_preempts 0 ep_flusher_state running ep_flusher_todo 0 ep_io_num_read 0 ep_io_num_write 215040 ep_io_read_bytes 0 ep_io_write_bytes 24514560 ep_item_begin_failed 0 ep_item_commit_failed 0 ep_item_flush_expired 0 ep_item_flush_failed 0 ep_kv_size 35840000 ep_max_data_size 839909376 ep_max_txn_size 1000 ep_mem_high_wat 629932032 ep_mem_low_wat 503945625 ep_min_data_age 0 ep_num_active_non_resident 0 ep_num_eject_failures 0 ep_num_eject_replicas 0 ep_num_expiry_pager_runs 0 ep_num_non_resident 0 ep_num_not_my_vbuckets 0 ep_num_pager_runs 0 ep_num_value_ejects 0 ep_oom_errors 0 ep_overhead 12960032 ep_pending_ops 0 ep_pending_ops_max 0 ep_pending_ops_max_duration 0 ep_pending_ops_total 0 ep_queue_age_cap 900 ep_queue_size 0 ep_storage_age 0 ep_storage_age_highwat 2 ep_storage_type featured ep_store_max_concurrency 10 ep_store_max_readers 9 ep_store_max_readwrite 1 ep_tap_bg_fetch_requeued 0 ep_tap_bg_fetched 0 ep_tap_keepalive 300 ep_tmp_oom_errors 0 ep_too_old 0 ep_too_young 0 ep_total_cache_size 35840000 ep_total_del_items 0 ep_total_enqueued 215239 ep_total_new_items 204800 ep_total_persisted 215040 ep_vbucket_del 0 ep_vbucket_del_fail 0 ep_version 1.6.5.3.1_8_gce31baa ep_warmed_up 0 ep_warmup true ep_warmup_dups 0 ep_warmup_oom 0 ep_warmup_thread complete ep_warmup_time 29657 get_hits 165643 get_misses 0 incr_hits 0 incr_misses 0 libevent 2.0.7-rc limit_maxbytes 67108864 mem_used 48800032 pid 5769 pointer_size 32 rejected_conns 0 rusage_system 2.384149 rusage_user 3.948246 threads 4 time 1304813418 total_connections 14 uptime 317 version 1.4.4_382_g9df3289 INFO REPORT <6516.92.0> 2011-05-07 17:11:03 =============================================================================== ns_1@127.0.0.1:ns_doctor:82: Current node statuses: [{'ns_1@127.0.0.1', [{last_heard,{1304,813463,60274}}, {active_buckets,["default"]}, {memory, [{total,13963520}, {processes,3940836}, {processes_used,3935956}, {system,10022684}, {atom,518633}, {atom_used,496964}, {binary,104240}, {code,4014832}, {ets,500300}]}, {cluster_compatibility_version,1}, {version, [{kernel,"2.13.4"}, {mnesia,"4.4.12"}, {stdlib,"1.16.4"}, {os_mon,"2.2.4"}, {ns_server,"1.6.5.4r"}, {menelaus,"1.6.5.4r"}, {sasl,"2.1.8"}]}, {system_arch,"i686-pc-linux-gnu"}, {wall_clock,360}, {memory_data,{1050726400,778182656,{<6516.70.0>,486376}}}, {disk_data, [{"/",7583436,18}, {"/dev",508460,1}, {"/dev/shm",513048,0}, {"/var/run",513048,1}, {"/var/lock",513048,0}, {"/lib/init/rw",513048,0}, {"/boot",233191,8}]}, {meminfo, <<"MemTotal: 1026100 kB\nMemFree: 266280 kB\nBuffers: 152728 kB\nCached: 448504 kB\nSwapCached: 0 kB\nActive: 442368 kB\nInactive: 279588 kB\nActive(anon): 116540 kB\nInactive(anon): 4412 kB\nActive(file): 325828 kB\nInactive(file): 275176 kB\nUnevictable: 0 kB\nMlocked: 0 kB\nHighTotal: 137160 kB\nHighFree: 252 kB\nLowTotal: 888940 kB\nLowFree: 266028 kB\nSwapTotal: 397304 kB\nSwapFree: 397304 kB\nDirty: 136 kB\nWriteback: 0 kB\nAnonPages: 120724 kB\nMapped: 10180 kB\nShmem: 228 kB\nSlab: 27896 kB\nSReclaimable: 23140 kB\nSUnreclaim: 4756 kB\nKernelStack: 920 kB\nPageTables: 936 kB\nNFS_Unstable: 0 kB\nBounce: 0 kB\nWritebackTmp: 0 kB\nCommitLimit: 910352 kB\nCommitted_AS: 316616 kB\nVmallocTotal: 122880 kB\nVmallocUsed: 3540 kB\nVmallocChunk: 114424 kB\nHardwareCorrupted: 0 kB\nHugePages_Total: 0\nHugePages_Free: 0\nHugePages_Rsvd: 0\nHugePages_Surp: 0\nHugepagesize: 2048 kB\nDirectMap4k: 10232 kB\nDirectMap2M: 901120 kB\n">>}, {replication,[{"default",1.0}]}, {system_memory_data, [{system_total_memory,1050726400}, {free_swap,406839296}, {total_swap,406839296}, {cached_memory,459268096}, {buffered_memory,156393472}, {free_memory,272670720}, {total_memory,1050726400}]}, {statistics, [{wall_clock,{350410,1}}, {context_switches,{67916,0}}, {garbage_collection,{14837,78258301,0}}, {io,{{input,5511178},{output,3332679}}}, {reductions,{43638345,1857284}}, {run_queue,0}, {runtime,{5040,210}}]}]}] INFO REPORT <6516.224.0> 2011-05-07 17:11:59 =============================================================================== ns_1@127.0.0.1:stats_collector:71: Stats for bucket "default": auth_cmds 4 auth_errors 0 bucket_conns 5 bytes_read 39219743 bytes_written 38224785 cas_badval 0 cas_hits 0 cas_misses 0 cmd_flush 0 cmd_get 204800 cmd_set 215040 conn_yields 0 connection_structures 15 curr_connections 15 curr_items 204800 curr_items_tot 204800 daemon_connections 10 decr_hits 0 decr_misses 0 delete_hits 0 delete_misses 0 ep_bg_fetched 0 ep_commit_num 457 ep_commit_time 0 ep_commit_time_total 5 ep_data_age 1 ep_data_age_highwat 5 ep_db_cleaner_status complete ep_db_strategy multiMTVBDB ep_dbinit 1 ep_dbname /var/opt/membase/1.6.5.4r/data/ns_1/default ep_dbshards 4 ep_expired 0 ep_flush_duration 0 ep_flush_duration_highwat 1 ep_flush_duration_total 6 ep_flush_preempts 0 ep_flusher_state running ep_flusher_todo 0 ep_io_num_read 0 ep_io_num_write 215040 ep_io_read_bytes 0 ep_io_write_bytes 24514560 ep_item_begin_failed 0 ep_item_commit_failed 0 ep_item_flush_expired 0 ep_item_flush_failed 0 ep_kv_size 35840000 ep_max_data_size 839909376 ep_max_txn_size 1000 ep_mem_high_wat 629932032 ep_mem_low_wat 503945625 ep_min_data_age 0 ep_num_active_non_resident 0 ep_num_eject_failures 0 ep_num_eject_replicas 0 ep_num_expiry_pager_runs 0 ep_num_non_resident 0 ep_num_not_my_vbuckets 0 ep_num_pager_runs 0 ep_num_value_ejects 0 ep_oom_errors 0 ep_overhead 12960032 ep_pending_ops 0 ep_pending_ops_max 0 ep_pending_ops_max_duration 0 ep_pending_ops_total 0 ep_queue_age_cap 900 ep_queue_size 0 ep_storage_age 0 ep_storage_age_highwat 2 ep_storage_type featured ep_store_max_concurrency 10 ep_store_max_readers 9 ep_store_max_readwrite 1 ep_tap_bg_fetch_requeued 0 ep_tap_bg_fetched 0 ep_tap_keepalive 300 ep_tmp_oom_errors 0 ep_too_old 0 ep_too_young 0 ep_total_cache_size 35840000 ep_total_del_items 0 ep_total_enqueued 215239 ep_total_new_items 204800 ep_total_persisted 215040 ep_vbucket_del 0 ep_vbucket_del_fail 0 ep_version 1.6.5.3.1_8_gce31baa ep_warmed_up 0 ep_warmup true ep_warmup_dups 0 ep_warmup_oom 0 ep_warmup_thread complete ep_warmup_time 29657 get_hits 204800 get_misses 0 incr_hits 0 incr_misses 0 libevent 2.0.7-rc limit_maxbytes 67108864 mem_used 48800032 pid 5769 pointer_size 32 rejected_conns 0 rusage_system 2.432152 rusage_user 4.048253 threads 4 time 1304813519 total_connections 15 uptime 418 version 1.4.4_382_g9df3289 INFO REPORT <6516.92.0> 2011-05-07 17:12:03 =============================================================================== ns_1@127.0.0.1:ns_doctor:82: Current node statuses: [{'ns_1@127.0.0.1', [{last_heard,{1304,813523,60273}}, {active_buckets,["default"]}, {memory, [{total,14011736}, {processes,3952164}, {processes_used,3947236}, {system,10059572}, {atom,518633}, {atom_used,496964}, {binary,105648}, {code,4014832}, {ets,535132}]}, {cluster_compatibility_version,1}, {version, [{kernel,"2.13.4"}, {mnesia,"4.4.12"}, {stdlib,"1.16.4"}, {os_mon,"2.2.4"}, {ns_server,"1.6.5.4r"}, {menelaus,"1.6.5.4r"}, {sasl,"2.1.8"}]}, {system_arch,"i686-pc-linux-gnu"}, {wall_clock,420}, {memory_data,{1050726400,779350016,{<6516.70.0>,486376}}}, {disk_data, [{"/",7583436,18}, {"/dev",508460,1}, {"/dev/shm",513048,0}, {"/var/run",513048,1}, {"/var/lock",513048,0}, {"/lib/init/rw",513048,0}, {"/boot",233191,8}]}, {meminfo, <<"MemTotal: 1026100 kB\nMemFree: 266668 kB\nBuffers: 152728 kB\nCached: 447980 kB\nSwapCached: 0 kB\nActive: 442276 kB\nInactive: 279196 kB\nActive(anon): 116580 kB\nInactive(anon): 4412 kB\nActive(file): 325696 kB\nInactive(file): 274784 kB\nUnevictable: 0 kB\nMlocked: 0 kB\nHighTotal: 137160 kB\nHighFree: 252 kB\nLowTotal: 888940 kB\nLowFree: 266416 kB\nSwapTotal: 397304 kB\nSwapFree: 397304 kB\nDirty: 80 kB\nWriteback: 0 kB\nAnonPages: 120764 kB\nMapped: 10180 kB\nShmem: 228 kB\nSlab: 27884 kB\nSReclaimable: 23140 kB\nSUnreclaim: 4744 kB\nKernelStack: 920 kB\nPageTables: 936 kB\nNFS_Unstable: 0 kB\nBounce: 0 kB\nWritebackTmp: 0 kB\nCommitLimit: 910352 kB\nCommitted_AS: 316616 kB\nVmallocTotal: 122880 kB\nVmallocUsed: 3540 kB\nVmallocChunk: 114424 kB\nHardwareCorrupted: 0 kB\nHugePages_Total: 0\nHugePages_Free: 0\nHugePages_Rsvd: 0\nHugePages_Surp: 0\nHugepagesize: 2048 kB\nDirectMap4k: 10232 kB\nDirectMap2M: 901120 kB\n">>}, {replication,[{"default",1.0}]}, {system_memory_data, [{system_total_memory,1050726400}, {free_swap,406839296}, {total_swap,406839296}, {cached_memory,458731520}, {buffered_memory,156393472}, {free_memory,273068032}, {total_memory,1050726400}]}, {statistics, [{wall_clock,{410409,0}}, {context_switches,{78792,0}}, {garbage_collection,{17237,91449311,0}}, {io,{{input,6278142},{output,3850945}}}, {reductions,{54044305,1711067}}, {run_queue,0}, {runtime,{6100,170}}]}]}] INFO REPORT <6516.92.0> 2011-05-07 17:13:03 =============================================================================== ns_1@127.0.0.1:ns_doctor:82: Current node statuses: [{'ns_1@127.0.0.1', [{last_heard,{1304,813583,60263}}, {active_buckets,["default"]}, {memory, [{total,13931160}, {processes,3810048}, {processes_used,3797760}, {system,10121112}, {atom,518633}, {atom_used,496964}, {binary,131864}, {code,4014832}, {ets,569740}]}, {cluster_compatibility_version,1}, {version, [{kernel,"2.13.4"}, {mnesia,"4.4.12"}, {stdlib,"1.16.4"}, {os_mon,"2.2.4"}, {ns_server,"1.6.5.4r"}, {menelaus,"1.6.5.4r"}, {sasl,"2.1.8"}]}, {system_arch,"i686-pc-linux-gnu"}, {wall_clock,480}, {memory_data,{1050726400,778731520,{<6516.70.0>,486376}}}, {disk_data, [{"/",7583436,18}, {"/dev",508460,1}, {"/dev/shm",513048,0}, {"/var/run",513048,1}, {"/var/lock",513048,0}, {"/lib/init/rw",513048,0}, {"/boot",233191,8}]}, {meminfo, <<"MemTotal: 1026100 kB\nMemFree: 265644 kB\nBuffers: 152732 kB\nCached: 448240 kB\nSwapCached: 0 kB\nActive: 443140 kB\nInactive: 279336 kB\nActive(anon): 117328 kB\nInactive(anon): 4412 kB\nActive(file): 325812 kB\nInactive(file): 274924 kB\nUnevictable: 0 kB\nMlocked: 0 kB\nHighTotal: 137160 kB\nHighFree: 252 kB\nLowTotal: 888940 kB\nLowFree: 265392 kB\nSwapTotal: 397304 kB\nSwapFree: 397304 kB\nDirty: 72 kB\nWriteback: 0 kB\nAnonPages: 121512 kB\nMapped: 10180 kB\nShmem: 228 kB\nSlab: 27916 kB\nSReclaimable: 23140 kB\nSUnreclaim: 4776 kB\nKernelStack: 920 kB\nPageTables: 936 kB\nNFS_Unstable: 0 kB\nBounce: 0 kB\nWritebackTmp: 0 kB\nCommitLimit: 910352 kB\nCommitted_AS: 316616 kB\nVmallocTotal: 122880 kB\nVmallocUsed: 3540 kB\nVmallocChunk: 114424 kB\nHardwareCorrupted: 0 kB\nHugePages_Total: 0\nHugePages_Free: 0\nHugePages_Rsvd: 0\nHugePages_Surp: 0\nHugepagesize: 2048 kB\nDirectMap4k: 10232 kB\nDirectMap2M: 901120 kB\n">>}, {replication,[{"default",1.0}]}, {system_memory_data, [{system_total_memory,1050726400}, {free_swap,406839296}, {total_swap,406839296}, {cached_memory,458997760}, {buffered_memory,156397568}, {free_memory,272019456}, {total_memory,1050726400}]}, {statistics, [{wall_clock,{470409,0}}, {context_switches,{91022,0}}, {garbage_collection,{19898,104781620,0}}, {io,{{input,7407151},{output,5291714}}}, {reductions,{64966153,1638761}}, {run_queue,0}, {runtime,{7190,170}}]}]}] INFO REPORT <6516.224.0> 2011-05-07 17:13:39 =============================================================================== ns_1@127.0.0.1:stats_collector:71: Stats for bucket "default": auth_cmds 4 auth_errors 0 bucket_conns 5 bytes_read 39224799 bytes_written 39792852 cas_badval 0 cas_hits 0 cas_misses 0 cmd_flush 0 cmd_get 204800 cmd_set 215040 conn_yields 0 connection_structures 15 curr_connections 15 curr_items 204800 curr_items_tot 204800 daemon_connections 10 decr_hits 0 decr_misses 0 delete_hits 0 delete_misses 0 ep_bg_fetched 0 ep_commit_num 457 ep_commit_time 0 ep_commit_time_total 5 ep_data_age 1 ep_data_age_highwat 5 ep_db_cleaner_status complete ep_db_strategy multiMTVBDB ep_dbinit 1 ep_dbname /var/opt/membase/1.6.5.4r/data/ns_1/default ep_dbshards 4 ep_expired 0 ep_flush_duration 0 ep_flush_duration_highwat 1 ep_flush_duration_total 6 ep_flush_preempts 0 ep_flusher_state running ep_flusher_todo 0 ep_io_num_read 0 ep_io_num_write 215040 ep_io_read_bytes 0 ep_io_write_bytes 24514560 ep_item_begin_failed 0 ep_item_commit_failed 0 ep_item_flush_expired 0 ep_item_flush_failed 0 ep_kv_size 35840000 ep_max_data_size 839909376 ep_max_txn_size 1000 ep_mem_high_wat 629932032 ep_mem_low_wat 503945625 ep_min_data_age 0 ep_num_active_non_resident 0 ep_num_eject_failures 0 ep_num_eject_replicas 0 ep_num_expiry_pager_runs 0 ep_num_non_resident 0 ep_num_not_my_vbuckets 0 ep_num_pager_runs 0 ep_num_value_ejects 0 ep_oom_errors 0 ep_overhead 12960032 ep_pending_ops 0 ep_pending_ops_max 0 ep_pending_ops_max_duration 0 ep_pending_ops_total 0 ep_queue_age_cap 900 ep_queue_size 0 ep_storage_age 0 ep_storage_age_highwat 2 ep_storage_type featured ep_store_max_concurrency 10 ep_store_max_readers 9 ep_store_max_readwrite 1 ep_tap_bg_fetch_requeued 0 ep_tap_bg_fetched 0 ep_tap_keepalive 300 ep_tmp_oom_errors 0 ep_too_old 0 ep_too_young 0 ep_total_cache_size 35840000 ep_total_del_items 0 ep_total_enqueued 215239 ep_total_new_items 204800 ep_total_persisted 215040 ep_vbucket_del 0 ep_vbucket_del_fail 0 ep_version 1.6.5.3.1_8_gce31baa ep_warmed_up 0 ep_warmup true ep_warmup_dups 0 ep_warmup_oom 0 ep_warmup_thread complete ep_warmup_time 29657 get_hits 204800 get_misses 0 incr_hits 0 incr_misses 0 libevent 2.0.7-rc limit_maxbytes 67108864 mem_used 48800032 pid 5769 pointer_size 32 rejected_conns 0 rusage_system 2.436152 rusage_user 4.072254 threads 4 time 1304813619 total_connections 15 uptime 518 version 1.4.4_382_g9df3289 INFO REPORT <6516.92.0> 2011-05-07 17:14:03 =============================================================================== ns_1@127.0.0.1:ns_doctor:82: Current node statuses: [{'ns_1@127.0.0.1', [{last_heard,{1304,813643,60202}}, {active_buckets,["default"]}, {memory, [{total,14913744}, {processes,4785208}, {processes_used,4773848}, {system,10128536}, {atom,518633}, {atom_used,496964}, {binary,104944}, {code,4014832}, {ets,602788}]}, {cluster_compatibility_version,1}, {version, [{kernel,"2.13.4"}, {mnesia,"4.4.12"}, {stdlib,"1.16.4"}, {os_mon,"2.2.4"}, {ns_server,"1.6.5.4r"}, {menelaus,"1.6.5.4r"}, {sasl,"2.1.8"}]}, {system_arch,"i686-pc-linux-gnu"}, {wall_clock,540}, {memory_data,{1050726400,780644352,{<6516.70.0>,486376}}}, {disk_data, [{"/",7583436,18}, {"/dev",508460,1}, {"/dev/shm",513048,0}, {"/var/run",513048,1}, {"/var/lock",513048,0}, {"/lib/init/rw",513048,0}, {"/boot",233191,8}]}, {meminfo, <<"MemTotal: 1026100 kB\nMemFree: 263876 kB\nBuffers: 152736 kB\nCached: 448568 kB\nSwapCached: 0 kB\nActive: 446596 kB\nInactive: 277556 kB\nActive(anon): 118676 kB\nInactive(anon): 4412 kB\nActive(file): 327920 kB\nInactive(file): 273144 kB\nUnevictable: 0 kB\nMlocked: 0 kB\nHighTotal: 137160 kB\nHighFree: 252 kB\nLowTotal: 888940 kB\nLowFree: 263624 kB\nSwapTotal: 397304 kB\nSwapFree: 397304 kB\nDirty: 72 kB\nWriteback: 0 kB\nAnonPages: 122860 kB\nMapped: 10304 kB\nShmem: 232 kB\nSlab: 28008 kB\nSReclaimable: 23192 kB\nSUnreclaim: 4816 kB\nKernelStack: 952 kB\nPageTables: 1008 kB\nNFS_Unstable: 0 kB\nBounce: 0 kB\nWritebackTmp: 0 kB\nCommitLimit: 910352 kB\nCommitted_AS: 321168 kB\nVmallocTotal: 122880 kB\nVmallocUsed: 3540 kB\nVmallocChunk: 114424 kB\nHardwareCorrupted: 0 kB\nHugePages_Total: 0\nHugePages_Free: 0\nHugePages_Rsvd: 0\nHugePages_Surp: 0\nHugepagesize: 2048 kB\nDirectMap4k: 10232 kB\nDirectMap2M: 901120 kB\n">>}, {replication,[{"default",1.0}]}, {system_memory_data, [{system_total_memory,1050726400}, {free_swap,406839296}, {total_swap,406839296}, {cached_memory,459333632}, {buffered_memory,156401664}, {free_memory,270209024}, {total_memory,1050726400}]}, {statistics, [{wall_clock,{530410,1}}, {context_switches,{102123,0}}, {garbage_collection,{22388,116469634,0}}, {io,{{input,7977564},{output,6143063}}}, {reductions,{75471574,2087262}}, {run_queue,0}, {runtime,{8220,190}}]}]}] INFO REPORT <6516.92.0> 2011-05-07 17:15:03 =============================================================================== ns_1@127.0.0.1:ns_doctor:82: Current node statuses: [{'ns_1@127.0.0.1', [{last_heard,{1304,813703,60271}}, {active_buckets,["default"]}, {memory, [{total,15455168}, {processes,5221584}, {processes_used,5212240}, {system,10233584}, {atom,518633}, {atom_used,496964}, {binary,169704}, {code,4014832}, {ets,637676}]}, {cluster_compatibility_version,1}, {version, [{kernel,"2.13.4"}, {mnesia,"4.4.12"}, {stdlib,"1.16.4"}, {os_mon,"2.2.4"}, {ns_server,"1.6.5.4r"}, {menelaus,"1.6.5.4r"}, {sasl,"2.1.8"}]}, {system_arch,"i686-pc-linux-gnu"}, {wall_clock,600}, {memory_data,{1050726400,781803520,{<6516.990.0>,671544}}}, {disk_data, [{"/",7583436,18}, {"/dev",508460,1}, {"/dev/shm",513048,0}, {"/var/run",513048,1}, {"/var/lock",513048,0}, {"/lib/init/rw",513048,0}, {"/boot",233191,8}]}, {meminfo, <<"MemTotal: 1026100 kB\nMemFree: 263248 kB\nBuffers: 152736 kB\nCached: 448800 kB\nSwapCached: 0 kB\nActive: 446692 kB\nInactive: 277936 kB\nActive(anon): 118912 kB\nInactive(anon): 4412 kB\nActive(file): 327780 kB\nInactive(file): 273524 kB\nUnevictable: 0 kB\nMlocked: 0 kB\nHighTotal: 137160 kB\nHighFree: 252 kB\nLowTotal: 888940 kB\nLowFree: 262996 kB\nSwapTotal: 397304 kB\nSwapFree: 397304 kB\nDirty: 80 kB\nWriteback: 0 kB\nAnonPages: 123092 kB\nMapped: 10304 kB\nShmem: 232 kB\nSlab: 28040 kB\nSReclaimable: 23212 kB\nSUnreclaim: 4828 kB\nKernelStack: 944 kB\nPageTables: 1008 kB\nNFS_Unstable: 0 kB\nBounce: 0 kB\nWritebackTmp: 0 kB\nCommitLimit: 910352 kB\nCommitted_AS: 321168 kB\nVmallocTotal: 122880 kB\nVmallocUsed: 3540 kB\nVmallocChunk: 114424 kB\nHardwareCorrupted: 0 kB\nHugePages_Total: 0\nHugePages_Free: 0\nHugePages_Rsvd: 0\nHugePages_Surp: 0\nHugepagesize: 2048 kB\nDirectMap4k: 10232 kB\nDirectMap2M: 901120 kB\n">>}, {replication,[{"default",1.0}]}, {system_memory_data, [{system_total_memory,1050726400}, {free_swap,406839296}, {total_swap,406839296}, {cached_memory,459571200}, {buffered_memory,156401664}, {free_memory,269565952}, {total_memory,1050726400}]}, {statistics, [{wall_clock,{590410,1}}, {context_switches,{113798,0}}, {garbage_collection,{25020,127621416,0}}, {io,{{input,9621923},{output,8025769}}}, {reductions,{85929315,2040360}}, {run_queue,0}, {runtime,{9230,220}}]}]}] INFO REPORT <6516.224.0> 2011-05-07 17:15:19 =============================================================================== ns_1@127.0.0.1:stats_collector:71: Stats for bucket "default": auth_cmds 4 auth_errors 0 bucket_conns 5 bytes_read 39229855 bytes_written 41361319 cas_badval 0 cas_hits 0 cas_misses 0 cmd_flush 0 cmd_get 204800 cmd_set 215040 conn_yields 0 connection_structures 15 curr_connections 15 curr_items 204800 curr_items_tot 204800 daemon_connections 10 decr_hits 0 decr_misses 0 delete_hits 0 delete_misses 0 ep_bg_fetched 0 ep_commit_num 457 ep_commit_time 0 ep_commit_time_total 5 ep_data_age 1 ep_data_age_highwat 5 ep_db_cleaner_status complete ep_db_strategy multiMTVBDB ep_dbinit 1 ep_dbname /var/opt/membase/1.6.5.4r/data/ns_1/default ep_dbshards 4 ep_expired 0 ep_flush_duration 0 ep_flush_duration_highwat 1 ep_flush_duration_total 6 ep_flush_preempts 0 ep_flusher_state running ep_flusher_todo 0 ep_io_num_read 0 ep_io_num_write 215040 ep_io_read_bytes 0 ep_io_write_bytes 24514560 ep_item_begin_failed 0 ep_item_commit_failed 0 ep_item_flush_expired 0 ep_item_flush_failed 0 ep_kv_size 35840000 ep_max_data_size 839909376 ep_max_txn_size 1000 ep_mem_high_wat 629932032 ep_mem_low_wat 503945625 ep_min_data_age 0 ep_num_active_non_resident 0 ep_num_eject_failures 0 ep_num_eject_replicas 0 ep_num_expiry_pager_runs 0 ep_num_non_resident 0 ep_num_not_my_vbuckets 0 ep_num_pager_runs 0 ep_num_value_ejects 0 ep_oom_errors 0 ep_overhead 12960032 ep_pending_ops 0 ep_pending_ops_max 0 ep_pending_ops_max_duration 0 ep_pending_ops_total 0 ep_queue_age_cap 900 ep_queue_size 0 ep_storage_age 0 ep_storage_age_highwat 2 ep_storage_type featured ep_store_max_concurrency 10 ep_store_max_readers 9 ep_store_max_readwrite 1 ep_tap_bg_fetch_requeued 0 ep_tap_bg_fetched 0 ep_tap_keepalive 300 ep_tmp_oom_errors 0 ep_too_old 0 ep_too_young 0 ep_total_cache_size 35840000 ep_total_del_items 0 ep_total_enqueued 215239 ep_total_new_items 204800 ep_total_persisted 215040 ep_vbucket_del 0 ep_vbucket_del_fail 0 ep_version 1.6.5.3.1_8_gce31baa ep_warmed_up 0 ep_warmup true ep_warmup_dups 0 ep_warmup_oom 0 ep_warmup_thread complete ep_warmup_time 29657 get_hits 204800 get_misses 0 incr_hits 0 incr_misses 0 libevent 2.0.7-rc limit_maxbytes 67108864 mem_used 48800032 pid 5769 pointer_size 32 rejected_conns 0 rusage_system 2.440152 rusage_user 4.116257 threads 4 time 1304813719 total_connections 15 uptime 618 version 1.4.4_382_g9df3289 INFO REPORT <6516.92.0> 2011-05-07 17:16:03 =============================================================================== ns_1@127.0.0.1:ns_doctor:82: Current node statuses: [{'ns_1@127.0.0.1', [{last_heard,{1304,813763,60248}}, {active_buckets,["default"]}, {memory, [{total,14660072}, {processes,4450528}, {processes_used,4439792}, {system,10209544}, {atom,518633}, {atom_used,496964}, {binary,126984}, {code,4014832}, {ets,657308}]}, {cluster_compatibility_version,1}, {version, [{kernel,"2.13.4"}, {mnesia,"4.4.12"}, {stdlib,"1.16.4"}, {os_mon,"2.2.4"}, {ns_server,"1.6.5.4r"}, {menelaus,"1.6.5.4r"}, {sasl,"2.1.8"}]}, {system_arch,"i686-pc-linux-gnu"}, {wall_clock,660}, {memory_data,{1050726400,783216640,{<6516.70.0>,486376}}}, {disk_data, [{"/",7583436,18}, {"/dev",508460,1}, {"/dev/shm",513048,0}, {"/var/run",513048,1}, {"/var/lock",513048,0}, {"/lib/init/rw",513048,0}, {"/boot",233191,8}]}, {meminfo, <<"MemTotal: 1026100 kB\nMemFree: 262148 kB\nBuffers: 152736 kB\nCached: 449052 kB\nSwapCached: 0 kB\nActive: 447680 kB\nInactive: 278068 kB\nActive(anon): 119780 kB\nInactive(anon): 4412 kB\nActive(file): 327900 kB\nInactive(file): 273656 kB\nUnevictable: 0 kB\nMlocked: 0 kB\nHighTotal: 137160 kB\nHighFree: 252 kB\nLowTotal: 888940 kB\nLowFree: 261896 kB\nSwapTotal: 397304 kB\nSwapFree: 397304 kB\nDirty: 68 kB\nWriteback: 0 kB\nAnonPages: 123960 kB\nMapped: 10304 kB\nShmem: 232 kB\nSlab: 28024 kB\nSReclaimable: 23212 kB\nSUnreclaim: 4812 kB\nKernelStack: 944 kB\nPageTables: 1012 kB\nNFS_Unstable: 0 kB\nBounce: 0 kB\nWritebackTmp: 0 kB\nCommitLimit: 910352 kB\nCommitted_AS: 321280 kB\nVmallocTotal: 122880 kB\nVmallocUsed: 3540 kB\nVmallocChunk: 114424 kB\nHardwareCorrupted: 0 kB\nHugePages_Total: 0\nHugePages_Free: 0\nHugePages_Rsvd: 0\nHugePages_Surp: 0\nHugepagesize: 2048 kB\nDirectMap4k: 10232 kB\nDirectMap2M: 901120 kB\n">>}, {replication,[{"default",1.0}]}, {system_memory_data, [{system_total_memory,1050726400}, {free_swap,406839296}, {total_swap,406839296}, {cached_memory,459829248}, {buffered_memory,156401664}, {free_memory,268439552}, {total_memory,1050726400}]}, {statistics, [{wall_clock,{650410,1}}, {context_switches,{125141,0}}, {garbage_collection,{27327,141366924,0}}, {io,{{input,9642415},{output,8351055}}}, {reductions,{96652868,1757453}}, {run_queue,0}, {runtime,{10280,180}}]}]}] INFO REPORT <6516.224.0> 2011-05-07 17:16:59 =============================================================================== ns_1@127.0.0.1:stats_collector:71: Stats for bucket "default": auth_cmds 4 auth_errors 0 bucket_conns 5 bytes_read 39234911 bytes_written 42929786 cas_badval 0 cas_hits 0 cas_misses 0 cmd_flush 0 cmd_get 204800 cmd_set 215040 conn_yields 0 connection_structures 15 curr_connections 15 curr_items 204800 curr_items_tot 204800 daemon_connections 10 decr_hits 0 decr_misses 0 delete_hits 0 delete_misses 0 ep_bg_fetched 0 ep_commit_num 457 ep_commit_time 0 ep_commit_time_total 5 ep_data_age 1 ep_data_age_highwat 5 ep_db_cleaner_status complete ep_db_strategy multiMTVBDB ep_dbinit 1 ep_dbname /var/opt/membase/1.6.5.4r/data/ns_1/default ep_dbshards 4 ep_expired 0 ep_flush_duration 0 ep_flush_duration_highwat 1 ep_flush_duration_total 6 ep_flush_preempts 0 ep_flusher_state running ep_flusher_todo 0 ep_io_num_read 0 ep_io_num_write 215040 ep_io_read_bytes 0 ep_io_write_bytes 24514560 ep_item_begin_failed 0 ep_item_commit_failed 0 ep_item_flush_expired 0 ep_item_flush_failed 0 ep_kv_size 35840000 ep_max_data_size 839909376 ep_max_txn_size 1000 ep_mem_high_wat 629932032 ep_mem_low_wat 503945625 ep_min_data_age 0 ep_num_active_non_resident 0 ep_num_eject_failures 0 ep_num_eject_replicas 0 ep_num_expiry_pager_runs 0 ep_num_non_resident 0 ep_num_not_my_vbuckets 0 ep_num_pager_runs 0 ep_num_value_ejects 0 ep_oom_errors 0 ep_overhead 12960032 ep_pending_ops 0 ep_pending_ops_max 0 ep_pending_ops_max_duration 0 ep_pending_ops_total 0 ep_queue_age_cap 900 ep_queue_size 0 ep_storage_age 0 ep_storage_age_highwat 2 ep_storage_type featured ep_store_max_concurrency 10 ep_store_max_readers 9 ep_store_max_readwrite 1 ep_tap_bg_fetch_requeued 0 ep_tap_bg_fetched 0 ep_tap_keepalive 300 ep_tmp_oom_errors 0 ep_too_old 0 ep_too_young 0 ep_total_cache_size 35840000 ep_total_del_items 0 ep_total_enqueued 215239 ep_total_new_items 204800 ep_total_persisted 215040 ep_vbucket_del 0 ep_vbucket_del_fail 0 ep_version 1.6.5.3.1_8_gce31baa ep_warmed_up 0 ep_warmup true ep_warmup_dups 0 ep_warmup_oom 0 ep_warmup_thread complete ep_warmup_time 29657 get_hits 204800 get_misses 0 incr_hits 0 incr_misses 0 libevent 2.0.7-rc limit_maxbytes 67108864 mem_used 48800032 pid 5769 pointer_size 32 rejected_conns 0 rusage_system 2.440152 rusage_user 4.148259 threads 4 time 1304813819 total_connections 15 uptime 718 version 1.4.4_382_g9df3289 INFO REPORT <6516.92.0> 2011-05-07 17:17:03 =============================================================================== ns_1@127.0.0.1:ns_doctor:82: Current node statuses: [{'ns_1@127.0.0.1', [{last_heard,{1304,813823,60198}}, {active_buckets,["default"]}, {memory, [{total,15027880}, {processes,4783400}, {processes_used,4774472}, {system,10244480}, {atom,518633}, {atom_used,496964}, {binary,126864}, {code,4014832}, {ets,690348}]}, {cluster_compatibility_version,1}, {version, [{kernel,"2.13.4"}, {mnesia,"4.4.12"}, {stdlib,"1.16.4"}, {os_mon,"2.2.4"}, {ns_server,"1.6.5.4r"}, {menelaus,"1.6.5.4r"}, {sasl,"2.1.8"}]}, {system_arch,"i686-pc-linux-gnu"}, {wall_clock,720}, {memory_data,{1050726400,783749120,{<6516.70.0>,486376}}}, {disk_data, [{"/",7583436,18}, {"/dev",508460,1}, {"/dev/shm",513048,0}, {"/var/run",513048,1}, {"/var/lock",513048,0}, {"/lib/init/rw",513048,0}, {"/boot",233191,8}]}, {meminfo, <<"MemTotal: 1026100 kB\nMemFree: 260852 kB\nBuffers: 152740 kB\nCached: 449328 kB\nSwapCached: 0 kB\nActive: 448768 kB\nInactive: 278212 kB\nActive(anon): 120748 kB\nInactive(anon): 4412 kB\nActive(file): 328020 kB\nInactive(file): 273800 kB\nUnevictable: 0 kB\nMlocked: 0 kB\nHighTotal: 137160 kB\nHighFree: 252 kB\nLowTotal: 888940 kB\nLowFree: 260600 kB\nSwapTotal: 397304 kB\nSwapFree: 397304 kB\nDirty: 92 kB\nWriteback: 0 kB\nAnonPages: 124920 kB\nMapped: 10328 kB\nShmem: 240 kB\nSlab: 28100 kB\nSReclaimable: 23228 kB\nSUnreclaim: 4872 kB\nKernelStack: 968 kB\nPageTables: 1060 kB\nNFS_Unstable: 0 kB\nBounce: 0 kB\nWritebackTmp: 0 kB\nCommitLimit: 910352 kB\nCommitted_AS: 325100 kB\nVmallocTotal: 122880 kB\nVmallocUsed: 3540 kB\nVmallocChunk: 114424 kB\nHardwareCorrupted: 0 kB\nHugePages_Total: 0\nHugePages_Free: 0\nHugePages_Rsvd: 0\nHugePages_Surp: 0\nHugepagesize: 2048 kB\nDirectMap4k: 10232 kB\nDirectMap2M: 901120 kB\n">>}, {replication,[{"default",1.0}]}, {system_memory_data, [{system_total_memory,1050726400}, {free_swap,406839296}, {total_swap,406839296}, {cached_memory,460111872}, {buffered_memory,156405760}, {free_memory,267112448}, {total_memory,1050726400}]}, {statistics, [{wall_clock,{710410,1}}, {context_switches,{135669,0}}, {garbage_collection,{29498,152821056,0}}, {io,{{input,9657727},{output,8622689}}}, {reductions,{106958501,1753649}}, {run_queue,0}, {runtime,{11250,170}}]}]}] INFO REPORT <6516.92.0> 2011-05-07 17:18:03 =============================================================================== ns_1@127.0.0.1:ns_doctor:82: Current node statuses: [{'ns_1@127.0.0.1', [{last_heard,{1304,813883,60238}}, {active_buckets,["default"]}, {memory, [{total,14431432}, {processes,4162368}, {processes_used,4151008}, {system,10269064}, {atom,518633}, {atom_used,496964}, {binary,117064}, {code,4014832}, {ets,725180}]}, {cluster_compatibility_version,1}, {version, [{kernel,"2.13.4"}, {mnesia,"4.4.12"}, {stdlib,"1.16.4"}, {os_mon,"2.2.4"}, {ns_server,"1.6.5.4r"}, {menelaus,"1.6.5.4r"}, {sasl,"2.1.8"}]}, {system_arch,"i686-pc-linux-gnu"}, {wall_clock,780}, {memory_data,{1050726400,783122432,{<6516.70.0>,486376}}}, {disk_data, [{"/",7583436,18}, {"/dev",508460,1}, {"/dev/shm",513048,0}, {"/var/run",513048,1}, {"/var/lock",513048,0}, {"/lib/init/rw",513048,0}, {"/boot",233191,8}]}, {meminfo, <<"MemTotal: 1026100 kB\nMemFree: 261472 kB\nBuffers: 152748 kB\nCached: 448368 kB\nSwapCached: 0 kB\nActive: 448780 kB\nInactive: 277508 kB\nActive(anon): 121000 kB\nInactive(anon): 4412 kB\nActive(file): 327780 kB\nInactive(file): 273096 kB\nUnevictable: 0 kB\nMlocked: 0 kB\nHighTotal: 137160 kB\nHighFree: 252 kB\nLowTotal: 888940 kB\nLowFree: 261220 kB\nSwapTotal: 397304 kB\nSwapFree: 397304 kB\nDirty: 88 kB\nWriteback: 0 kB\nAnonPages: 125176 kB\nMapped: 10328 kB\nShmem: 240 kB\nSlab: 28100 kB\nSReclaimable: 23232 kB\nSUnreclaim: 4868 kB\nKernelStack: 968 kB\nPageTables: 1060 kB\nNFS_Unstable: 0 kB\nBounce: 0 kB\nWritebackTmp: 0 kB\nCommitLimit: 910352 kB\nCommitted_AS: 325100 kB\nVmallocTotal: 122880 kB\nVmallocUsed: 3540 kB\nVmallocChunk: 114424 kB\nHardwareCorrupted: 0 kB\nHugePages_Total: 0\nHugePages_Free: 0\nHugePages_Rsvd: 0\nHugePages_Surp: 0\nHugepagesize: 2048 kB\nDirectMap4k: 10232 kB\nDirectMap2M: 901120 kB\n">>}, {replication,[{"default",1.0}]}, {system_memory_data, [{system_total_memory,1050726400}, {free_swap,406839296}, {total_swap,406839296}, {cached_memory,459128832}, {buffered_memory,156413952}, {free_memory,267747328}, {total_memory,1050726400}]}, {statistics, [{wall_clock,{770410,1}}, {context_switches,{146711,0}}, {garbage_collection,{31787,165061258,0}}, {io,{{input,10428463},{output,9332067}}}, {reductions,{117393754,1714805}}, {run_queue,0}, {runtime,{12290,170}}]}]}] INFO REPORT <6516.224.0> 2011-05-07 17:18:39 =============================================================================== ns_1@127.0.0.1:stats_collector:71: Stats for bucket "default": auth_cmds 4 auth_errors 0 bucket_conns 5 bytes_read 39239943 bytes_written 44493500 cas_badval 0 cas_hits 0 cas_misses 0 cmd_flush 0 cmd_get 204800 cmd_set 215040 conn_yields 0 connection_structures 15 curr_connections 15 curr_items 204800 curr_items_tot 204800 daemon_connections 10 decr_hits 0 decr_misses 0 delete_hits 0 delete_misses 0 ep_bg_fetched 0 ep_commit_num 457 ep_commit_time 0 ep_commit_time_total 5 ep_data_age 1 ep_data_age_highwat 5 ep_db_cleaner_status complete ep_db_strategy multiMTVBDB ep_dbinit 1 ep_dbname /var/opt/membase/1.6.5.4r/data/ns_1/default ep_dbshards 4 ep_expired 0 ep_flush_duration 0 ep_flush_duration_highwat 1 ep_flush_duration_total 6 ep_flush_preempts 0 ep_flusher_state running ep_flusher_todo 0 ep_io_num_read 0 ep_io_num_write 215040 ep_io_read_bytes 0 ep_io_write_bytes 24514560 ep_item_begin_failed 0 ep_item_commit_failed 0 ep_item_flush_expired 0 ep_item_flush_failed 0 ep_kv_size 35840000 ep_max_data_size 839909376 ep_max_txn_size 1000 ep_mem_high_wat 629932032 ep_mem_low_wat 503945625 ep_min_data_age 0 ep_num_active_non_resident 0 ep_num_eject_failures 0 ep_num_eject_replicas 0 ep_num_expiry_pager_runs 0 ep_num_non_resident 0 ep_num_not_my_vbuckets 0 ep_num_pager_runs 0 ep_num_value_ejects 0 ep_oom_errors 0 ep_overhead 12960032 ep_pending_ops 0 ep_pending_ops_max 0 ep_pending_ops_max_duration 0 ep_pending_ops_total 0 ep_queue_age_cap 900 ep_queue_size 0 ep_storage_age 0 ep_storage_age_highwat 2 ep_storage_type featured ep_store_max_concurrency 10 ep_store_max_readers 9 ep_store_max_readwrite 1 ep_tap_bg_fetch_requeued 0 ep_tap_bg_fetched 0 ep_tap_keepalive 300 ep_tmp_oom_errors 0 ep_too_old 0 ep_too_young 0 ep_total_cache_size 35840000 ep_total_del_items 0 ep_total_enqueued 215239 ep_total_new_items 204800 ep_total_persisted 215040 ep_vbucket_del 0 ep_vbucket_del_fail 0 ep_version 1.6.5.3.1_8_gce31baa ep_warmed_up 0 ep_warmup true ep_warmup_dups 0 ep_warmup_oom 0 ep_warmup_thread complete ep_warmup_time 29657 get_hits 204800 get_misses 0 incr_hits 0 incr_misses 0 libevent 2.0.7-rc limit_maxbytes 67108864 mem_used 48800032 pid 5769 pointer_size 32 rejected_conns 0 rusage_system 2.440152 rusage_user 4.192262 threads 4 time 1304813918 total_connections 15 uptime 817 version 1.4.4_382_g9df3289 INFO REPORT <6516.92.0> 2011-05-07 17:19:03 =============================================================================== ns_1@127.0.0.1:ns_doctor:82: Current node statuses: [{'ns_1@127.0.0.1', [{last_heard,{1304,813943,60237}}, {active_buckets,["default"]}, {memory, [{total,15378520}, {processes,5041064}, {processes_used,5030216}, {system,10337456}, {atom,518633}, {atom_used,496964}, {binary,151456}, {code,4014832}, {ets,758380}]}, {cluster_compatibility_version,1}, {version, [{kernel,"2.13.4"}, {mnesia,"4.4.12"}, {stdlib,"1.16.4"}, {os_mon,"2.2.4"}, {ns_server,"1.6.5.4r"}, {menelaus,"1.6.5.4r"}, {sasl,"2.1.8"}]}, {system_arch,"i686-pc-linux-gnu"}, {wall_clock,840}, {memory_data,{1050726400,783503360,{<6516.1399.0>,786172}}}, {disk_data, [{"/",7583436,18}, {"/dev",508460,1}, {"/dev/shm",513048,0}, {"/var/run",513048,1}, {"/var/lock",513048,0}, {"/lib/init/rw",513048,0}, {"/boot",233191,8}]}, {meminfo, <<"MemTotal: 1026100 kB\nMemFree: 260464 kB\nBuffers: 152748 kB\nCached: 448616 kB\nSwapCached: 0 kB\nActive: 449944 kB\nInactive: 277648 kB\nActive(anon): 122056 kB\nInactive(anon): 4412 kB\nActive(file): 327888 kB\nInactive(file): 273236 kB\nUnevictable: 0 kB\nMlocked: 0 kB\nHighTotal: 137160 kB\nHighFree: 252 kB\nLowTotal: 888940 kB\nLowFree: 260212 kB\nSwapTotal: 397304 kB\nSwapFree: 397304 kB\nDirty: 68 kB\nWriteback: 0 kB\nAnonPages: 126284 kB\nMapped: 10332 kB\nShmem: 240 kB\nSlab: 28104 kB\nSReclaimable: 23232 kB\nSUnreclaim: 4872 kB\nKernelStack: 984 kB\nPageTables: 1064 kB\nNFS_Unstable: 0 kB\nBounce: 0 kB\nWritebackTmp: 0 kB\nCommitLimit: 910352 kB\nCommitted_AS: 330428 kB\nVmallocTotal: 122880 kB\nVmallocUsed: 3540 kB\nVmallocChunk: 114424 kB\nHardwareCorrupted: 0 kB\nHugePages_Total: 0\nHugePages_Free: 0\nHugePages_Rsvd: 0\nHugePages_Surp: 0\nHugepagesize: 2048 kB\nDirectMap4k: 10232 kB\nDirectMap2M: 901120 kB\n">>}, {replication,[{"default",1.0}]}, {system_memory_data, [{system_total_memory,1050726400}, {free_swap,406839296}, {total_swap,406839296}, {cached_memory,459382784}, {buffered_memory,156413952}, {free_memory,266715136}, {total_memory,1050726400}]}, {statistics, [{wall_clock,{830410,1}}, {context_switches,{157388,0}}, {garbage_collection,{34054,177047545,0}}, {io,{{input,10445931},{output,9630171}}}, {reductions,{127816431,2095919}}, {run_queue,0}, {runtime,{13300,180}}]}]}] INFO REPORT <6516.92.0> 2011-05-07 17:20:03 =============================================================================== ns_1@127.0.0.1:ns_doctor:82: Current node statuses: [{'ns_1@127.0.0.1', [{last_heard,{1304,814003,59749}}, {active_buckets,["default"]}, {memory, [{total,15404728}, {processes,5097984}, {processes_used,5087552}, {system,10306744}, {atom,518633}, {atom_used,496964}, {binary,119760}, {code,4014832}, {ets,758428}]}, {cluster_compatibility_version,1}, {version, [{kernel,"2.13.4"}, {mnesia,"4.4.12"}, {stdlib,"1.16.4"}, {os_mon,"2.2.4"}, {ns_server,"1.6.5.4r"}, {menelaus,"1.6.5.4r"}, {sasl,"2.1.8"}]}, {system_arch,"i686-pc-linux-gnu"}, {wall_clock,900}, {memory_data,{1050726400,783732736,{<6516.268.0>,556860}}}, {disk_data, [{"/",7583436,18}, {"/dev",508460,1}, {"/dev/shm",513048,0}, {"/var/run",513048,1}, {"/var/lock",513048,0}, {"/lib/init/rw",513048,0}, {"/boot",233191,8}]}, {meminfo, <<"MemTotal: 1026100 kB\nMemFree: 260860 kB\nBuffers: 152748 kB\nCached: 448876 kB\nSwapCached: 0 kB\nActive: 449044 kB\nInactive: 277796 kB\nActive(anon): 121044 kB\nInactive(anon): 4412 kB\nActive(file): 328000 kB\nInactive(file): 273384 kB\nUnevictable: 0 kB\nMlocked: 0 kB\nHighTotal: 137160 kB\nHighFree: 252 kB\nLowTotal: 888940 kB\nLowFree: 260608 kB\nSwapTotal: 397304 kB\nSwapFree: 397304 kB\nDirty: 80 kB\nWriteback: 0 kB\nAnonPages: 125216 kB\nMapped: 10328 kB\nShmem: 240 kB\nSlab: 28104 kB\nSReclaimable: 23232 kB\nSUnreclaim: 4872 kB\nKernelStack: 960 kB\nPageTables: 1060 kB\nNFS_Unstable: 0 kB\nBounce: 0 kB\nWritebackTmp: 0 kB\nCommitLimit: 910352 kB\nCommitted_AS: 325100 kB\nVmallocTotal: 122880 kB\nVmallocUsed: 3540 kB\nVmallocChunk: 114424 kB\nHardwareCorrupted: 0 kB\nHugePages_Total: 0\nHugePages_Free: 0\nHugePages_Rsvd: 0\nHugePages_Surp: 0\nHugepagesize: 2048 kB\nDirectMap4k: 10232 kB\nDirectMap2M: 901120 kB\n">>}, {replication,[{"default",1.0}]}, {system_memory_data, [{system_total_memory,1050726400}, {free_swap,406839296}, {total_swap,406839296}, {cached_memory,459649024}, {buffered_memory,156413952}, {free_memory,267120640}, {total_memory,1050726400}]}, {statistics, [{wall_clock,{890409,0}}, {context_switches,{169461,0}}, {garbage_collection,{36642,190664472,0}}, {io,{{input,10504173},{output,9997618}}}, {reductions,{138765218,2058826}}, {run_queue,0}, {runtime,{14380,200}}]}]}] INFO REPORT <6516.224.0> 2011-05-07 17:20:19 =============================================================================== ns_1@127.0.0.1:stats_collector:71: Stats for bucket "default": auth_cmds 4 auth_errors 0 bucket_conns 5 bytes_read 39244999 bytes_written 46061967 cas_badval 0 cas_hits 0 cas_misses 0 cmd_flush 0 cmd_get 204800 cmd_set 215040 conn_yields 0 connection_structures 15 curr_connections 15 curr_items 204800 curr_items_tot 204800 daemon_connections 10 decr_hits 0 decr_misses 0 delete_hits 0 delete_misses 0 ep_bg_fetched 0 ep_commit_num 457 ep_commit_time 0 ep_commit_time_total 5 ep_data_age 1 ep_data_age_highwat 5 ep_db_cleaner_status complete ep_db_strategy multiMTVBDB ep_dbinit 1 ep_dbname /var/opt/membase/1.6.5.4r/data/ns_1/default ep_dbshards 4 ep_expired 0 ep_flush_duration 0 ep_flush_duration_highwat 1 ep_flush_duration_total 6 ep_flush_preempts 0 ep_flusher_state running ep_flusher_todo 0 ep_io_num_read 0 ep_io_num_write 215040 ep_io_read_bytes 0 ep_io_write_bytes 24514560 ep_item_begin_failed 0 ep_item_commit_failed 0 ep_item_flush_expired 0 ep_item_flush_failed 0 ep_kv_size 35840000 ep_max_data_size 839909376 ep_max_txn_size 1000 ep_mem_high_wat 629932032 ep_mem_low_wat 503945625 ep_min_data_age 0 ep_num_active_non_resident 0 ep_num_eject_failures 0 ep_num_eject_replicas 0 ep_num_expiry_pager_runs 0 ep_num_non_resident 0 ep_num_not_my_vbuckets 0 ep_num_pager_runs 0 ep_num_value_ejects 0 ep_oom_errors 0 ep_overhead 12960032 ep_pending_ops 0 ep_pending_ops_max 0 ep_pending_ops_max_duration 0 ep_pending_ops_total 0 ep_queue_age_cap 900 ep_queue_size 0 ep_storage_age 0 ep_storage_age_highwat 2 ep_storage_type featured ep_store_max_concurrency 10 ep_store_max_readers 9 ep_store_max_readwrite 1 ep_tap_bg_fetch_requeued 0 ep_tap_bg_fetched 0 ep_tap_keepalive 300 ep_tmp_oom_errors 0 ep_too_old 0 ep_too_young 0 ep_total_cache_size 35840000 ep_total_del_items 0 ep_total_enqueued 215239 ep_total_new_items 204800 ep_total_persisted 215040 ep_vbucket_del 0 ep_vbucket_del_fail 0 ep_version 1.6.5.3.1_8_gce31baa ep_warmed_up 0 ep_warmup true ep_warmup_dups 0 ep_warmup_oom 0 ep_warmup_thread complete ep_warmup_time 29657 get_hits 204800 get_misses 0 incr_hits 0 incr_misses 0 libevent 2.0.7-rc limit_maxbytes 67108864 mem_used 48800032 pid 5769 pointer_size 32 rejected_conns 0 rusage_system 2.444152 rusage_user 4.228264 threads 4 time 1304814018 total_connections 15 uptime 917 version 1.4.4_382_g9df3289 INFO REPORT <6516.92.0> 2011-05-07 17:21:03 =============================================================================== ns_1@127.0.0.1:ns_doctor:82: Current node statuses: [{'ns_1@127.0.0.1', [{last_heard,{1304,814063,59760}}, {active_buckets,["default"]}, {memory, [{total,14346448}, {processes,3999400}, {processes_used,3986648}, {system,10347048}, {atom,518633}, {atom_used,496964}, {binary,125216}, {code,4014832}, {ets,794580}]}, {cluster_compatibility_version,1}, {version, [{kernel,"2.13.4"}, {mnesia,"4.4.12"}, {stdlib,"1.16.4"}, {os_mon,"2.2.4"}, {ns_server,"1.6.5.4r"}, {menelaus,"1.6.5.4r"}, {sasl,"2.1.8"}]}, {system_arch,"i686-pc-linux-gnu"}, {wall_clock,960}, {memory_data,{1050726400,783884288,{<6516.268.0>,556860}}}, {disk_data, [{"/",7583436,18}, {"/dev",508460,1}, {"/dev/shm",513048,0}, {"/var/run",513048,1}, {"/var/lock",513048,0}, {"/lib/init/rw",513048,0}, {"/boot",233191,8}]}, {meminfo, <<"MemTotal: 1026100 kB\nMemFree: 260612 kB\nBuffers: 152752 kB\nCached: 449096 kB\nSwapCached: 0 kB\nActive: 448932 kB\nInactive: 278156 kB\nActive(anon): 121068 kB\nInactive(anon): 4412 kB\nActive(file): 327864 kB\nInactive(file): 273744 kB\nUnevictable: 0 kB\nMlocked: 0 kB\nHighTotal: 137160 kB\nHighFree: 252 kB\nLowTotal: 888940 kB\nLowFree: 260360 kB\nSwapTotal: 397304 kB\nSwapFree: 397304 kB\nDirty: 88 kB\nWriteback: 0 kB\nAnonPages: 125240 kB\nMapped: 10328 kB\nShmem: 240 kB\nSlab: 28104 kB\nSReclaimable: 23236 kB\nSUnreclaim: 4868 kB\nKernelStack: 960 kB\nPageTables: 1060 kB\nNFS_Unstable: 0 kB\nBounce: 0 kB\nWritebackTmp: 0 kB\nCommitLimit: 910352 kB\nCommitted_AS: 325100 kB\nVmallocTotal: 122880 kB\nVmallocUsed: 3540 kB\nVmallocChunk: 114424 kB\nHardwareCorrupted: 0 kB\nHugePages_Total: 0\nHugePages_Free: 0\nHugePages_Rsvd: 0\nHugePages_Surp: 0\nHugepagesize: 2048 kB\nDirectMap4k: 10232 kB\nDirectMap2M: 901120 kB\n">>}, {replication,[{"default",1.0}]}, {system_memory_data, [{system_total_memory,1050726400}, {free_swap,406839296}, {total_swap,406839296}, {cached_memory,459874304}, {buffered_memory,156418048}, {free_memory,266866688}, {total_memory,1050726400}]}, {statistics, [{wall_clock,{950410,1}}, {context_switches,{179956,0}}, {garbage_collection,{38997,201365028,0}}, {io,{{input,11272174},{output,10992219}}}, {reductions,{148940339,1712831}}, {run_queue,0}, {runtime,{15340,160}}]}]}] INFO REPORT <6516.224.0> 2011-05-07 17:21:59 =============================================================================== ns_1@127.0.0.1:stats_collector:71: Stats for bucket "default": auth_cmds 4 auth_errors 0 bucket_conns 5 bytes_read 39250055 bytes_written 47630461 cas_badval 0 cas_hits 0 cas_misses 0 cmd_flush 0 cmd_get 204800 cmd_set 215040 conn_yields 0 connection_structures 15 curr_connections 15 curr_items 204800 curr_items_tot 204800 daemon_connections 10 decr_hits 0 decr_misses 0 delete_hits 0 delete_misses 0 ep_bg_fetched 0 ep_commit_num 457 ep_commit_time 0 ep_commit_time_total 5 ep_data_age 1 ep_data_age_highwat 5 ep_db_cleaner_status complete ep_db_strategy multiMTVBDB ep_dbinit 1 ep_dbname /var/opt/membase/1.6.5.4r/data/ns_1/default ep_dbshards 4 ep_expired 0 ep_flush_duration 0 ep_flush_duration_highwat 1 ep_flush_duration_total 6 ep_flush_preempts 0 ep_flusher_state running ep_flusher_todo 0 ep_io_num_read 0 ep_io_num_write 215040 ep_io_read_bytes 0 ep_io_write_bytes 24514560 ep_item_begin_failed 0 ep_item_commit_failed 0 ep_item_flush_expired 0 ep_item_flush_failed 0 ep_kv_size 35840000 ep_max_data_size 839909376 ep_max_txn_size 1000 ep_mem_high_wat 629932032 ep_mem_low_wat 503945625 ep_min_data_age 0 ep_num_active_non_resident 0 ep_num_eject_failures 0 ep_num_eject_replicas 0 ep_num_expiry_pager_runs 0 ep_num_non_resident 0 ep_num_not_my_vbuckets 0 ep_num_pager_runs 0 ep_num_value_ejects 0 ep_oom_errors 0 ep_overhead 12960032 ep_pending_ops 0 ep_pending_ops_max 0 ep_pending_ops_max_duration 0 ep_pending_ops_total 0 ep_queue_age_cap 900 ep_queue_size 0 ep_storage_age 0 ep_storage_age_highwat 2 ep_storage_type featured ep_store_max_concurrency 10 ep_store_max_readers 9 ep_store_max_readwrite 1 ep_tap_bg_fetch_requeued 0 ep_tap_bg_fetched 0 ep_tap_keepalive 300 ep_tmp_oom_errors 0 ep_too_old 0 ep_too_young 0 ep_total_cache_size 35840000 ep_total_del_items 0 ep_total_enqueued 215239 ep_total_new_items 204800 ep_total_persisted 215040 ep_vbucket_del 0 ep_vbucket_del_fail 0 ep_version 1.6.5.3.1_8_gce31baa ep_warmed_up 0 ep_warmup true ep_warmup_dups 0 ep_warmup_oom 0 ep_warmup_thread complete ep_warmup_time 29657 get_hits 204800 get_misses 0 incr_hits 0 incr_misses 0 libevent 2.0.7-rc limit_maxbytes 67108864 mem_used 48800032 pid 5769 pointer_size 32 rejected_conns 0 rusage_system 2.448153 rusage_user 4.264266 threads 4 time 1304814118 total_connections 15 uptime 1017 version 1.4.4_382_g9df3289 INFO REPORT <6516.92.0> 2011-05-07 17:22:03 =============================================================================== ns_1@127.0.0.1:ns_doctor:82: Current node statuses: [{'ns_1@127.0.0.1', [{last_heard,{1304,814123,59808}}, {active_buckets,["default"]}, {memory, [{total,14358136}, {processes,4017528}, {processes_used,4004776}, {system,10340608}, {atom,518633}, {atom_used,496964}, {binary,118656}, {code,4014832}, {ets,794412}]}, {cluster_compatibility_version,1}, {version, [{kernel,"2.13.4"}, {mnesia,"4.4.12"}, {stdlib,"1.16.4"}, {os_mon,"2.2.4"}, {ns_server,"1.6.5.4r"}, {menelaus,"1.6.5.4r"}, {sasl,"2.1.8"}]}, {system_arch,"i686-pc-linux-gnu"}, {wall_clock,1020}, {memory_data,{1050726400,784138240,{<6516.268.0>,556860}}}, {disk_data, [{"/",7583436,18}, {"/dev",508460,1}, {"/dev/shm",513048,0}, {"/var/run",513048,1}, {"/var/lock",513048,0}, {"/lib/init/rw",513048,0}, {"/boot",233191,8}]}, {meminfo, <<"MemTotal: 1026100 kB\nMemFree: 260364 kB\nBuffers: 152756 kB\nCached: 449348 kB\nSwapCached: 0 kB\nActive: 449052 kB\nInactive: 278304 kB\nActive(anon): 121080 kB\nInactive(anon): 4412 kB\nActive(file): 327972 kB\nInactive(file): 273892 kB\nUnevictable: 0 kB\nMlocked: 0 kB\nHighTotal: 137160 kB\nHighFree: 252 kB\nLowTotal: 888940 kB\nLowFree: 260112 kB\nSwapTotal: 397304 kB\nSwapFree: 397304 kB\nDirty: 68 kB\nWriteback: 0 kB\nAnonPages: 125256 kB\nMapped: 10328 kB\nShmem: 240 kB\nSlab: 28108 kB\nSReclaimable: 23236 kB\nSUnreclaim: 4872 kB\nKernelStack: 960 kB\nPageTables: 1060 kB\nNFS_Unstable: 0 kB\nBounce: 0 kB\nWritebackTmp: 0 kB\nCommitLimit: 910352 kB\nCommitted_AS: 325100 kB\nVmallocTotal: 122880 kB\nVmallocUsed: 3540 kB\nVmallocChunk: 114424 kB\nHardwareCorrupted: 0 kB\nHugePages_Total: 0\nHugePages_Free: 0\nHugePages_Rsvd: 0\nHugePages_Surp: 0\nHugepagesize: 2048 kB\nDirectMap4k: 10232 kB\nDirectMap2M: 901120 kB\n">>}, {replication,[{"default",1.0}]}, {system_memory_data, [{system_total_memory,1050726400}, {free_swap,406839296}, {total_swap,406839296}, {cached_memory,460132352}, {buffered_memory,156422144}, {free_memory,266612736}, {total_memory,1050726400}]}, {statistics, [{wall_clock,{1010410,1}}, {context_switches,{189857,0}}, {garbage_collection,{41106,211440916,0}}, {io,{{input,11284324},{output,11251054}}}, {reductions,{158985154,1713004}}, {run_queue,0}, {runtime,{16310,170}}]}]}] INFO REPORT <6516.92.0> 2011-05-07 17:23:03 =============================================================================== ns_1@127.0.0.1:ns_doctor:82: Current node statuses: [{'ns_1@127.0.0.1', [{last_heard,{1304,814183,60233}}, {active_buckets,["default"]}, {memory, [{total,15980192}, {processes,5599336}, {processes_used,5589944}, {system,10380856}, {atom,518633}, {atom_used,496964}, {binary,119632}, {code,4014832}, {ets,827452}]}, {cluster_compatibility_version,1}, {version, [{kernel,"2.13.4"}, {mnesia,"4.4.12"}, {stdlib,"1.16.4"}, {os_mon,"2.2.4"}, {ns_server,"1.6.5.4r"}, {menelaus,"1.6.5.4r"}, {sasl,"2.1.8"}]}, {system_arch,"i686-pc-linux-gnu"}, {wall_clock,1080}, {memory_data,{1050726400,784494592,{<6516.268.0>,556860}}}, {disk_data, [{"/",7583436,18}, {"/dev",508460,1}, {"/dev/shm",513048,0}, {"/var/run",513048,1}, {"/var/lock",513048,0}, {"/lib/init/rw",513048,0}, {"/boot",233191,8}]}, {meminfo, <<"MemTotal: 1026100 kB\nMemFree: 260116 kB\nBuffers: 152756 kB\nCached: 449604 kB\nSwapCached: 0 kB\nActive: 449196 kB\nInactive: 278432 kB\nActive(anon): 121104 kB\nInactive(anon): 4412 kB\nActive(file): 328092 kB\nInactive(file): 274020 kB\nUnevictable: 0 kB\nMlocked: 0 kB\nHighTotal: 137160 kB\nHighFree: 252 kB\nLowTotal: 888940 kB\nLowFree: 259864 kB\nSwapTotal: 397304 kB\nSwapFree: 397304 kB\nDirty: 88 kB\nWriteback: 0 kB\nAnonPages: 125280 kB\nMapped: 10328 kB\nShmem: 240 kB\nSlab: 28108 kB\nSReclaimable: 23236 kB\nSUnreclaim: 4872 kB\nKernelStack: 960 kB\nPageTables: 1060 kB\nNFS_Unstable: 0 kB\nBounce: 0 kB\nWritebackTmp: 0 kB\nCommitLimit: 910352 kB\nCommitted_AS: 325100 kB\nVmallocTotal: 122880 kB\nVmallocUsed: 3540 kB\nVmallocChunk: 114424 kB\nHardwareCorrupted: 0 kB\nHugePages_Total: 0\nHugePages_Free: 0\nHugePages_Rsvd: 0\nHugePages_Surp: 0\nHugepagesize: 2048 kB\nDirectMap4k: 10232 kB\nDirectMap2M: 901120 kB\n">>}, {replication,[{"default",1.0}]}, {system_memory_data, [{system_total_memory,1050726400}, {free_swap,406839296}, {total_swap,406839296}, {cached_memory,460394496}, {buffered_memory,156422144}, {free_memory,266358784}, {total_memory,1050726400}]}, {statistics, [{wall_clock,{1070410,1}}, {context_switches,{201336,0}}, {garbage_collection,{43617,223145629,0}}, {io,{{input,12176218},{output,12435281}}}, {reductions,{169548009,1908131}}, {run_queue,0}, {runtime,{17380,190}}]}]}] INFO REPORT <6516.224.0> 2011-05-07 17:23:39 =============================================================================== ns_1@127.0.0.1:stats_collector:71: Stats for bucket "default": auth_cmds 4 auth_errors 0 bucket_conns 5 bytes_read 39255111 bytes_written 49199087 cas_badval 0 cas_hits 0 cas_misses 0 cmd_flush 0 cmd_get 204800 cmd_set 215040 conn_yields 0 connection_structures 15 curr_connections 15 curr_items 204800 curr_items_tot 204800 daemon_connections 10 decr_hits 0 decr_misses 0 delete_hits 0 delete_misses 0 ep_bg_fetched 0 ep_commit_num 457 ep_commit_time 0 ep_commit_time_total 5 ep_data_age 1 ep_data_age_highwat 5 ep_db_cleaner_status complete ep_db_strategy multiMTVBDB ep_dbinit 1 ep_dbname /var/opt/membase/1.6.5.4r/data/ns_1/default ep_dbshards 4 ep_expired 0 ep_flush_duration 0 ep_flush_duration_highwat 1 ep_flush_duration_total 6 ep_flush_preempts 0 ep_flusher_state running ep_flusher_todo 0 ep_io_num_read 0 ep_io_num_write 215040 ep_io_read_bytes 0 ep_io_write_bytes 24514560 ep_item_begin_failed 0 ep_item_commit_failed 0 ep_item_flush_expired 0 ep_item_flush_failed 0 ep_kv_size 35840000 ep_max_data_size 839909376 ep_max_txn_size 1000 ep_mem_high_wat 629932032 ep_mem_low_wat 503945625 ep_min_data_age 0 ep_num_active_non_resident 0 ep_num_eject_failures 0 ep_num_eject_replicas 0 ep_num_expiry_pager_runs 0 ep_num_non_resident 0 ep_num_not_my_vbuckets 0 ep_num_pager_runs 0 ep_num_value_ejects 0 ep_oom_errors 0 ep_overhead 12960032 ep_pending_ops 0 ep_pending_ops_max 0 ep_pending_ops_max_duration 0 ep_pending_ops_total 0 ep_queue_age_cap 900 ep_queue_size 0 ep_storage_age 0 ep_storage_age_highwat 2 ep_storage_type featured ep_store_max_concurrency 10 ep_store_max_readers 9 ep_store_max_readwrite 1 ep_tap_bg_fetch_requeued 0 ep_tap_bg_fetched 0 ep_tap_keepalive 300 ep_tmp_oom_errors 0 ep_too_old 0 ep_too_young 0 ep_total_cache_size 35840000 ep_total_del_items 0 ep_total_enqueued 215239 ep_total_new_items 204800 ep_total_persisted 215040 ep_vbucket_del 0 ep_vbucket_del_fail 0 ep_version 1.6.5.3.1_8_gce31baa ep_warmed_up 0 ep_warmup true ep_warmup_dups 0 ep_warmup_oom 0 ep_warmup_thread complete ep_warmup_time 29657 get_hits 204800 get_misses 0 incr_hits 0 incr_misses 0 libevent 2.0.7-rc limit_maxbytes 67108864 mem_used 48800032 pid 5769 pointer_size 32 rejected_conns 0 rusage_system 2.452153 rusage_user 4.296268 threads 4 time 1304814219 total_connections 15 uptime 1118 version 1.4.4_382_g9df3289 INFO REPORT <6516.92.0> 2011-05-07 17:24:03 =============================================================================== ns_1@127.0.0.1:ns_doctor:82: Current node statuses: [{'ns_1@127.0.0.1', [{last_heard,{1304,814243,59765}}, {active_buckets,["default"]}, {memory, [{total,14881760}, {processes,4461840}, {processes_used,4450592}, {system,10419920}, {atom,518633}, {atom_used,496964}, {binary,125096}, {code,4014832}, {ets,862084}]}, {cluster_compatibility_version,1}, {version, [{kernel,"2.13.4"}, {mnesia,"4.4.12"}, {stdlib,"1.16.4"}, {os_mon,"2.2.4"}, {ns_server,"1.6.5.4r"}, {menelaus,"1.6.5.4r"}, {sasl,"2.1.8"}]}, {system_arch,"i686-pc-linux-gnu"}, {wall_clock,1140}, {memory_data,{1050726400,783503360,{<6516.268.0>,556860}}}, {disk_data, [{"/",7583436,18}, {"/dev",508460,1}, {"/dev/shm",513048,0}, {"/var/run",513048,1}, {"/var/lock",513048,0}, {"/lib/init/rw",513048,0}, {"/boot",233191,8}]}, {meminfo, <<"MemTotal: 1026100 kB\nMemFree: 260984 kB\nBuffers: 152760 kB\nCached: 448640 kB\nSwapCached: 0 kB\nActive: 448988 kB\nInactive: 277724 kB\nActive(anon): 121140 kB\nInactive(anon): 4412 kB\nActive(file): 327848 kB\nInactive(file): 273312 kB\nUnevictable: 0 kB\nMlocked: 0 kB\nHighTotal: 137160 kB\nHighFree: 252 kB\nLowTotal: 888940 kB\nLowFree: 260732 kB\nSwapTotal: 397304 kB\nSwapFree: 397304 kB\nDirty: 84 kB\nWriteback: 0 kB\nAnonPages: 125312 kB\nMapped: 10328 kB\nShmem: 240 kB\nSlab: 28112 kB\nSReclaimable: 23240 kB\nSUnreclaim: 4872 kB\nKernelStack: 960 kB\nPageTables: 1060 kB\nNFS_Unstable: 0 kB\nBounce: 0 kB\nWritebackTmp: 0 kB\nCommitLimit: 910352 kB\nCommitted_AS: 325100 kB\nVmallocTotal: 122880 kB\nVmallocUsed: 3540 kB\nVmallocChunk: 114424 kB\nHardwareCorrupted: 0 kB\nHugePages_Total: 0\nHugePages_Free: 0\nHugePages_Rsvd: 0\nHugePages_Surp: 0\nHugepagesize: 2048 kB\nDirectMap4k: 10232 kB\nDirectMap2M: 901120 kB\n">>}, {replication,[{"default",1.0}]}, {system_memory_data, [{system_total_memory,1050726400}, {free_swap,406839296}, {total_swap,406839296}, {cached_memory,459407360}, {buffered_memory,156426240}, {free_memory,267247616}, {total_memory,1050726400}]}, {statistics, [{wall_clock,{1130410,1}}, {context_switches,{212088,0}}, {garbage_collection,{45881,234716299,0}}, {io,{{input,12944592},{output,13318702}}}, {reductions,{179858313,1786867}}, {run_queue,0}, {runtime,{18410,180}}]}]}] INFO REPORT <6516.92.0> 2011-05-07 17:25:03 =============================================================================== ns_1@127.0.0.1:ns_doctor:82: Current node statuses: [{'ns_1@127.0.0.1', [{last_heard,{1304,814303,60248}}, {active_buckets,["default"]}, {memory, [{total,15380440}, {processes,4899672}, {processes_used,4888888}, {system,10480768}, {atom,518633}, {atom_used,496964}, {binary,151984}, {code,4014832}, {ets,895388}]}, {cluster_compatibility_version,1}, {version, [{kernel,"2.13.4"}, {mnesia,"4.4.12"}, {stdlib,"1.16.4"}, {os_mon,"2.2.4"}, {ns_server,"1.6.5.4r"}, {menelaus,"1.6.5.4r"}, {sasl,"2.1.8"}]}, {system_arch,"i686-pc-linux-gnu"}, {wall_clock,1200}, {memory_data,{1050726400,783757312,{<6516.268.0>,556860}}}, {disk_data, [{"/",7583436,18}, {"/dev",508460,1}, {"/dev/shm",513048,0}, {"/var/run",513048,1}, {"/var/lock",513048,0}, {"/lib/init/rw",513048,0}, {"/boot",233191,8}]}, {meminfo, <<"MemTotal: 1026100 kB\nMemFree: 260736 kB\nBuffers: 152764 kB\nCached: 448892 kB\nSwapCached: 0 kB\nActive: 449136 kB\nInactive: 277856 kB\nActive(anon): 121172 kB\nInactive(anon): 4412 kB\nActive(file): 327964 kB\nInactive(file): 273444 kB\nUnevictable: 0 kB\nMlocked: 0 kB\nHighTotal: 137160 kB\nHighFree: 252 kB\nLowTotal: 888940 kB\nLowFree: 260484 kB\nSwapTotal: 397304 kB\nSwapFree: 397304 kB\nDirty: 72 kB\nWriteback: 0 kB\nAnonPages: 125344 kB\nMapped: 10328 kB\nShmem: 240 kB\nSlab: 28112 kB\nSReclaimable: 23240 kB\nSUnreclaim: 4872 kB\nKernelStack: 960 kB\nPageTables: 1060 kB\nNFS_Unstable: 0 kB\nBounce: 0 kB\nWritebackTmp: 0 kB\nCommitLimit: 910352 kB\nCommitted_AS: 325100 kB\nVmallocTotal: 122880 kB\nVmallocUsed: 3540 kB\nVmallocChunk: 114424 kB\nHardwareCorrupted: 0 kB\nHugePages_Total: 0\nHugePages_Free: 0\nHugePages_Rsvd: 0\nHugePages_Surp: 0\nHugepagesize: 2048 kB\nDirectMap4k: 10232 kB\nDirectMap2M: 901120 kB\n">>}, {replication,[{"default",1.0}]}, {system_memory_data, [{system_total_memory,1050726400}, {free_swap,406839296}, {total_swap,406839296}, {cached_memory,459665408}, {buffered_memory,156430336}, {free_memory,266993664}, {total_memory,1050726400}]}, {statistics, [{wall_clock,{1190410,1}}, {context_switches,{222739,0}}, {garbage_collection,{48170,246629311,0}}, {io,{{input,12961074},{output,13606562}}}, {reductions,{190248539,1787282}}, {run_queue,0}, {runtime,{19410,180}}]}]}] INFO REPORT <6516.224.0> 2011-05-07 17:25:19 =============================================================================== ns_1@127.0.0.1:stats_collector:71: Stats for bucket "default": auth_cmds 4 auth_errors 0 bucket_conns 5 bytes_read 39260167 bytes_written 50767713 cas_badval 0 cas_hits 0 cas_misses 0 cmd_flush 0 cmd_get 204800 cmd_set 215040 conn_yields 0 connection_structures 15 curr_connections 15 curr_items 204800 curr_items_tot 204800 daemon_connections 10 decr_hits 0 decr_misses 0 delete_hits 0 delete_misses 0 ep_bg_fetched 0 ep_commit_num 457 ep_commit_time 0 ep_commit_time_total 5 ep_data_age 1 ep_data_age_highwat 5 ep_db_cleaner_status complete ep_db_strategy multiMTVBDB ep_dbinit 1 ep_dbname /var/opt/membase/1.6.5.4r/data/ns_1/default ep_dbshards 4 ep_expired 0 ep_flush_duration 0 ep_flush_duration_highwat 1 ep_flush_duration_total 6 ep_flush_preempts 0 ep_flusher_state running ep_flusher_todo 0 ep_io_num_read 0 ep_io_num_write 215040 ep_io_read_bytes 0 ep_io_write_bytes 24514560 ep_item_begin_failed 0 ep_item_commit_failed 0 ep_item_flush_expired 0 ep_item_flush_failed 0 ep_kv_size 35840000 ep_max_data_size 839909376 ep_max_txn_size 1000 ep_mem_high_wat 629932032 ep_mem_low_wat 503945625 ep_min_data_age 0 ep_num_active_non_resident 0 ep_num_eject_failures 0 ep_num_eject_replicas 0 ep_num_expiry_pager_runs 0 ep_num_non_resident 0 ep_num_not_my_vbuckets 0 ep_num_pager_runs 0 ep_num_value_ejects 0 ep_oom_errors 0 ep_overhead 12960032 ep_pending_ops 0 ep_pending_ops_max 0 ep_pending_ops_max_duration 0 ep_pending_ops_total 0 ep_queue_age_cap 900 ep_queue_size 0 ep_storage_age 0 ep_storage_age_highwat 2 ep_storage_type featured ep_store_max_concurrency 10 ep_store_max_readers 9 ep_store_max_readwrite 1 ep_tap_bg_fetch_requeued 0 ep_tap_bg_fetched 0 ep_tap_keepalive 300 ep_tmp_oom_errors 0 ep_too_old 0 ep_too_young 0 ep_total_cache_size 35840000 ep_total_del_items 0 ep_total_enqueued 215239 ep_total_new_items 204800 ep_total_persisted 215040 ep_vbucket_del 0 ep_vbucket_del_fail 0 ep_version 1.6.5.3.1_8_gce31baa ep_warmed_up 0 ep_warmup true ep_warmup_dups 0 ep_warmup_oom 0 ep_warmup_thread complete ep_warmup_time 29657 get_hits 204800 get_misses 0 incr_hits 0 incr_misses 0 libevent 2.0.7-rc limit_maxbytes 67108864 mem_used 48800032 pid 5769 pointer_size 32 rejected_conns 0 rusage_system 2.460153 rusage_user 4.328270 threads 4 time 1304814319 total_connections 15 uptime 1218 version 1.4.4_382_g9df3289 INFO REPORT <6516.92.0> 2011-05-07 17:26:03 =============================================================================== ns_1@127.0.0.1:ns_doctor:82: Current node statuses: [{'ns_1@127.0.0.1', [{last_heard,{1304,814363,60233}}, {active_buckets,["default"]}, {memory, [{total,15598752}, {processes,5145104}, {processes_used,5135360}, {system,10453648}, {atom,518633}, {atom_used,496964}, {binary,124096}, {code,4014832}, {ets,895380}]}, {cluster_compatibility_version,1}, {version, [{kernel,"2.13.4"}, {mnesia,"4.4.12"}, {stdlib,"1.16.4"}, {os_mon,"2.2.4"}, {ns_server,"1.6.5.4r"}, {menelaus,"1.6.5.4r"}, {sasl,"2.1.8"}]}, {system_arch,"i686-pc-linux-gnu"}, {wall_clock,1260}, {memory_data,{1050726400,784113664,{<6516.268.0>,556860}}}, {disk_data, [{"/",7583436,18}, {"/dev",508460,1}, {"/dev/shm",513048,0}, {"/var/run",513048,1}, {"/var/lock",513048,0}, {"/lib/init/rw",513048,0}, {"/boot",233191,8}]}, {meminfo, <<"MemTotal: 1026100 kB\nMemFree: 260488 kB\nBuffers: 152764 kB\nCached: 449144 kB\nSwapCached: 0 kB\nActive: 449260 kB\nInactive: 277996 kB\nActive(anon): 121184 kB\nInactive(anon): 4412 kB\nActive(file): 328076 kB\nInactive(file): 273584 kB\nUnevictable: 0 kB\nMlocked: 0 kB\nHighTotal: 137160 kB\nHighFree: 252 kB\nLowTotal: 888940 kB\nLowFree: 260236 kB\nSwapTotal: 397304 kB\nSwapFree: 397304 kB\nDirty: 80 kB\nWriteback: 0 kB\nAnonPages: 125356 kB\nMapped: 10328 kB\nShmem: 240 kB\nSlab: 28112 kB\nSReclaimable: 23240 kB\nSUnreclaim: 4872 kB\nKernelStack: 960 kB\nPageTables: 1060 kB\nNFS_Unstable: 0 kB\nBounce: 0 kB\nWritebackTmp: 0 kB\nCommitLimit: 910352 kB\nCommitted_AS: 325100 kB\nVmallocTotal: 122880 kB\nVmallocUsed: 3540 kB\nVmallocChunk: 114424 kB\nHardwareCorrupted: 0 kB\nHugePages_Total: 0\nHugePages_Free: 0\nHugePages_Rsvd: 0\nHugePages_Surp: 0\nHugepagesize: 2048 kB\nDirectMap4k: 10232 kB\nDirectMap2M: 901120 kB\n">>}, {replication,[{"default",1.0}]}, {system_memory_data, [{system_total_memory,1050726400}, {free_swap,406839296}, {total_swap,406839296}, {cached_memory,459923456}, {buffered_memory,156430336}, {free_memory,266739712}, {total_memory,1050726400}]}, {statistics, [{wall_clock,{1250410,1}}, {context_switches,{233194,0}}, {garbage_collection,{50366,257913276,0}}, {io,{{input,12975810},{output,13877191}}}, {reductions,{200541468,1787480}}, {run_queue,0}, {runtime,{20410,170}}]}]}] INFO REPORT <6516.224.0> 2011-05-07 17:26:59 =============================================================================== ns_1@127.0.0.1:stats_collector:71: Stats for bucket "default": auth_cmds 4 auth_errors 0 bucket_conns 5 bytes_read 39265223 bytes_written 52336339 cas_badval 0 cas_hits 0 cas_misses 0 cmd_flush 0 cmd_get 204800 cmd_set 215040 conn_yields 0 connection_structures 15 curr_connections 15 curr_items 204800 curr_items_tot 204800 daemon_connections 10 decr_hits 0 decr_misses 0 delete_hits 0 delete_misses 0 ep_bg_fetched 0 ep_commit_num 457 ep_commit_time 0 ep_commit_time_total 5 ep_data_age 1 ep_data_age_highwat 5 ep_db_cleaner_status complete ep_db_strategy multiMTVBDB ep_dbinit 1 ep_dbname /var/opt/membase/1.6.5.4r/data/ns_1/default ep_dbshards 4 ep_expired 0 ep_flush_duration 0 ep_flush_duration_highwat 1 ep_flush_duration_total 6 ep_flush_preempts 0 ep_flusher_state running ep_flusher_todo 0 ep_io_num_read 0 ep_io_num_write 215040 ep_io_read_bytes 0 ep_io_write_bytes 24514560 ep_item_begin_failed 0 ep_item_commit_failed 0 ep_item_flush_expired 0 ep_item_flush_failed 0 ep_kv_size 35840000 ep_max_data_size 839909376 ep_max_txn_size 1000 ep_mem_high_wat 629932032 ep_mem_low_wat 503945625 ep_min_data_age 0 ep_num_active_non_resident 0 ep_num_eject_failures 0 ep_num_eject_replicas 0 ep_num_expiry_pager_runs 0 ep_num_non_resident 0 ep_num_not_my_vbuckets 0 ep_num_pager_runs 0 ep_num_value_ejects 0 ep_oom_errors 0 ep_overhead 12960032 ep_pending_ops 0 ep_pending_ops_max 0 ep_pending_ops_max_duration 0 ep_pending_ops_total 0 ep_queue_age_cap 900 ep_queue_size 0 ep_storage_age 0 ep_storage_age_highwat 2 ep_storage_type featured ep_store_max_concurrency 10 ep_store_max_readers 9 ep_store_max_readwrite 1 ep_tap_bg_fetch_requeued 0 ep_tap_bg_fetched 0 ep_tap_keepalive 300 ep_tmp_oom_errors 0 ep_too_old 0 ep_too_young 0 ep_total_cache_size 35840000 ep_total_del_items 0 ep_total_enqueued 215239 ep_total_new_items 204800 ep_total_persisted 215040 ep_vbucket_del 0 ep_vbucket_del_fail 0 ep_version 1.6.5.3.1_8_gce31baa ep_warmed_up 0 ep_warmup true ep_warmup_dups 0 ep_warmup_oom 0 ep_warmup_thread complete ep_warmup_time 29657 get_hits 204800 get_misses 0 incr_hits 0 incr_misses 0 libevent 2.0.7-rc limit_maxbytes 67108864 mem_used 48800032 pid 5769 pointer_size 32 rejected_conns 0 rusage_system 2.464154 rusage_user 4.380273 threads 4 time 1304814419 total_connections 15 uptime 1318 version 1.4.4_382_g9df3289 INFO REPORT <6516.92.0> 2011-05-07 17:27:03 =============================================================================== ns_1@127.0.0.1:ns_doctor:82: Current node statuses: [{'ns_1@127.0.0.1', [{last_heard,{1304,814423,60250}}, {active_buckets,["default"]}, {memory, [{total,15475336}, {processes,4940892}, {processes_used,4930668}, {system,10534444}, {atom,518633}, {atom_used,496964}, {binary,169352}, {code,4014832}, {ets,930084}]}, {cluster_compatibility_version,1}, {version, [{kernel,"2.13.4"}, {mnesia,"4.4.12"}, {stdlib,"1.16.4"}, {os_mon,"2.2.4"}, {ns_server,"1.6.5.4r"}, {menelaus,"1.6.5.4r"}, {sasl,"2.1.8"}]}, {system_arch,"i686-pc-linux-gnu"}, {wall_clock,1320}, {memory_data,{1050726400,784392192,{<6516.268.0>,556860}}}, {disk_data, [{"/",7583436,18}, {"/dev",508460,1}, {"/dev/shm",513048,0}, {"/var/run",513048,1}, {"/var/lock",513048,0}, {"/lib/init/rw",513048,0}, {"/boot",233191,8}]}, {meminfo, <<"MemTotal: 1026100 kB\nMemFree: 260240 kB\nBuffers: 152764 kB\nCached: 449368 kB\nSwapCached: 0 kB\nActive: 449184 kB\nInactive: 278328 kB\nActive(anon): 121208 kB\nInactive(anon): 4412 kB\nActive(file): 327976 kB\nInactive(file): 273916 kB\nUnevictable: 0 kB\nMlocked: 0 kB\nHighTotal: 137160 kB\nHighFree: 252 kB\nLowTotal: 888940 kB\nLowFree: 259988 kB\nSwapTotal: 397304 kB\nSwapFree: 397304 kB\nDirty: 84 kB\nWriteback: 0 kB\nAnonPages: 125384 kB\nMapped: 10328 kB\nShmem: 240 kB\nSlab: 28116 kB\nSReclaimable: 23244 kB\nSUnreclaim: 4872 kB\nKernelStack: 960 kB\nPageTables: 1060 kB\nNFS_Unstable: 0 kB\nBounce: 0 kB\nWritebackTmp: 0 kB\nCommitLimit: 910352 kB\nCommitted_AS: 325100 kB\nVmallocTotal: 122880 kB\nVmallocUsed: 3540 kB\nVmallocChunk: 114424 kB\nHardwareCorrupted: 0 kB\nHugePages_Total: 0\nHugePages_Free: 0\nHugePages_Rsvd: 0\nHugePages_Surp: 0\nHugepagesize: 2048 kB\nDirectMap4k: 10232 kB\nDirectMap2M: 901120 kB\n">>}, {replication,[{"default",1.0}]}, {system_memory_data, [{system_total_memory,1050726400}, {free_swap,406839296}, {total_swap,406839296}, {cached_memory,460152832}, {buffered_memory,156430336}, {free_memory,266485760}, {total_memory,1050726400}]}, {statistics, [{wall_clock,{1310410,1}}, {context_switches,{244793,0}}, {garbage_collection,{52898,271269093,0}}, {io,{{input,13749236},{output,14882139}}}, {reductions,{211202786,1791473}}, {run_queue,0}, {runtime,{21460,180}}]}]}] INFO REPORT <6516.62.0> 2011-05-07 17:27:43 =============================================================================== ns_1@127.0.0.1:ns_cluster:90: handling add_node("10.1.4.244", 8091, ..) INFO REPORT <6516.62.0> 2011-05-07 17:27:43 =============================================================================== ns_1@127.0.0.1:ns_cluster:247: Decided to change address to "10.1.4.243" INFO REPORT <6516.110.0> 2011-05-07 17:27:43 =============================================================================== ns_1@127.0.0.1:ns_mnesia:144: Info from Mnesia: {mnesia_checkpoint,{{1304,814463,178212},'ns_1@127.0.0.1'}} starting: <6516.2117.0> INFO REPORT <6516.110.0> 2011-05-07 17:27:43 =============================================================================== ns_1@127.0.0.1:ns_mnesia:144: Info from Mnesia: {mnesia_checkpoint,{{1304,814463,178212},'ns_1@127.0.0.1'}} terminated: shutdown INFO REPORT <6516.62.0> 2011-05-07 17:27:43 =============================================================================== ns_1@127.0.0.1:ns_cluster:249: prepared mnesia. INFO REPORT <6516.62.0> 2011-05-07 17:27:43 =============================================================================== ns_server_sup plug pulled. Killing [ns_mnesia,ns_bad_bucket_worker, ns_bad_bucket_sup,ns_moxi_sup,ns_tick], keeping [ns_log, ns_log_events, ns_mail_sup, ns_node_disco_sup, ns_heart, ns_doctor, menelaus, ns_port_sup, ns_tick_event, ns_stats_event, ns_good_bucket_worker, ns_good_bucket_sup, ns_orchestrator] INFO REPORT <6516.7.0> 2011-05-07 17:27:43 =============================================================================== application mnesia exited stopped type temporary INFO REPORT <6516.110.0> 2011-05-07 17:27:43 =============================================================================== ns_1@127.0.0.1:ns_mnesia:202: Shut Mnesia down: shutdown. Exiting. INFO REPORT <3.78.0> 2011-05-07 17:27:43 =============================================================================== ns_log: logging ns_node_disco:5:Node nonode@nohost saw that node 'ns_1@127.0.0.1' went down. INFO REPORT <3.55.0> 2011-05-07 17:27:43 =============================================================================== Adjusted IP to "10.1.4.243" PROGRESS REPORT <3.2125.0> 2011-05-07 17:27:43 =============================================================================== supervisor {local,net_sup} started [{pid,<3.2126.0>}, {name,erl_epmd}, {mfa,{erl_epmd,start_link,[]}}, {restart_type,permanent}, {shutdown,2000}, {child_type,worker}] PROGRESS REPORT <3.2125.0> 2011-05-07 17:27:43 =============================================================================== supervisor {local,net_sup} started [{pid,<3.2127.0>}, {name,auth}, {mfa,{auth,start_link,[]}}, {restart_type,permanent}, {shutdown,2000}, {child_type,worker}] INFO REPORT <0.78.0> 2011-05-07 17:27:43 =============================================================================== ns_log: logging ns_node_disco:4:Node 'ns_1@10.1.4.243' saw that node 'ns_1@10.1.4.243' came up. PROGRESS REPORT <0.2125.0> 2011-05-07 17:27:43 =============================================================================== supervisor {local,net_sup} started [{pid,<0.2128.0>}, {name,net_kernel}, {mfa,{net_kernel,start_link,[['ns_1@10.1.4.243',longnames]]}}, {restart_type,permanent}, {shutdown,2000}, {child_type,worker}] PROGRESS REPORT <0.11.0> 2011-05-07 17:27:43 =============================================================================== supervisor {local,kernel_sup} started [{pid,<0.2125.0>}, {name,net_sup_dynamic}, {mfa,{erl_distribution,start_link,[['ns_1@10.1.4.243',longnames]]}}, {restart_type,permanent}, {shutdown,1000}, {child_type,supervisor}] INFO REPORT <0.55.0> 2011-05-07 17:27:43 =============================================================================== Re-setting cookie {elyrdmyctqykhepq,'ns_1@10.1.4.243'} INFO REPORT <0.55.0> 2011-05-07 17:27:43 =============================================================================== saving ip config to "/etc/opt/membase/1.6.5.4r/ip" INFO REPORT <0.55.0> 2011-05-07 17:27:43 =============================================================================== save_address_config: ok INFO REPORT <0.62.0> 2011-05-07 17:27:43 =============================================================================== ns_1@10.1.4.243:ns_mnesia:69: Renaming node from 'ns_1@127.0.0.1' to 'ns_1@10.1.4.243'. INFO REPORT <0.62.0> 2011-05-07 17:27:43 =============================================================================== ns_1@10.1.4.243:ns_mnesia:73: Deleting old schema. INFO REPORT <0.62.0> 2011-05-07 17:27:43 =============================================================================== ns_1@10.1.4.243:ns_mnesia:75: Installing new backup as fallback. ERROR REPORT <0.6.0> 2011-05-07 17:27:43 =============================================================================== Truncated log event: {info_msg,<0.52.0>, {<0.65.0>,"renaming node conf ~p -> ~p:~n ~p ->~n ~p~n", [buckets,buckets, [{configs,[{"default", [{num_replicas,1}, {ram_quota,839909376}, {auth_type,sasl}, {sasl_password,[]}, {type,membase}, {num_vbuckets,1024}, {db_shards,4}, {ht_size,3079}, {tap_keepalive,0}, {tap_noop_interval,20}, {max_txn_size,1000}, {ht_locks,5}, {servers,['ns_1@127.0.0.1']}, {map,[['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1',undefined], ['ns_1@127.0.0.1'|...], [...]|...]}]}]}], [{configs,[{"default", [{num_replicas,1}, {ram_quota,839909376}, {auth_type,sasl}, {sasl_password,[]}, {type,membase}, {num_vbuckets,1024}, {db_shards,4}, {ht_size,3079}, {tap_keepalive,0}, {tap_noop_interval,20}, {max_txn_size,1000}, {ht_locks,5}, {servers,['ns_1@10.1.4.243']}, {map,[['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243'|...], [...]|...]}]}]}]]}} INFO REPORT <0.65.0> 2011-05-07 17:27:43 =============================================================================== renaming node conf {node,'ns_1@127.0.0.1',memcached} -> {node, 'ns_1@10.1.4.243', memcached}: [{dbdir,"/var/opt/membase/1.6.5.4r/data/ns_1"}, {port,11210}, {admin_user,"_admin"}, {admin_pass,"_admin"}, {bucket_engine,"./bin/bucket_engine/bucket_engine.so"}, {engines,[{membase,[{engine,"bin/ep_engine/ep.so"}, {initfile,"/etc/opt/membase/1.6.5.4r/init.sql"}]}, {memcached,[{engine,"bin/memcached/default_engine.so"}]}]}, {verbosity,[]}] -> [{dbdir,"/var/opt/membase/1.6.5.4r/data/ns_1"}, {port,11210}, {admin_user,"_admin"}, {admin_pass,"_admin"}, {bucket_engine,"./bin/bucket_engine/bucket_engine.so"}, {engines,[{membase,[{engine,"bin/ep_engine/ep.so"}, {initfile,"/etc/opt/membase/1.6.5.4r/init.sql"}]}, {memcached,[{engine,"bin/memcached/default_engine.so"}]}]}, {verbosity,[]}] INFO REPORT <0.65.0> 2011-05-07 17:27:43 =============================================================================== renaming node conf nodes_wanted -> nodes_wanted: ['ns_1@127.0.0.1'] -> ['ns_1@10.1.4.243'] INFO REPORT <0.65.0> 2011-05-07 17:27:43 =============================================================================== renaming node conf {node,'ns_1@127.0.0.1',isasl} -> {node,'ns_1@10.1.4.243', isasl}: [{path,"/var/opt/membase/1.6.5.4r/data/ns_1/isasl.pw"}] -> [{path,"/var/opt/membase/1.6.5.4r/data/ns_1/isasl.pw"}] INFO REPORT <0.65.0> 2011-05-07 17:27:43 =============================================================================== renaming node conf {node,'ns_1@127.0.0.1',membership} -> {node, 'ns_1@10.1.4.243', membership}: active -> active INFO REPORT <0.65.0> 2011-05-07 17:27:43 =============================================================================== renaming node conf {node,'ns_1@127.0.0.1',ns_log} -> {node,'ns_1@10.1.4.243', ns_log}: [{filename,"/var/opt/membase/1.6.5.4r/data/ns_1/ns_log"}] -> [{filename,"/var/opt/membase/1.6.5.4r/data/ns_1/ns_log"}] INFO REPORT <0.62.0> 2011-05-07 17:27:43 =============================================================================== ns_1@10.1.4.243:ns_cluster:260: Renamed node. New name is 'ns_1@10.1.4.243'. INFO REPORT <0.64.0> 2011-05-07 17:27:43 =============================================================================== config change: {node,'ns_1@10.1.4.243',ns_log} -> [{filename,"/var/opt/membase/1.6.5.4r/data/ns_1/ns_log"}] PROGRESS REPORT <0.2149.0> 2011-05-07 17:27:43 =============================================================================== supervisor {local,mnesia_sup} started [{pid,<0.2150.0>}, {name,mnesia_event}, {mfa,{mnesia_sup,start_event,[]}}, {restart_type,permanent}, {shutdown,30000}, {child_type,worker}] INFO REPORT <0.64.0> 2011-05-07 17:27:43 =============================================================================== config change: {node,'ns_1@10.1.4.243',membership} -> active INFO REPORT <0.64.0> 2011-05-07 17:27:43 =============================================================================== config change: {node,'ns_1@10.1.4.243',isasl} -> [{path,"/var/opt/membase/1.6.5.4r/data/ns_1/isasl.pw"}] INFO REPORT <0.64.0> 2011-05-07 17:27:43 =============================================================================== ns_node_disco_conf_events config on nodes_wanted INFO REPORT <0.64.0> 2011-05-07 17:27:43 =============================================================================== config change: nodes_wanted -> ['ns_1@10.1.4.243'] INFO REPORT <0.2153.0> 2011-05-07 17:27:43 =============================================================================== ns_node_disco cookie_sync INFO REPORT <0.2153.0> 2011-05-07 17:27:43 =============================================================================== ns_node_disco: nodes_wanted updated: ['ns_1@10.1.4.243'], with cookie: elyrdmyctqykhepq PROGRESS REPORT <0.2151.0> 2011-05-07 17:27:43 =============================================================================== supervisor {local,mnesia_kernel_sup} started [{pid,<0.2152.0>}, {name,mnesia_monitor}, {mfa,{mnesia_monitor,start,[]}}, {restart_type,permanent}, {shutdown,3000}, {child_type,worker}] INFO REPORT <0.2153.0> 2011-05-07 17:27:43 =============================================================================== ns_node_disco: nodes_wanted pong: ['ns_1@10.1.4.243'], with cookie: elyrdmyctqykhepq INFO REPORT <0.64.0> 2011-05-07 17:27:43 =============================================================================== config change: {node,'ns_1@10.1.4.243',memcached} -> [{dbdir,"/var/opt/membase/1.6.5.4r/data/ns_1"}, {port,11210}, {admin_user,"_admin"}, {admin_pass,"_admin"}, {bucket_engine,"./bin/bucket_engine/bucket_engine.so"}, {engines,[{membase,[{engine,"bin/ep_engine/ep.so"}, {initfile,"/etc/opt/membase/1.6.5.4r/init.sql"}]}, {memcached,[{engine,"bin/memcached/default_engine.so"}]}]}, {verbosity,[]}] PROGRESS REPORT <0.2151.0> 2011-05-07 17:27:43 =============================================================================== supervisor {local,mnesia_kernel_sup} started [{pid,<0.2154.0>}, {name,mnesia_subscr}, {mfa,{mnesia_subscr,start,[]}}, {restart_type,permanent}, {shutdown,3000}, {child_type,worker}] PROGRESS REPORT <0.2151.0> 2011-05-07 17:27:43 =============================================================================== supervisor {local,mnesia_kernel_sup} started [{pid,<0.2155.0>}, {name,mnesia_locker}, {mfa,{mnesia_locker,start,[]}}, {restart_type,permanent}, {shutdown,3000}, {child_type,worker}] PROGRESS REPORT <0.2151.0> 2011-05-07 17:27:43 =============================================================================== supervisor {local,mnesia_kernel_sup} started [{pid,<0.2156.0>}, {name,mnesia_recover}, {mfa,{mnesia_recover,start,[]}}, {restart_type,permanent}, {shutdown,180000}, {child_type,worker}] INFO REPORT <0.64.0> 2011-05-07 17:27:43 =============================================================================== config change: buckets -> [{configs,[{"default", [{num_replicas,1}, {ram_quota,839909376}, {auth_type,sasl}, {sasl_password,[]}, {type,membase}, {num_vbuckets,1024}, {db_shards,4}, {ht_size,3079}, {tap_keepalive,0}, {tap_noop_interval,20}, {max_txn_size,1000}, {ht_locks,5}, {servers,['ns_1@10.1.4.243']}, {map,[['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243'|...], [...]|...]}]}]}] INFO REPORT <0.64.0> 2011-05-07 17:27:43 =============================================================================== ns_node_disco_conf_events config all INFO REPORT <0.85.0> 2011-05-07 17:27:43 =============================================================================== Pushing config INFO REPORT <0.85.0> 2011-05-07 17:27:43 =============================================================================== Pushing config done PROGRESS REPORT <0.2151.0> 2011-05-07 17:27:43 =============================================================================== supervisor {local,mnesia_kernel_sup} started [{pid,<0.2157.0>}, {name,mnesia_tm}, {mfa,{mnesia_tm,start,[]}}, {restart_type,permanent}, {shutdown,30000}, {child_type,worker}] PROGRESS REPORT <0.2151.0> 2011-05-07 17:27:43 =============================================================================== supervisor {local,mnesia_kernel_sup} started [{pid,<0.2193.0>}, {name,mnesia_checkpoint_sup}, {mfa,{mnesia_checkpoint_sup,start,[]}}, {restart_type,permanent}, {shutdown,infinity}, {child_type,supervisor}] PROGRESS REPORT <0.2151.0> 2011-05-07 17:27:43 =============================================================================== supervisor {local,mnesia_kernel_sup} started [{pid,<0.2194.0>}, {name,mnesia_snmp_sup}, {mfa,{mnesia_snmp_sup,start,[]}}, {restart_type,permanent}, {shutdown,infinity}, {child_type,supervisor}] PROGRESS REPORT <0.2151.0> 2011-05-07 17:27:43 =============================================================================== supervisor {local,mnesia_kernel_sup} started [{pid,<0.2195.0>}, {name,mnesia_controller}, {mfa,{mnesia_controller,start,[]}}, {restart_type,permanent}, {shutdown,3000}, {child_type,worker}] PROGRESS REPORT <0.2151.0> 2011-05-07 17:27:43 =============================================================================== supervisor {local,mnesia_kernel_sup} started [{pid,<0.2196.0>}, {name,mnesia_late_loader}, {mfa,{mnesia_late_loader,start,[]}}, {restart_type,permanent}, {shutdown,3000}, {child_type,worker}] PROGRESS REPORT <0.2149.0> 2011-05-07 17:27:43 =============================================================================== supervisor {local,mnesia_sup} started [{pid,<0.2151.0>}, {name,mnesia_kernel_sup}, {mfa,{mnesia_kernel_sup,start,[]}}, {restart_type,permanent}, {shutdown,infinity}, {child_type,supervisor}] PROGRESS REPORT <0.7.0> 2011-05-07 17:27:43 =============================================================================== application mnesia started_at 'ns_1@10.1.4.243' INFO REPORT <0.2145.0> 2011-05-07 17:27:43 =============================================================================== ns_1@10.1.4.243:ns_mnesia:258: Using existing disk schema on ['ns_1@10.1.4.243']. INFO REPORT <0.2145.0> 2011-05-07 17:27:43 =============================================================================== ns_1@10.1.4.243:ns_mnesia:196: Current config: [{access_module,mnesia}, {auto_repair,true}, {backup_module,mnesia_backup}, {checkpoints,[]}, {db_nodes,['ns_1@10.1.4.243']}, {debug,verbose}, {directory, "/var/opt/membase/1.6.5.4r/mnesia"}, {dump_log_load_regulation, false}, {dump_log_time_threshold, 180000}, {dump_log_update_in_place, true}, {dump_log_write_threshold, 1000}, {embedded_mnemosyne,false}, {event_module,mnesia_event}, {extra_db_nodes,[]}, {fallback_activated,false}, {held_locks,[]}, {ignore_fallback_at_startup, false}, {fallback_error_function, {mnesia,lkill}}, {is_running,yes}, {local_tables, ['stats_archiver-default-day', 'stats_archiver-default-month', 'stats_archiver-default-week', 'stats_archiver-default-year', schema, 'stats_archiver-default-minute', 'stats_archiver-default-hour']}, {lock_queue,[]}, {log_version,"4.3"}, {master_node_tables,[]}, {max_wait_for_decision,10000}, {protocol_version,{7,6}}, {running_db_nodes, ['ns_1@10.1.4.243']}, {schema_location,opt_disc}, {schema_version,{3,0}}, {subscribers, [<0.2150.0>,<0.2145.0>]}, {tables, ['stats_archiver-default-day', 'stats_archiver-default-month', 'stats_archiver-default-week', 'stats_archiver-default-year', schema, 'stats_archiver-default-minute', 'stats_archiver-default-hour']}, {transaction_commits,2}, {transaction_failures,0}, {transaction_log_writes,0}, {transaction_restarts,0}, {transactions,[]}, {use_dir,true}, {core_dir,false}, {no_table_loaders,2}, {dc_dump_limit,4}, {version,"4.4.12"}] PROGRESS REPORT <0.70.0> 2011-05-07 17:27:43 =============================================================================== supervisor {local,ns_server_sup} started [{pid,<0.2145.0>}, {name,ns_mnesia}, {mfa,{ns_mnesia,start_link,[]}}, {restart_type,permanent}, {shutdown,10000}, {child_type,worker}] INFO REPORT <0.62.0> 2011-05-07 17:27:43 =============================================================================== Restarting ns_mnesia: {ok,<0.2145.0>} PROGRESS REPORT <0.70.0> 2011-05-07 17:27:43 =============================================================================== supervisor {local,ns_server_sup} started [{pid,<0.2202.0>}, {name,ns_bad_bucket_worker}, {mfa,{work_queue,start_link,[ns_bad_bucket_worker]}}, {restart_type,permanent}, {shutdown,10}, {child_type,worker}] INFO REPORT <0.62.0> 2011-05-07 17:27:43 =============================================================================== Restarting ns_bad_bucket_worker: {ok,<0.2202.0>} PROGRESS REPORT <0.2203.0> 2011-05-07 17:27:43 =============================================================================== supervisor {local,ns_bad_bucket_sup} started [{pid,<0.2204.0>}, {name,{stats_collector,"default"}}, {mfa,{stats_collector,start_link,["default"]}}, {restart_type,permanent}, {shutdown,10}, {child_type,worker}] PROGRESS REPORT <0.2203.0> 2011-05-07 17:27:43 =============================================================================== supervisor {local,ns_bad_bucket_sup} started [{pid,<0.2205.0>}, {name,{stats_archiver,"default"}}, {mfa,{stats_archiver,start_link,["default"]}}, {restart_type,permanent}, {shutdown,10}, {child_type,worker}] PROGRESS REPORT <0.2203.0> 2011-05-07 17:27:43 =============================================================================== supervisor {local,ns_bad_bucket_sup} started [{pid,<0.2208.0>}, {name,{stats_reader,"default"}}, {mfa,{stats_reader,start_link,["default"]}}, {restart_type,permanent}, {shutdown,10}, {child_type,worker}] PROGRESS REPORT <0.70.0> 2011-05-07 17:27:43 =============================================================================== supervisor {local,ns_server_sup} started [{pid,<0.2203.0>}, {name,ns_bad_bucket_sup}, {mfa,{ns_bucket_sup,start_link, [ns_bad_bucket_sup, #Fun, ns_bad_bucket_worker]}}, {restart_type,permanent}, {shutdown,infinity}, {child_type,supervisor}] INFO REPORT <0.62.0> 2011-05-07 17:27:43 =============================================================================== Restarting ns_bad_bucket_sup: {ok,<0.2203.0>} PROGRESS REPORT <0.70.0> 2011-05-07 17:27:43 =============================================================================== supervisor {local,ns_server_sup} started [{pid,<0.2209.0>}, {name,ns_moxi_sup}, {mfa,{ns_moxi_sup,start_link,[]}}, {restart_type,permanent}, {shutdown,infinity}, {child_type,supervisor}] INFO REPORT <0.62.0> 2011-05-07 17:27:43 =============================================================================== Restarting ns_moxi_sup: {ok,<0.2209.0>} INFO REPORT <0.70.0> 2011-05-07 17:27:43 =============================================================================== ns_1@10.1.4.243:misc:725: start_singleton(gen_server, ns_tick, [], []): started as <0.2210.0> on 'ns_1@10.1.4.243' PROGRESS REPORT <0.70.0> 2011-05-07 17:27:43 =============================================================================== supervisor {local,ns_server_sup} started [{pid,<0.2210.0>}, {name,ns_tick}, {mfa,{ns_tick,start_link,[]}}, {restart_type,permanent}, {shutdown,10}, {child_type,worker}] INFO REPORT <0.62.0> 2011-05-07 17:27:43 =============================================================================== Restarting ns_tick: {ok,<0.2210.0>} PROGRESS REPORT <0.70.0> 2011-05-07 17:27:43 =============================================================================== supervisor {local,ns_server_sup} started [{pid,<0.2211.0>}, {name,ns_doctor}, {mfa,{ns_doctor,start_link,[]}}, {restart_type,permanent}, {shutdown,10}, {child_type,worker}] INFO REPORT <0.62.0> 2011-05-07 17:27:43 =============================================================================== ns_1@10.1.4.243:ns_server_sup:140: Restarted grey child ns_doctor: {ok, <0.2211.0>} INFO REPORT <0.62.0> 2011-05-07 17:27:43 =============================================================================== ns_1@10.1.4.243:ns_cluster:264: Started ns_server_sup childs back. INFO REPORT <0.2211.0> 2011-05-07 17:27:43 =============================================================================== ns_1@10.1.4.243:ns_doctor:78: Got initial status [{'ns_1@10.1.4.243', [{last_heard, {1304,814463,512101}}, {active_buckets, ["default"]}, {memory, [{total,14654952}, {processes,4828360}, {processes_used,4819688}, {system,9826592}, {atom,519117}, {atom_used,497147}, {binary,81472}, {code,4019423}, {ets,297484}]}, {cluster_compatibility_version, 1}, {version, [{kernel,"2.13.4"}, {mnesia,"4.4.12"}, {stdlib,"1.16.4"}, {os_mon,"2.2.4"}, {ns_server,"1.6.5.4r"}, {menelaus,"1.6.5.4r"}, {sasl,"2.1.8"}]}, {system_arch, "i686-pc-linux-gnu"}, {wall_clock,1360}, {memory_data, {1050726400,784392192, {<0.268.0>,556860}}}, {disk_data, [{"/",7583436,18}, {"/dev",508460,1}, {"/dev/shm",513048,0}, {"/var/run",513048,1}, {"/var/lock",513048,0}, {"/lib/init/rw",513048, 0}, {"/boot",233191,8}]}, {meminfo, <<"MemTotal: 1026100 kB\nMemFree: 259992 kB\nBuffers: 152764 kB\nCached: 449588 kB\nSwapCached: 0 kB\nActive: 449296 kB\nInactive: 278444 kB\nActive(anon): 121216 kB\nInactive(anon): 4412 kB\nActive(file): 328080 kB\nInactive(file): 274032 kB\nUnevictable: 0 kB\nMlocked: 0 kB\nHighTotal: 137160 kB\nHighFree: 252 kB\nLowTotal: 888940 kB\nLowFree: 259740 kB\nSwapTotal: 397304 kB\nSwapFree: 397304 kB\nDirty: 44 kB\nWriteback: 0 kB\nAnonPages: 125388 kB\nMapped: 10328 kB\nShmem: 240 kB\nSlab: 28116 kB\nSReclaimable: 23244 kB\nSUnreclaim: 4872 kB\nKernelStack: 960 kB\nPageTables: 1060 kB\nNFS_Unstable: 0 kB\nBounce: 0 kB\nWritebackTmp: 0 kB\nCommitLimit: 910352 kB\nCommitted_AS: 325100 kB\nVmallocTotal: 122880 kB\nVmallocUsed: 3540 kB\nVmallocChunk: 114424 kB\nHardwareCorrupted: 0 kB\nHugePages_Total: 0\nHugePages_Free: 0\nHugePages_Rsvd: 0\nHugePages_Surp: 0\nHugepagesize: 2048 kB\nDirectMap4k: 10232 kB\nDirectMap2M: 901120 kB\n">>}, {replication, [{"default",1.0}]}, {system_memory_data, [{system_total_memory, 1050726400}, {free_swap,406839296}, {total_swap,406839296}, {cached_memory, 460378112}, {buffered_memory, 156430336}, {free_memory,266231808}, {total_memory, 1050726400}]}, {statistics, [{wall_clock,{1360410,1}}, {context_switches, {253819,0}}, {garbage_collection, {54700,281477609,0}}, {io, {{input,13763331}, {output,15112129}}}, {reductions, {219902178,1711861}}, {run_queue,0}, {runtime, {22310,170}}]}]}] ERROR REPORT <0.2208.0> 2011-05-07 17:27:43 =============================================================================== ** Generic server 'stats_reader-default' terminating ** Last message in was {latest,minute,1} ** When Server state == {state,"default"} ** Reason for termination == ** {aborted,{no_exists,['stats_archiver-default-minute']}} CRASH REPORT <0.2208.0> 2011-05-07 17:27:43 =============================================================================== Crashing process initial_call {stats_reader,init,['Argument__1']} pid <0.2208.0> registered_name 'stats_reader-default' error_info {exit,{aborted,{no_exists,['stats_archiver-default-minute']}}, [{gen_server,terminate,6},{proc_lib,init_p_do_apply,3}]} ancestors [ns_bad_bucket_sup,ns_server_sup,ns_server_cluster_sup,<0.53.0>] messages [] links [<0.2203.0>] dictionary [] trap_exit false status running heap_size 377 stack_size 24 reductions 288 SUPERVISOR REPORT <0.2203.0> 2011-05-07 17:27:43 =============================================================================== Reporting supervisor {local,ns_bad_bucket_sup} Child process errorContext child_terminated reason {aborted,{no_exists,['stats_archiver-default-minute']}} pid <0.2208.0> name {stats_reader,"default"} start_function {stats_reader,start_link,["default"]} restart_type permanent shutdown 10 child_type worker ERROR REPORT <0.62.0> 2011-05-07 17:27:43 =============================================================================== ns_1@10.1.4.243:stats_reader:176: Some nodes didn't respond: ['ns_1@10.1.4.243'] PROGRESS REPORT <0.2203.0> 2011-05-07 17:27:43 =============================================================================== supervisor {local,ns_bad_bucket_sup} started [{pid,<0.2218.0>}, {name,{stats_reader,"default"}}, {mfa,{stats_reader,start_link,["default"]}}, {restart_type,permanent}, {shutdown,10}, {child_type,worker}] INFO REPORT <0.62.0> 2011-05-07 17:27:43 =============================================================================== ns_1@10.1.4.243:ns_cluster:300: Posting node info to engage_cluster on {"10.1.4.244", 8091}: {struct, [{availableStorage, {struct, [{hdd, [{struct, [{path,<<"/">>}, {sizeKBytes,7583436}, {usagePercent,18}]}, {struct, [{path,<<"/dev">>}, {sizeKBytes,508460}, {usagePercent,1}]}, {struct, [{path,<<"/dev/shm">>}, {sizeKBytes,513048}, {usagePercent,0}]}, {struct, [{path,<<"/var/run">>}, {sizeKBytes,513048}, {usagePercent,1}]}, {struct, [{path,<<"/var/lock">>}, {sizeKBytes,513048}, {usagePercent,0}]}, {struct, [{path,<<"/lib/init/rw">>}, {sizeKBytes,513048}, {usagePercent,0}]}, {struct, [{path,<<"/boot">>}, {sizeKBytes,233191}, {usagePercent,8}]}]}]}}, {memoryQuota,801}, {storageTotals, {struct, [{ram, {struct, [{usedByData,0}, {total,1050726400}, {quotaTotal,839909376}, {used,784392192}]}}, {hdd, {struct, [{usedByData,28868608}, {total,7765438464}, {quotaTotal,7765438464}, {used,1397778923}, {free,6367659541}]}}]}}, {storage, {struct, [{ssd,[]}, {hdd, [{struct, [{path,<<"/var/opt/membase/1.6.5.4r/data/ns_1">>}, {quotaMb,none}, {state,ok}]}]}]}}, {uptime,<<"1360">>}, {memoryTotal,1050726400}, {memoryFree,266334208}, {mcdMemoryReserved,801}, {mcdMemoryAllocated,801}, {otpNode,<<"ns_1@10.1.4.243">>}, {otpCookie,<<"elyrdmyctqykhepq">>}, {clusterMembership,<<"active">>}, {status,<<"healthy">>}, {hostname,<<"10.1.4.243:8091">>}, {clusterCompatibility,1}, {version,<<"1.6.5.4r">>}, {os,<<"i686-pc-linux-gnu">>}, {ports,{struct,[{proxy,11211},{direct,11210}]}}]} PROGRESS REPORT <0.2254.0> 2011-05-07 17:27:43 =============================================================================== supervisor {local,inets_sup} started [{pid,<0.2255.0>}, {name,ftp_sup}, {mfa,{ftp_sup,start_link,[]}}, {restart_type,permanent}, {shutdown,infinity}, {child_type,supervisor}] PROGRESS REPORT <0.2260.0> 2011-05-07 17:27:43 =============================================================================== supervisor {local,httpc_profile_sup} started [{pid,<0.2267.0>}, {name,httpc_manager}, {mfa,{httpc_manager,start_link,[{default,only_session_cookies}]}}, {restart_type,permanent}, {shutdown,4000}, {child_type,worker}] PROGRESS REPORT <0.2256.0> 2011-05-07 17:27:43 =============================================================================== supervisor {local,httpc_sup} started [{pid,<0.2260.0>}, {name,httpc_profile_sup}, {mfa,{httpc_profile_sup,start_link, [[{httpc,{default,only_session_cookies}}]]}}, {restart_type,permanent}, {shutdown,infinity}, {child_type,supervisor}] PROGRESS REPORT <0.2256.0> 2011-05-07 17:27:43 =============================================================================== supervisor {local,httpc_sup} started [{pid,<0.2271.0>}, {name,httpc_handler_sup}, {mfa,{httpc_handler_sup,start_link,[]}}, {restart_type,permanent}, {shutdown,infinity}, {child_type,supervisor}] PROGRESS REPORT <0.2254.0> 2011-05-07 17:27:43 =============================================================================== supervisor {local,inets_sup} started [{pid,<0.2256.0>}, {name,httpc_sup}, {mfa,{httpc_sup,start_link, [[{httpc,{default,only_session_cookies}}]]}}, {restart_type,permanent}, {shutdown,infinity}, {child_type,supervisor}] PROGRESS REPORT <0.2254.0> 2011-05-07 17:27:43 =============================================================================== supervisor {local,inets_sup} started [{pid,<0.2272.0>}, {name,httpd_sup}, {mfa,{httpd_sup,start_link,[[]]}}, {restart_type,permanent}, {shutdown,infinity}, {child_type,supervisor}] PROGRESS REPORT <0.2254.0> 2011-05-07 17:27:43 =============================================================================== supervisor {local,inets_sup} started [{pid,<0.2273.0>}, {name,tftp_sup}, {mfa,{tftp_sup,start_link,[[]]}}, {restart_type,permanent}, {shutdown,infinity}, {child_type,supervisor}] PROGRESS REPORT <0.7.0> 2011-05-07 17:27:43 =============================================================================== application inets started_at 'ns_1@10.1.4.243' INFO REPORT <0.62.0> 2011-05-07 17:27:43 =============================================================================== ns_1@10.1.4.243:ns_cluster:306: Reply from engage_cluster on {"10.1.4.244", 8091}: {ok,{struct,[{<<"availableStorage">>, {struct,[{<<"hdd">>, [{struct,[{<<"path">>,<<"/">>}, {<<"sizeKBytes">>,7583436}, {<<"usagePercent">>,16}]}, {struct,[{<<"path">>,<<"/dev">>}, {<<"sizeKBytes">>,508460}, {<<"usagePercent">>,1}]}, {struct,[{<<"path">>,<<"/dev/shm">>}, {<<"sizeKBytes">>,513048}, {<<"usagePercent">>,0}]}, {struct,[{<<"path">>,<<"/var/run">>}, {<<"sizeKBytes">>,513048}, {<<"usagePercent">>,1}]}, {struct,[{<<"path">>,<<"/var/lock">>}, {<<"sizeKBytes">>,513048}, {<<"usagePercent">>,0}]}, {struct,[{<<"path">>,<<"/lib/init/rw">>}, {<<"sizeKBytes">>,513048}, {<<"usagePercent">>,0}]}, {struct,[{<<"path">>,<<"/boot">>}, {<<"sizeKBytes">>,233191}, {<<"usagePercent">>,8}]}]}]}}, {<<"memoryQuota">>,801}, {<<"storageTotals">>, {struct,[{<<"ram">>, {struct,[{<<"usedByData">>,0}, {<<"total">>,1050726400}, {<<"quotaTotal">>,839909376}, {<<"used">>,1031241728}]}}, {<<"hdd">>, {struct,[{<<"usedByData">>,0}, {<<"total">>,7765438464.0}, {<<"quotaTotal">>,7765438464.0}, {<<"used">>,1242470154}, {<<"free">>,6522968310.0}]}}]}}, {<<"storage">>, {struct,[{<<"ssd">>,[]}, {<<"hdd">>, [{struct,[{<<"path">>, <<"/var/opt/membase/1.6.5.4r/data/ns_1">>}, {<<"quotaMb">>,<<"none">>}, {<<"state">>,<<"ok">>}]}]}]}}, {<<"uptime">>,<<"1262">>}, {<<"memoryTotal">>,1050726400}, {<<"memoryFree">>,19484672}, {<<"mcdMemoryReserved">>,801}, {<<"mcdMemoryAllocated">>,801}, {<<"otpNode">>,<<"ns_1@10.1.4.244">>}, {<<"otpCookie">>,<<"qivnihjeclgfstlb">>}, {<<"clusterMembership">>,<<"active">>}, {<<"status">>,<<"healthy">>}, {<<"hostname">>,<<"10.1.4.244:8091">>}, {<<"clusterCompatibility">>,1}, {<<"version">>,<<"1.6.5.4r">>}, {<<"os">>,<<"i686-pc-linux-gnu">>}, {<<"ports">>, {struct,[{<<"proxy">>,11211},{<<"direct">>,11210}]}}]}} INFO REPORT <0.62.0> 2011-05-07 17:27:43 =============================================================================== ns_1@10.1.4.243:ns_cluster:371: port_please("ns_1", "10.1.4.244") = 21100 INFO REPORT <0.62.0> 2011-05-07 17:27:43 =============================================================================== ns_1@10.1.4.243:ns_cluster:461: Started node add transaction by adding node 'ns_1@10.1.4.244' to nodes_wanted INFO REPORT <0.64.0> 2011-05-07 17:27:43 =============================================================================== ns_node_disco_conf_events config on nodes_wanted INFO REPORT <0.64.0> 2011-05-07 17:27:43 =============================================================================== config change: nodes_wanted -> ['ns_1@10.1.4.243','ns_1@10.1.4.244'] INFO REPORT <0.2276.0> 2011-05-07 17:27:43 =============================================================================== ns_node_disco cookie_sync INFO REPORT <0.2276.0> 2011-05-07 17:27:43 =============================================================================== ns_node_disco: nodes_wanted updated: ['ns_1@10.1.4.243','ns_1@10.1.4.244'], with cookie: elyrdmyctqykhepq INFO REPORT <0.64.0> 2011-05-07 17:27:43 =============================================================================== ns_node_disco_conf_events config all INFO REPORT <0.85.0> 2011-05-07 17:27:43 =============================================================================== Pushing config INFO REPORT <0.85.0> 2011-05-07 17:27:43 =============================================================================== Pushing config done INFO REPORT <0.64.0> 2011-05-07 17:27:43 =============================================================================== config change: {node,'ns_1@10.1.4.244',membership} -> inactiveAdded INFO REPORT <0.64.0> 2011-05-07 17:27:43 =============================================================================== ns_node_disco_conf_events config all INFO REPORT <0.85.0> 2011-05-07 17:27:43 =============================================================================== Pushing config INFO REPORT <0.85.0> 2011-05-07 17:27:43 =============================================================================== Pushing config done INFO REPORT <0.62.0> 2011-05-07 17:27:43 =============================================================================== ns_1@10.1.4.243:ns_cluster:431: Posting the following to complete_join on "10.1.4.244:8091": {struct, [{<<"targetNode">>,'ns_1@10.1.4.244'}, {availableStorage, {struct, [{hdd, [{struct, [{path,<<"/">>}, {sizeKBytes,7583436}, {usagePercent,18}]}, {struct, [{path,<<"/dev">>}, {sizeKBytes,508460}, {usagePercent,1}]}, {struct, [{path,<<"/dev/shm">>}, {sizeKBytes,513048}, {usagePercent,0}]}, {struct, [{path,<<"/var/run">>}, {sizeKBytes,513048}, {usagePercent,1}]}, {struct, [{path,<<"/var/lock">>}, {sizeKBytes,513048}, {usagePercent,0}]}, {struct, [{path,<<"/lib/init/rw">>}, {sizeKBytes,513048}, {usagePercent,0}]}, {struct, [{path,<<"/boot">>}, {sizeKBytes,233191}, {usagePercent,8}]}]}]}}, {memoryQuota,801}, {storageTotals, {struct, [{ram, {struct, [{usedByData,48800032}, {total,1050726400}, {quotaTotal,839909376}, {used,784392192}]}}, {hdd, {struct, [{usedByData,28868608}, {total,7765438464}, {quotaTotal,7765438464}, {used,1397778923}, {free,6367659541}]}}]}}, {storage, {struct, [{ssd,[]}, {hdd, [{struct, [{path,<<"/var/opt/membase/1.6.5.4r/data/ns_1">>}, {quotaMb,none}, {state,ok}]}]}]}}, {uptime,<<"1360">>}, {memoryTotal,1050726400}, {memoryFree,266334208}, {mcdMemoryReserved,801}, {mcdMemoryAllocated,801}, {otpNode,<<"ns_1@10.1.4.243">>}, {otpCookie,<<"elyrdmyctqykhepq">>}, {clusterMembership,<<"active">>}, {status,<<"healthy">>}, {hostname,<<"10.1.4.243:8091">>}, {clusterCompatibility,1}, {version,<<"1.6.5.4r">>}, {os,<<"i686-pc-linux-gnu">>}, {ports,{struct,[{proxy,11211},{direct,11210}]}}]} INFO REPORT <0.2276.0> 2011-05-07 17:27:43 =============================================================================== ns_node_disco: nodes_wanted pong: ['ns_1@10.1.4.243'], with cookie: elyrdmyctqykhepq INFO REPORT <0.78.0> 2011-05-07 17:27:43 =============================================================================== ns_log: logging ns_node_disco:4:Node 'ns_1@10.1.4.243' saw that node 'ns_1@10.1.4.244' came up. INFO REPORT <0.64.0> 2011-05-07 17:27:44 =============================================================================== ns_node_disco_conf_events config on otp INFO REPORT <0.64.0> 2011-05-07 17:27:44 =============================================================================== config change: otp -> [{cookie,elyrdmyctqykhepq}] INFO REPORT <0.2301.0> 2011-05-07 17:27:44 =============================================================================== ns_node_disco cookie_sync INFO REPORT <0.2301.0> 2011-05-07 17:27:44 =============================================================================== ns_node_disco: nodes_wanted updated: ['ns_1@10.1.4.243','ns_1@10.1.4.244'], with cookie: elyrdmyctqykhepq INFO REPORT <0.64.0> 2011-05-07 17:27:44 =============================================================================== config change: {node,'ns_1@127.0.0.1',ns_log} -> [{filename,"/var/opt/membase/1.6.5.4r/data/ns_1/ns_log"}] INFO REPORT <0.2301.0> 2011-05-07 17:27:44 =============================================================================== ns_node_disco: nodes_wanted pong: ['ns_1@10.1.4.243','ns_1@10.1.4.244'], with cookie: elyrdmyctqykhepq INFO REPORT <0.64.0> 2011-05-07 17:27:44 =============================================================================== config change: {node,'ns_1@10.1.4.244',memcached} -> [{dbdir,"/var/opt/membase/1.6.5.4r/data/ns_1"}, {port,11210}, {admin_user,"_admin"}, {admin_pass,"_admin"}, {bucket_engine,"./bin/bucket_engine/bucket_engine.so"}, {engines,[{membase,[{engine,"bin/ep_engine/ep.so"}, {initfile,"/etc/opt/membase/1.6.5.4r/init.sql"}]}, {memcached,[{engine,"bin/memcached/default_engine.so"}]}]}, {verbosity,[]}] INFO REPORT <0.64.0> 2011-05-07 17:27:44 =============================================================================== config change: {node,'ns_1@127.0.0.1',membership} -> active INFO REPORT <0.64.0> 2011-05-07 17:27:44 =============================================================================== config change: {node,'ns_1@10.1.4.244',ns_log} -> [{filename,"/var/opt/membase/1.6.5.4r/data/ns_1/ns_log"}] INFO REPORT <0.64.0> 2011-05-07 17:27:44 =============================================================================== config change: {node,'ns_1@127.0.0.1',isasl} -> [{path,"/var/opt/membase/1.6.5.4r/data/ns_1/isasl.pw"}] INFO REPORT <0.64.0> 2011-05-07 17:27:44 =============================================================================== config change: {node,'ns_1@10.1.4.244',isasl} -> [{path,"/var/opt/membase/1.6.5.4r/data/ns_1/isasl.pw"}] INFO REPORT <0.64.0> 2011-05-07 17:27:44 =============================================================================== config change: {node,'ns_1@127.0.0.1',memcached} -> [{port,11210}, {dbdir,"/var/opt/membase/1.6.5.4r/data/ns_1"}, {admin_user,"_admin"}, {admin_pass,"_admin"}, {bucket_engine,"./bin/bucket_engine/bucket_engine.so"}, {engines,[{membase,[{engine,"bin/ep_engine/ep.so"}, {initfile,"/etc/opt/membase/1.6.5.4r/init.sql"}]}, {memcached,[{engine,"bin/memcached/default_engine.so"}]}]}, {verbosity,[]}] INFO REPORT <0.64.0> 2011-05-07 17:27:44 =============================================================================== ns_node_disco_conf_events config all INFO REPORT <0.85.0> 2011-05-07 17:27:44 =============================================================================== Pushing config INFO REPORT <0.85.0> 2011-05-07 17:27:44 =============================================================================== Pushing config done INFO REPORT <0.2204.0> 2011-05-07 17:27:44 =============================================================================== ns_1@10.1.4.243:stats_collector:71: Stats for bucket "default": auth_cmds 5 auth_errors 0 bucket_conns 6 bytes_read 39267499 bytes_written 53054678 cas_badval 0 cas_hits 0 cas_misses 0 cmd_flush 0 cmd_get 204800 cmd_set 215040 conn_yields 0 connection_structures 16 curr_connections 16 curr_items 204800 curr_items_tot 204800 daemon_connections 10 decr_hits 0 decr_misses 0 delete_hits 0 delete_misses 0 ep_bg_fetched 0 ep_commit_num 457 ep_commit_time 0 ep_commit_time_total 5 ep_data_age 1 ep_data_age_highwat 5 ep_db_cleaner_status complete ep_db_strategy multiMTVBDB ep_dbinit 1 ep_dbname /var/opt/membase/1.6.5.4r/data/ns_1/default ep_dbshards 4 ep_expired 0 ep_flush_duration 0 ep_flush_duration_highwat 1 ep_flush_duration_total 6 ep_flush_preempts 0 ep_flusher_state running ep_flusher_todo 0 ep_io_num_read 0 ep_io_num_write 215040 ep_io_read_bytes 0 ep_io_write_bytes 24514560 ep_item_begin_failed 0 ep_item_commit_failed 0 ep_item_flush_expired 0 ep_item_flush_failed 0 ep_kv_size 35840000 ep_max_data_size 839909376 ep_max_txn_size 1000 ep_mem_high_wat 629932032 ep_mem_low_wat 503945625 ep_min_data_age 0 ep_num_active_non_resident 0 ep_num_eject_failures 0 ep_num_eject_replicas 0 ep_num_expiry_pager_runs 0 ep_num_non_resident 0 ep_num_not_my_vbuckets 0 ep_num_pager_runs 0 ep_num_value_ejects 0 ep_oom_errors 0 ep_overhead 12960032 ep_pending_ops 0 ep_pending_ops_max 0 ep_pending_ops_max_duration 0 ep_pending_ops_total 0 ep_queue_age_cap 900 ep_queue_size 0 ep_storage_age 0 ep_storage_age_highwat 2 ep_storage_type featured ep_store_max_concurrency 10 ep_store_max_readers 9 ep_store_max_readwrite 1 ep_tap_bg_fetch_requeued 0 ep_tap_bg_fetched 0 ep_tap_keepalive 300 ep_tmp_oom_errors 0 ep_too_old 0 ep_too_young 0 ep_total_cache_size 35840000 ep_total_del_items 0 ep_total_enqueued 215239 ep_total_new_items 204800 ep_total_persisted 215040 ep_vbucket_del 0 ep_vbucket_del_fail 0 ep_version 1.6.5.3.1_8_gce31baa ep_warmed_up 0 ep_warmup true ep_warmup_dups 0 ep_warmup_oom 0 ep_warmup_thread complete ep_warmup_time 29657 get_hits 204800 get_misses 0 incr_hits 0 incr_misses 0 libevent 2.0.7-rc limit_maxbytes 67108864 mem_used 48800032 pid 5769 pointer_size 32 rejected_conns 0 rusage_system 2.464154 rusage_user 4.400275 threads 4 time 1304814463 total_connections 16 uptime 1362 version 1.4.4_382_g9df3289 INFO REPORT <0.85.0> 2011-05-07 17:27:44 =============================================================================== Pulling config from: 'ns_1@10.1.4.244' INFO REPORT <0.62.0> 2011-05-07 17:27:45 =============================================================================== ns_1@10.1.4.243:ns_cluster:437: Reply from complete_join on "10.1.4.244:8091": {ok,[]} INFO REPORT <0.62.0> 2011-05-07 17:27:45 =============================================================================== ns_1@10.1.4.243:ns_cluster:92: add_node("10.1.4.244", 8091, ..) -> {ok, 'ns_1@10.1.4.244'} INFO REPORT <0.77.0> 2011-05-07 17:27:48 =============================================================================== Detected a new node (from node 'ns_1@10.1.4.243'). Moving config around. INFO REPORT <0.77.0> 2011-05-07 17:27:48 =============================================================================== ns_node_disco_log: nodes changed: ['ns_1@10.1.4.243','ns_1@10.1.4.244'] INFO REPORT <0.85.0> 2011-05-07 17:27:48 =============================================================================== Pulling config INFO REPORT <0.85.0> 2011-05-07 17:27:48 =============================================================================== Pulling config from: 'ns_1@10.1.4.244' INFO REPORT <0.85.0> 2011-05-07 17:27:48 =============================================================================== Pulling config done INFO REPORT <0.109.0> 2011-05-07 17:27:59 =============================================================================== ns_log: logging ns_orchestrator:4:Starting rebalance, KeepNodes = ['ns_1@10.1.4.243','ns_1@10.1.4.244'], EjectNodes = [] INFO REPORT <0.64.0> 2011-05-07 17:27:59 =============================================================================== config change: rebalance_status -> running INFO REPORT <0.64.0> 2011-05-07 17:27:59 =============================================================================== ns_node_disco_conf_events config all INFO REPORT <0.85.0> 2011-05-07 17:27:59 =============================================================================== Pushing config INFO REPORT <0.85.0> 2011-05-07 17:27:59 =============================================================================== Pushing config done INFO REPORT <0.64.0> 2011-05-07 17:27:59 =============================================================================== config change: buckets -> [{configs,[{"default", [{num_replicas,1}, {ram_quota,839909376}, {auth_type,sasl}, {sasl_password,[]}, {type,membase}, {num_vbuckets,1024}, {db_shards,4}, {ht_size,3079}, {tap_keepalive,0}, {tap_noop_interval,20}, {max_txn_size,1000}, {ht_locks,5}, {servers,['ns_1@10.1.4.243','ns_1@10.1.4.244']}, {map,[['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243'|...], [...]|...]}]}]}] INFO REPORT <0.64.0> 2011-05-07 17:27:59 =============================================================================== ns_node_disco_conf_events config all INFO REPORT <0.85.0> 2011-05-07 17:27:59 =============================================================================== Pushing config INFO REPORT <0.85.0> 2011-05-07 17:27:59 =============================================================================== Pushing config done INFO REPORT <0.64.0> 2011-05-07 17:27:59 =============================================================================== config change: {node,'ns_1@10.1.4.244',membership} -> active INFO REPORT <0.64.0> 2011-05-07 17:27:59 =============================================================================== ns_node_disco_conf_events config all INFO REPORT <0.85.0> 2011-05-07 17:27:59 =============================================================================== Pushing config INFO REPORT <0.85.0> 2011-05-07 17:27:59 =============================================================================== Pushing config done ERROR REPORT <0.1506.0> 2011-05-07 17:27:59 =============================================================================== ns_1@10.1.4.243:stats_reader:176: Some nodes didn't respond: ['ns_1@10.1.4.244'] ERROR REPORT <0.1487.0> 2011-05-07 17:27:59 =============================================================================== ns_1@10.1.4.243:stats_reader:176: Some nodes didn't respond: ['ns_1@10.1.4.244'] ERROR REPORT <0.1950.0> 2011-05-07 17:27:59 =============================================================================== ns_1@10.1.4.243:stats_reader:176: Some nodes didn't respond: ['ns_1@10.1.4.244'] INFO REPORT <0.2350.0> 2011-05-07 17:28:00 =============================================================================== ns_1@10.1.4.243:ns_rebalancer:420: Waiting for ['ns_1@10.1.4.244'] INFO REPORT <0.2350.0> 2011-05-07 17:28:01 =============================================================================== ns_1@10.1.4.243:ns_rebalancer:420: Waiting for ['ns_1@10.1.4.244'] INFO REPORT <0.2350.0> 2011-05-07 17:28:02 =============================================================================== ns_1@10.1.4.243:ns_rebalancer:420: Waiting for ['ns_1@10.1.4.244'] INFO REPORT <0.109.0> 2011-05-07 17:28:03 =============================================================================== ns_1@10.1.4.243:ns_orchestrator:164: Skipping janitor in state rebalancing: {rebalancing_state, <0.2350.0>, {dict, 2, 16, 16, 8, 80, 48, {[], [], [], [], [], [], [], [], [], [], [], [], [], [], [], []}, {{[], [], [], [['ns_1@10.1.4.243'| 0.0]], [['ns_1@10.1.4.244'| 0.0]], [], [], [], [], [], [], [], [], [], [], []}}}} INFO REPORT <0.2350.0> 2011-05-07 17:28:03 =============================================================================== ns_1@10.1.4.243:ns_rebalancer:420: Waiting for ['ns_1@10.1.4.244'] INFO REPORT <0.2350.0> 2011-05-07 17:28:04 =============================================================================== ns_1@10.1.4.243:ns_rebalancer:420: Waiting for ['ns_1@10.1.4.244'] INFO REPORT <0.2350.0> 2011-05-07 17:28:05 =============================================================================== ns_1@10.1.4.243:ns_rebalancer:420: Waiting for ['ns_1@10.1.4.244'] INFO REPORT <0.2350.0> 2011-05-07 17:28:07 =============================================================================== ns_1@10.1.4.243:ns_rebalancer:420: Waiting for ['ns_1@10.1.4.244'] INFO REPORT <0.2350.0> 2011-05-07 17:28:08 =============================================================================== ns_1@10.1.4.243:ns_rebalancer:420: Waiting for ['ns_1@10.1.4.244'] INFO REPORT <0.2350.0> 2011-05-07 17:28:09 =============================================================================== ns_1@10.1.4.243:ns_rebalancer:420: Waiting for ['ns_1@10.1.4.244'] INFO REPORT <0.2350.0> 2011-05-07 17:28:10 =============================================================================== ns_1@10.1.4.243:ns_rebalancer:420: Waiting for ['ns_1@10.1.4.244'] INFO REPORT <0.2350.0> 2011-05-07 17:28:11 =============================================================================== ns_1@10.1.4.243:ns_rebalancer:420: Waiting for ['ns_1@10.1.4.244'] INFO REPORT <0.109.0> 2011-05-07 17:28:11 =============================================================================== ns_log: logging ns_orchestrator:2:Rebalance exited with reason wait_for_memcached_failed INFO REPORT <0.64.0> 2011-05-07 17:28:11 =============================================================================== config change: rebalance_status -> {none,<<"Rebalance failed. See logs for detailed reason. You can try rebalance again.">>} INFO REPORT <0.64.0> 2011-05-07 17:28:11 =============================================================================== ns_node_disco_conf_events config all INFO REPORT <0.85.0> 2011-05-07 17:28:11 =============================================================================== Pushing config INFO REPORT <0.85.0> 2011-05-07 17:28:11 =============================================================================== Pushing config done INFO REPORT <0.85.0> 2011-05-07 17:28:37 =============================================================================== Pulling config from: 'ns_1@10.1.4.244' INFO REPORT <0.2211.0> 2011-05-07 17:28:43 =============================================================================== ns_1@10.1.4.243:ns_doctor:82: Current node statuses: [{'ns_1@10.1.4.243', [{last_heard,{1304,814523,60377}}, {active_buckets,["default"]}, {memory, [{total,18452816}, {processes,7621300}, {processes_used,7612692}, {system,10831516}, {atom,535797}, {atom_used,521109}, {binary,173448}, {code,4282433}, {ets,886700}]}, {cluster_compatibility_version,1}, {version, [{kernel,"2.13.4"}, {mnesia,"4.4.12"}, {stdlib,"1.16.4"}, {os_mon,"2.2.4"}, {inets,"5.2"}, {ns_server,"1.6.5.4r"}, {menelaus,"1.6.5.4r"}, {sasl,"2.1.8"}]}, {system_arch,"i686-pc-linux-gnu"}, {wall_clock,1420}, {memory_data,{1050726400,787955712,{<0.65.0>,786116}}}, {disk_data, [{"/",7583436,18}, {"/dev",508460,1}, {"/dev/shm",513048,0}, {"/var/run",513048,1}, {"/var/lock",513048,0}, {"/lib/init/rw",513048,0}, {"/boot",233191,8}]}, {meminfo, <<"MemTotal: 1026100 kB\nMemFree: 253372 kB\nBuffers: 152768 kB\nCached: 450436 kB\nSwapCached: 0 kB\nActive: 456648 kB\nInactive: 277544 kB\nActive(anon): 126816 kB\nInactive(anon): 4412 kB\nActive(file): 329832 kB\nInactive(file): 273132 kB\nUnevictable: 0 kB\nMlocked: 0 kB\nHighTotal: 137160 kB\nHighFree: 252 kB\nLowTotal: 888940 kB\nLowFree: 253120 kB\nSwapTotal: 397304 kB\nSwapFree: 397304 kB\nDirty: 76 kB\nWriteback: 0 kB\nAnonPages: 131000 kB\nMapped: 10344 kB\nShmem: 240 kB\nSlab: 28196 kB\nSReclaimable: 23284 kB\nSUnreclaim: 4912 kB\nKernelStack: 968 kB\nPageTables: 1084 kB\nNFS_Unstable: 0 kB\nBounce: 0 kB\nWritebackTmp: 0 kB\nCommitLimit: 910352 kB\nCommitted_AS: 335232 kB\nVmallocTotal: 122880 kB\nVmallocUsed: 3540 kB\nVmallocChunk: 114424 kB\nHardwareCorrupted: 0 kB\nHugePages_Total: 0\nHugePages_Free: 0\nHugePages_Rsvd: 0\nHugePages_Surp: 0\nHugepagesize: 2048 kB\nDirectMap4k: 10232 kB\nDirectMap2M: 901120 kB\n">>}, {replication,[{"default",1.0}]}, {system_memory_data, [{system_total_memory,1050726400}, {free_swap,406839296}, {total_swap,406839296}, {cached_memory,461246464}, {buffered_memory,156434432}, {free_memory,259452928}, {total_memory,1050726400}]}, {statistics, [{wall_clock,{1410410,1}}, {context_switches,{274686,0}}, {garbage_collection,{60083,313145453,0}}, {io,{{input,17608800},{output,20092949}}}, {reductions,{230407718,1942501}}, {run_queue,0}, {runtime,{23560,180}}]}]}, {'ns_1@10.1.4.244', [{last_heard,{1304,814523,166458}}, {active_buckets,["default"]}, {memory, [{total,14765856}, {processes,4402072}, {processes_used,4394136}, {system,10363784}, {atom,533861}, {atom_used,518115}, {binary,275832}, {code,4246930}, {ets,368740}]}, {cluster_compatibility_version,1}, {version, [{kernel,"2.13.4"}, {mnesia,"4.4.12"}, {stdlib,"1.16.4"}, {os_mon,"2.2.4"}, {inets,"5.2"}, {ns_server,"1.6.5.4r"}, {menelaus,"1.6.5.4r"}, {sasl,"2.1.8"}]}, {system_arch,"i686-pc-linux-gnu"}, {wall_clock,1321}, {memory_data,{1050726400,1031942144,{<6517.21.0>,98288}}}, {disk_data, [{"/",7583436,16}, {"/dev",508460,1}, {"/dev/shm",513048,0}, {"/var/run",513048,1}, {"/var/lock",513048,0}, {"/lib/init/rw",513048,0}, {"/boot",233191,8}]}, {meminfo, <<"MemTotal: 1026100 kB\nMemFree: 32632 kB\nBuffers: 20252 kB\nCached: 239644 kB\nSwapCached: 0 kB\nActive: 132540 kB\nInactive: 182452 kB\nActive(anon): 27708 kB\nInactive(anon): 27640 kB\nActive(file): 104832 kB\nInactive(file): 154812 kB\nUnevictable: 0 kB\nMlocked: 0 kB\nHighTotal: 137160 kB\nHighFree: 252 kB\nLowTotal: 888940 kB\nLowFree: 32380 kB\nSwapTotal: 397304 kB\nSwapFree: 397304 kB\nDirty: 284 kB\nWriteback: 0 kB\nAnonPages: 55128 kB\nMapped: 9968 kB\nShmem: 228 kB\nSlab: 21544 kB\nSReclaimable: 16204 kB\nSUnreclaim: 5340 kB\nKernelStack: 920 kB\nPageTables: 808 kB\nNFS_Unstable: 0 kB\nBounce: 0 kB\nWritebackTmp: 0 kB\nCommitLimit: 910352 kB\nCommitted_AS: 252320 kB\nVmallocTotal: 122880 kB\nVmallocUsed: 3540 kB\nVmallocChunk: 114844 kB\nHardwareCorrupted: 0 kB\nHugePages_Total: 0\nHugePages_Free: 0\nHugePages_Rsvd: 0\nHugePages_Surp: 0\nHugepagesize: 2048 kB\nDirectMap4k: 10232 kB\nDirectMap2M: 901120 kB\n">>}, {replication,[{"default",1.0}]}, {system_memory_data, [{system_total_memory,1050726400}, {free_swap,406839296}, {total_swap,406839296}, {cached_memory,245395456}, {buffered_memory,20738048}, {free_memory,33415168}, {total_memory,1050726400}]}, {statistics, [{wall_clock,{1312638,1}}, {context_switches,{78901,0}}, {garbage_collection,{15119,34578258,0}}, {io,{{input,5934982},{output,1856472}}}, {reductions,{6624071,244255}}, {run_queue,0}, {runtime,{1530,30}}]}]}] INFO REPORT <0.2204.0> 2011-05-07 17:29:24 =============================================================================== ns_1@10.1.4.243:stats_collector:71: Stats for bucket "default": auth_cmds 8 auth_errors 0 bucket_conns 9 bytes_read 39272614 bytes_written 54585859 cas_badval 0 cas_hits 0 cas_misses 0 cmd_flush 0 cmd_get 204800 cmd_set 215040 conn_yields 0 connection_structures 19 curr_connections 19 curr_items 204800 curr_items_tot 204800 daemon_connections 10 decr_hits 0 decr_misses 0 delete_hits 0 delete_misses 0 ep_bg_fetched 0 ep_commit_num 457 ep_commit_time 0 ep_commit_time_total 5 ep_data_age 1 ep_data_age_highwat 5 ep_db_cleaner_status complete ep_db_strategy multiMTVBDB ep_dbinit 1 ep_dbname /var/opt/membase/1.6.5.4r/data/ns_1/default ep_dbshards 4 ep_expired 0 ep_flush_duration 0 ep_flush_duration_highwat 1 ep_flush_duration_total 6 ep_flush_preempts 0 ep_flusher_state running ep_flusher_todo 0 ep_io_num_read 0 ep_io_num_write 215040 ep_io_read_bytes 0 ep_io_write_bytes 24514560 ep_item_begin_failed 0 ep_item_commit_failed 0 ep_item_flush_expired 0 ep_item_flush_failed 0 ep_kv_size 35840000 ep_max_data_size 839909376 ep_max_txn_size 1000 ep_mem_high_wat 629932032 ep_mem_low_wat 503945625 ep_min_data_age 0 ep_num_active_non_resident 0 ep_num_eject_failures 0 ep_num_eject_replicas 0 ep_num_expiry_pager_runs 0 ep_num_non_resident 0 ep_num_not_my_vbuckets 0 ep_num_pager_runs 0 ep_num_value_ejects 0 ep_oom_errors 0 ep_overhead 12960032 ep_pending_ops 0 ep_pending_ops_max 0 ep_pending_ops_max_duration 0 ep_pending_ops_total 0 ep_queue_age_cap 900 ep_queue_size 0 ep_storage_age 0 ep_storage_age_highwat 2 ep_storage_type featured ep_store_max_concurrency 10 ep_store_max_readers 9 ep_store_max_readwrite 1 ep_tap_bg_fetch_requeued 0 ep_tap_bg_fetched 0 ep_tap_keepalive 300 ep_tmp_oom_errors 0 ep_too_old 0 ep_too_young 0 ep_total_cache_size 35840000 ep_total_del_items 0 ep_total_enqueued 215239 ep_total_new_items 204800 ep_total_persisted 215040 ep_vbucket_del 0 ep_vbucket_del_fail 0 ep_version 1.6.5.3.1_8_gce31baa ep_warmed_up 0 ep_warmup true ep_warmup_dups 0 ep_warmup_oom 0 ep_warmup_thread complete ep_warmup_time 29657 get_hits 204800 get_misses 0 incr_hits 0 incr_misses 0 libevent 2.0.7-rc limit_maxbytes 67108864 mem_used 48800032 pid 5769 pointer_size 32 rejected_conns 0 rusage_system 2.468154 rusage_user 4.444277 threads 4 time 1304814563 total_connections 19 uptime 1462 version 1.4.4_382_g9df3289 INFO REPORT <0.85.0> 2011-05-07 17:29:26 =============================================================================== Pulling config from: 'ns_1@10.1.4.244' INFO REPORT <0.109.0> 2011-05-07 17:29:31 =============================================================================== ns_log: logging ns_orchestrator:4:Starting rebalance, KeepNodes = ['ns_1@10.1.4.243'], EjectNodes = ['ns_1@10.1.4.244'] INFO REPORT <0.64.0> 2011-05-07 17:29:31 =============================================================================== config change: {node,'ns_1@10.1.4.243',membership} -> active INFO REPORT <0.64.0> 2011-05-07 17:29:31 =============================================================================== ns_node_disco_conf_events config all INFO REPORT <0.85.0> 2011-05-07 17:29:31 =============================================================================== Pushing config INFO REPORT <0.85.0> 2011-05-07 17:29:31 =============================================================================== Pushing config done INFO REPORT <0.64.0> 2011-05-07 17:29:31 =============================================================================== config change: rebalance_status -> running INFO REPORT <0.64.0> 2011-05-07 17:29:31 =============================================================================== ns_node_disco_conf_events config all INFO REPORT <0.85.0> 2011-05-07 17:29:31 =============================================================================== Pushing config INFO REPORT <0.85.0> 2011-05-07 17:29:31 =============================================================================== Pushing config done INFO REPORT <0.64.0> 2011-05-07 17:29:31 =============================================================================== config change: buckets -> [{configs,[{"default", [{num_replicas,1}, {ram_quota,839909376}, {auth_type,sasl}, {sasl_password,[]}, {type,membase}, {num_vbuckets,1024}, {db_shards,4}, {ht_size,3079}, {tap_keepalive,0}, {tap_noop_interval,20}, {max_txn_size,1000}, {ht_locks,5}, {servers,['ns_1@10.1.4.243']}, {map,[['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243'|...], [...]|...]}]}]}] INFO REPORT <0.64.0> 2011-05-07 17:29:31 =============================================================================== ns_node_disco_conf_events config all INFO REPORT <0.85.0> 2011-05-07 17:29:31 =============================================================================== Pushing config INFO REPORT <0.2720.0> 2011-05-07 17:29:31 =============================================================================== ns_1@10.1.4.243:ns_cluster:201: Shunning 'ns_1@10.1.4.244' INFO REPORT <0.109.0> 2011-05-07 17:29:31 =============================================================================== ns_log: logging ns_orchestrator:1:Rebalance completed successfully. INFO REPORT <0.85.0> 2011-05-07 17:29:31 =============================================================================== Pushing config done INFO REPORT <0.64.0> 2011-05-07 17:29:31 =============================================================================== config change: {node,'ns_1@10.1.4.244',membership} -> inactiveAdded INFO REPORT <0.64.0> 2011-05-07 17:29:31 =============================================================================== ns_node_disco_conf_events config all INFO REPORT <0.85.0> 2011-05-07 17:29:31 =============================================================================== Pushing config INFO REPORT <0.85.0> 2011-05-07 17:29:31 =============================================================================== Pushing config done INFO REPORT <0.64.0> 2011-05-07 17:29:31 =============================================================================== ns_node_disco_conf_events config on nodes_wanted INFO REPORT <0.64.0> 2011-05-07 17:29:31 =============================================================================== config change: nodes_wanted -> ['ns_1@10.1.4.243'] INFO REPORT <0.2750.0> 2011-05-07 17:29:31 =============================================================================== ns_node_disco cookie_sync INFO REPORT <0.2750.0> 2011-05-07 17:29:31 =============================================================================== ns_node_disco: nodes_wanted updated: ['ns_1@10.1.4.243'], with cookie: elyrdmyctqykhepq INFO REPORT <0.2750.0> 2011-05-07 17:29:31 =============================================================================== ns_node_disco: nodes_wanted pong: ['ns_1@10.1.4.243'], with cookie: elyrdmyctqykhepq INFO REPORT <0.64.0> 2011-05-07 17:29:31 =============================================================================== ns_node_disco_conf_events config all INFO REPORT <0.85.0> 2011-05-07 17:29:31 =============================================================================== Pushing config INFO REPORT <0.85.0> 2011-05-07 17:29:31 =============================================================================== Pushing config done INFO REPORT <0.64.0> 2011-05-07 17:29:31 =============================================================================== config change: rebalance_status -> none INFO REPORT <0.64.0> 2011-05-07 17:29:31 =============================================================================== ns_node_disco_conf_events config all INFO REPORT <0.85.0> 2011-05-07 17:29:31 =============================================================================== Pushing config INFO REPORT <0.85.0> 2011-05-07 17:29:31 =============================================================================== Pushing config done ERROR REPORT <0.109.0> 2011-05-07 17:29:31 =============================================================================== ns_1@10.1.4.243:ns_orchestrator:199: Got unexpected message {'EXIT', <6517.954.0>, normal} in state idle with data {idle_state, []} INFO REPORT <0.78.0> 2011-05-07 17:29:31 =============================================================================== ns_log: logging ns_node_disco:5:Node 'ns_1@10.1.4.243' saw that node 'ns_1@10.1.4.244' went down. INFO REPORT <0.77.0> 2011-05-07 17:29:36 =============================================================================== Detected a new node (from node 'ns_1@10.1.4.243'). Moving config around. INFO REPORT <0.77.0> 2011-05-07 17:29:36 =============================================================================== ns_node_disco_log: nodes changed: ['ns_1@10.1.4.243'] INFO REPORT <0.85.0> 2011-05-07 17:29:36 =============================================================================== Pulling config INFO REPORT <0.85.0> 2011-05-07 17:29:36 =============================================================================== Pulling config done INFO REPORT <0.2211.0> 2011-05-07 17:29:43 =============================================================================== ns_1@10.1.4.243:ns_doctor:82: Current node statuses: [{'ns_1@10.1.4.243', [{last_heard,{1304,814583,60297}}, {active_buckets,["default"]}, {memory, [{total,19144840}, {processes,8272244}, {processes_used,8265092}, {system,10872596}, {atom,536281}, {atom_used,521793}, {binary,187160}, {code,4290757}, {ets,920068}]}, {cluster_compatibility_version,1}, {version, [{kernel,"2.13.4"}, {mnesia,"4.4.12"}, {stdlib,"1.16.4"}, {os_mon,"2.2.4"}, {inets,"5.2"}, {ns_server,"1.6.5.4r"}, {menelaus,"1.6.5.4r"}, {sasl,"2.1.8"}]}, {system_arch,"i686-pc-linux-gnu"}, {wall_clock,1480}, {memory_data,{1050726400,789749760,{<0.2516.0>,2542988}}}, {disk_data, [{"/",7583436,18}, {"/dev",508460,1}, {"/dev/shm",513048,0}, {"/var/run",513048,1}, {"/var/lock",513048,0}, {"/lib/init/rw",513048,0}, {"/boot",233191,8}]}, {meminfo, <<"MemTotal: 1026100 kB\nMemFree: 255960 kB\nBuffers: 152768 kB\nCached: 450708 kB\nSwapCached: 0 kB\nActive: 453908 kB\nInactive: 277680 kB\nActive(anon): 123940 kB\nInactive(anon): 4412 kB\nActive(file): 329968 kB\nInactive(file): 273268 kB\nUnevictable: 0 kB\nMlocked: 0 kB\nHighTotal: 137160 kB\nHighFree: 252 kB\nLowTotal: 888940 kB\nLowFree: 255708 kB\nSwapTotal: 397304 kB\nSwapFree: 397304 kB\nDirty: 124 kB\nWriteback: 0 kB\nAnonPages: 128120 kB\nMapped: 10344 kB\nShmem: 240 kB\nSlab: 28196 kB\nSReclaimable: 23284 kB\nSUnreclaim: 4912 kB\nKernelStack: 984 kB\nPageTables: 1092 kB\nNFS_Unstable: 0 kB\nBounce: 0 kB\nWritebackTmp: 0 kB\nCommitLimit: 910352 kB\nCommitted_AS: 333472 kB\nVmallocTotal: 122880 kB\nVmallocUsed: 3540 kB\nVmallocChunk: 114424 kB\nHardwareCorrupted: 0 kB\nHugePages_Total: 0\nHugePages_Free: 0\nHugePages_Rsvd: 0\nHugePages_Surp: 0\nHugepagesize: 2048 kB\nDirectMap4k: 10232 kB\nDirectMap2M: 901120 kB\n">>}, {replication,[{"default",1.0}]}, {system_memory_data, [{system_total_memory,1050726400}, {free_swap,406839296}, {total_swap,406839296}, {cached_memory,461524992}, {buffered_memory,156434432}, {free_memory,262103040}, {total_memory,1050726400}]}, {statistics, [{wall_clock,{1470410,1}}, {context_switches,{294540,0}}, {garbage_collection,{64527,350630099,0}}, {io,{{input,17916354},{output,20850960}}}, {reductions,{244591140,3875837}}, {run_queue,0}, {runtime,{24740,290}}]}]}, {'ns_1@10.1.4.244', [{last_heard,{1304,814571,166103}}, {active_buckets,["default"]}, {memory, [{total,14701192}, {processes,4417776}, {processes_used,4409840}, {system,10283416}, {atom,533861}, {atom_used,518115}, {binary,141416}, {code,4246930}, {ets,422596}]}, {cluster_compatibility_version,1}, {version, [{kernel,"2.13.4"}, {mnesia,"4.4.12"}, {stdlib,"1.16.4"}, {os_mon,"2.2.4"}, {inets,"5.2"}, {ns_server,"1.6.5.4r"}, {menelaus,"1.6.5.4r"}, {sasl,"2.1.8"}]}, {system_arch,"i686-pc-linux-gnu"}, {wall_clock,1369}, {memory_data,{1050726400,1005064192,{<6517.898.0>,901104}}}, {disk_data, [{"/",7583436,16}, {"/dev",508460,1}, {"/dev/shm",513048,0}, {"/var/run",513048,1}, {"/var/lock",513048,0}, {"/lib/init/rw",513048,0}, {"/boot",233191,8}]}, {meminfo, <<"MemTotal: 1026100 kB\nMemFree: 68136 kB\nBuffers: 20336 kB\nCached: 239872 kB\nSwapCached: 0 kB\nActive: 132692 kB\nInactive: 182712 kB\nActive(anon): 27808 kB\nInactive(anon): 27624 kB\nActive(file): 104884 kB\nInactive(file): 155088 kB\nUnevictable: 0 kB\nMlocked: 0 kB\nHighTotal: 137160 kB\nHighFree: 252 kB\nLowTotal: 888940 kB\nLowFree: 67884 kB\nSwapTotal: 397304 kB\nSwapFree: 397304 kB\nDirty: 60 kB\nWriteback: 0 kB\nAnonPages: 55204 kB\nMapped: 9972 kB\nShmem: 228 kB\nSlab: 21504 kB\nSReclaimable: 16204 kB\nSUnreclaim: 5300 kB\nKernelStack: 912 kB\nPageTables: 808 kB\nNFS_Unstable: 0 kB\nBounce: 0 kB\nWritebackTmp: 0 kB\nCommitLimit: 910352 kB\nCommitted_AS: 252320 kB\nVmallocTotal: 122880 kB\nVmallocUsed: 3540 kB\nVmallocChunk: 114844 kB\nHardwareCorrupted: 0 kB\nHugePages_Total: 0\nHugePages_Free: 0\nHugePages_Rsvd: 0\nHugePages_Surp: 0\nHugepagesize: 2048 kB\nDirectMap4k: 10232 kB\nDirectMap2M: 901120 kB\n">>}, {replication,[{"default",1.0}]}, {system_memory_data, [{system_total_memory,1050726400}, {free_swap,406839296}, {total_swap,406839296}, {cached_memory,245628928}, {buffered_memory,20824064}, {free_memory,69771264}, {total_memory,1050726400}]}, {statistics, [{wall_clock,{1362639,2}}, {context_switches,{84278,0}}, {garbage_collection,{16613,39678872,0}}, {io,{{input,6146655},{output,2079972}}}, {reductions,{7982461,297345}}, {run_queue,0}, {runtime,{1660,30}}]}]}] INFO REPORT <0.62.0> 2011-05-07 17:30:32 =============================================================================== ns_1@10.1.4.243:ns_cluster:90: handling add_node("10.1.4.244", 8091, ..) INFO REPORT <0.62.0> 2011-05-07 17:30:32 =============================================================================== ns_1@10.1.4.243:ns_cluster:300: Posting node info to engage_cluster on {"10.1.4.244", 8091}: {struct, [{availableStorage, {struct, [{hdd, [{struct, [{path,<<"/">>}, {sizeKBytes,7583436}, {usagePercent,18}]}, {struct, [{path,<<"/dev">>}, {sizeKBytes,508460}, {usagePercent,1}]}, {struct, [{path,<<"/dev/shm">>}, {sizeKBytes,513048}, {usagePercent,0}]}, {struct, [{path,<<"/var/run">>}, {sizeKBytes,513048}, {usagePercent,1}]}, {struct, [{path,<<"/var/lock">>}, {sizeKBytes,513048}, {usagePercent,0}]}, {struct, [{path,<<"/lib/init/rw">>}, {sizeKBytes,513048}, {usagePercent,0}]}, {struct, [{path,<<"/boot">>}, {sizeKBytes,233191}, {usagePercent,8}]}]}]}}, {memoryQuota,801}, {storageTotals, {struct, [{ram, {struct, [{usedByData,48800032}, {total,1050726400}, {quotaTotal,839909376}, {used,789106688}]}}, {hdd, {struct, [{usedByData,28868608}, {total,7765438464}, {quotaTotal,7765438464}, {used,1397778923}, {free,6367659541}]}}]}}, {storage, {struct, [{ssd,[]}, {hdd, [{struct, [{path,<<"/var/opt/membase/1.6.5.4r/data/ns_1">>}, {quotaMb,none}, {state,ok}]}]}]}}, {uptime,<<"1529">>}, {memoryTotal,1050726400}, {memoryFree,261619712}, {mcdMemoryReserved,801}, {mcdMemoryAllocated,801}, {otpNode,<<"ns_1@10.1.4.243">>}, {otpCookie,<<"elyrdmyctqykhepq">>}, {clusterMembership,<<"active">>}, {status,<<"healthy">>}, {hostname,<<"10.1.4.243:8091">>}, {clusterCompatibility,1}, {version,<<"1.6.5.4r">>}, {os,<<"i686-pc-linux-gnu">>}, {ports,{struct,[{proxy,11211},{direct,11210}]}}]} INFO REPORT <0.62.0> 2011-05-07 17:30:32 =============================================================================== ns_1@10.1.4.243:ns_cluster:306: Reply from engage_cluster on {"10.1.4.244", 8091}: {ok,{struct,[{<<"availableStorage">>, {struct,[{<<"hdd">>, [{struct,[{<<"path">>,<<"/">>}, {<<"sizeKBytes">>,7583436}, {<<"usagePercent">>,16}]}, {struct,[{<<"path">>,<<"/dev">>}, {<<"sizeKBytes">>,508460}, {<<"usagePercent">>,1}]}, {struct,[{<<"path">>,<<"/dev/shm">>}, {<<"sizeKBytes">>,513048}, {<<"usagePercent">>,0}]}, {struct,[{<<"path">>,<<"/var/run">>}, {<<"sizeKBytes">>,513048}, {<<"usagePercent">>,1}]}, {struct,[{<<"path">>,<<"/var/lock">>}, {<<"sizeKBytes">>,513048}, {<<"usagePercent">>,0}]}, {struct,[{<<"path">>,<<"/lib/init/rw">>}, {<<"sizeKBytes">>,513048}, {<<"usagePercent">>,0}]}, {struct,[{<<"path">>,<<"/boot">>}, {<<"sizeKBytes">>,233191}, {<<"usagePercent">>,8}]}]}]}}, {<<"memoryQuota">>,801}, {<<"storageTotals">>, {struct,[{<<"ram">>, {struct,[{<<"usedByData">>,0}, {<<"total">>,1050726400}, {<<"quotaTotal">>,839909376}, {<<"used">>,900386816}]}}, {<<"hdd">>, {struct,[{<<"usedByData">>,0}, {<<"total">>,7765438464.0}, {<<"quotaTotal">>,7765438464.0}, {<<"used">>,1242470154}, {<<"free">>,6522968310.0}]}}]}}, {<<"storage">>, {struct,[{<<"ssd">>,[]}, {<<"hdd">>, [{struct,[{<<"path">>, <<"/var/opt/membase/1.6.5.4r/data/ns_1">>}, {<<"quotaMb">>,<<"none">>}, {<<"state">>,<<"ok">>}]}]}]}}, {<<"uptime">>,<<"32">>}, {<<"memoryTotal">>,1050726400}, {<<"memoryFree">>,150339584}, {<<"mcdMemoryReserved">>,801}, {<<"mcdMemoryAllocated">>,801}, {<<"otpNode">>,<<"ns_1@10.1.4.244">>}, {<<"otpCookie">>,<<"orpmslzbvbfpydxk">>}, {<<"clusterMembership">>,<<"active">>}, {<<"status">>,<<"healthy">>}, {<<"hostname">>,<<"10.1.4.244:8091">>}, {<<"clusterCompatibility">>,1}, {<<"version">>,<<"1.6.5.4r">>}, {<<"os">>,<<"i686-pc-linux-gnu">>}, {<<"ports">>, {struct,[{<<"proxy">>,11211},{<<"direct">>,11210}]}}]}} INFO REPORT <0.62.0> 2011-05-07 17:30:32 =============================================================================== ns_1@10.1.4.243:ns_cluster:371: port_please("ns_1", "10.1.4.244") = 21100 INFO REPORT <0.62.0> 2011-05-07 17:30:32 =============================================================================== ns_1@10.1.4.243:ns_cluster:461: Started node add transaction by adding node 'ns_1@10.1.4.244' to nodes_wanted INFO REPORT <0.64.0> 2011-05-07 17:30:32 =============================================================================== ns_node_disco_conf_events config on nodes_wanted INFO REPORT <0.64.0> 2011-05-07 17:30:32 =============================================================================== config change: nodes_wanted -> ['ns_1@10.1.4.243','ns_1@10.1.4.244'] INFO REPORT <0.2999.0> 2011-05-07 17:30:32 =============================================================================== ns_node_disco cookie_sync INFO REPORT <0.2999.0> 2011-05-07 17:30:32 =============================================================================== ns_node_disco: nodes_wanted updated: ['ns_1@10.1.4.243','ns_1@10.1.4.244'], with cookie: elyrdmyctqykhepq INFO REPORT <0.64.0> 2011-05-07 17:30:32 =============================================================================== ns_node_disco_conf_events config all INFO REPORT <0.85.0> 2011-05-07 17:30:32 =============================================================================== Pushing config INFO REPORT <0.85.0> 2011-05-07 17:30:32 =============================================================================== Pushing config done INFO REPORT <0.64.0> 2011-05-07 17:30:32 =============================================================================== config change: {node,'ns_1@10.1.4.244',membership} -> inactiveAdded INFO REPORT <0.2999.0> 2011-05-07 17:30:32 =============================================================================== ns_node_disco: nodes_wanted pong: ['ns_1@10.1.4.243'], with cookie: elyrdmyctqykhepq INFO REPORT <0.64.0> 2011-05-07 17:30:32 =============================================================================== ns_node_disco_conf_events config all INFO REPORT <0.85.0> 2011-05-07 17:30:32 =============================================================================== Pushing config INFO REPORT <0.85.0> 2011-05-07 17:30:32 =============================================================================== Pushing config done INFO REPORT <0.62.0> 2011-05-07 17:30:32 =============================================================================== ns_1@10.1.4.243:ns_cluster:431: Posting the following to complete_join on "10.1.4.244:8091": {struct, [{<<"targetNode">>,'ns_1@10.1.4.244'}, {availableStorage, {struct, [{hdd, [{struct, [{path,<<"/">>}, {sizeKBytes,7583436}, {usagePercent,18}]}, {struct, [{path,<<"/dev">>}, {sizeKBytes,508460}, {usagePercent,1}]}, {struct, [{path,<<"/dev/shm">>}, {sizeKBytes,513048}, {usagePercent,0}]}, {struct, [{path,<<"/var/run">>}, {sizeKBytes,513048}, {usagePercent,1}]}, {struct, [{path,<<"/var/lock">>}, {sizeKBytes,513048}, {usagePercent,0}]}, {struct, [{path,<<"/lib/init/rw">>}, {sizeKBytes,513048}, {usagePercent,0}]}, {struct, [{path,<<"/boot">>}, {sizeKBytes,233191}, {usagePercent,8}]}]}]}}, {memoryQuota,801}, {storageTotals, {struct, [{ram, {struct, [{usedByData,48800032}, {total,1050726400}, {quotaTotal,839909376}, {used,789106688}]}}, {hdd, {struct, [{usedByData,28868608}, {total,7765438464}, {quotaTotal,7765438464}, {used,1397778923}, {free,6367659541}]}}]}}, {storage, {struct, [{ssd,[]}, {hdd, [{struct, [{path,<<"/var/opt/membase/1.6.5.4r/data/ns_1">>}, {quotaMb,none}, {state,ok}]}]}]}}, {uptime,<<"1529">>}, {memoryTotal,1050726400}, {memoryFree,261619712}, {mcdMemoryReserved,801}, {mcdMemoryAllocated,801}, {otpNode,<<"ns_1@10.1.4.243">>}, {otpCookie,<<"elyrdmyctqykhepq">>}, {clusterMembership,<<"active">>}, {status,<<"healthy">>}, {hostname,<<"10.1.4.243:8091">>}, {clusterCompatibility,1}, {version,<<"1.6.5.4r">>}, {os,<<"i686-pc-linux-gnu">>}, {ports,{struct,[{proxy,11211},{direct,11210}]}}]} INFO REPORT <0.78.0> 2011-05-07 17:30:32 =============================================================================== ns_log: logging ns_node_disco:4:Node 'ns_1@10.1.4.243' saw that node 'ns_1@10.1.4.244' came up. INFO REPORT <0.71.0> 2011-05-07 17:30:32 =============================================================================== ns_log: suppressing duplicate log ns_node_disco:4("Node 'ns_1@10.1.4.243' saw that node 'ns_1@10.1.4.244' came up.") because it's been seen 1 times in the past 168.947051 secs (last seen 168.947051 secs ago INFO REPORT <0.62.0> 2011-05-07 17:30:33 =============================================================================== ns_1@10.1.4.243:ns_cluster:437: Reply from complete_join on "10.1.4.244:8091": {ok,[]} INFO REPORT <0.62.0> 2011-05-07 17:30:33 =============================================================================== ns_1@10.1.4.243:ns_cluster:92: add_node("10.1.4.244", 8091, ..) -> {ok, 'ns_1@10.1.4.244'} INFO REPORT <0.109.0> 2011-05-07 17:30:34 =============================================================================== ns_log: logging ns_orchestrator:4:Starting rebalance, KeepNodes = ['ns_1@10.1.4.243','ns_1@10.1.4.244'], EjectNodes = [] INFO REPORT <0.71.0> 2011-05-07 17:30:34 =============================================================================== ns_log: suppressing duplicate log ns_orchestrator:4("Starting rebalance, KeepNodes = ['ns_1@10.1.4.243','ns_1@10.1.4.244'], EjectNodes = []\n") because it's been seen 1 times in the past 155.260353 secs (last seen 155.260353 secs ago INFO REPORT <0.64.0> 2011-05-07 17:30:34 =============================================================================== config change: {node,'ns_1@10.1.4.244',membership} -> active INFO REPORT <0.64.0> 2011-05-07 17:30:34 =============================================================================== config change: {node,'ns_1@10.1.4.243',membership} -> active INFO REPORT <0.64.0> 2011-05-07 17:30:34 =============================================================================== ns_node_disco_conf_events config all INFO REPORT <0.85.0> 2011-05-07 17:30:34 =============================================================================== Pushing config INFO REPORT <0.85.0> 2011-05-07 17:30:34 =============================================================================== Pushing config done INFO REPORT <0.64.0> 2011-05-07 17:30:34 =============================================================================== config change: rebalance_status -> running INFO REPORT <0.64.0> 2011-05-07 17:30:34 =============================================================================== ns_node_disco_conf_events config all INFO REPORT <0.85.0> 2011-05-07 17:30:34 =============================================================================== Pushing config INFO REPORT <0.85.0> 2011-05-07 17:30:34 =============================================================================== Pushing config done INFO REPORT <0.64.0> 2011-05-07 17:30:34 =============================================================================== config change: buckets -> [{configs,[{"default", [{num_replicas,1}, {ram_quota,839909376}, {auth_type,sasl}, {sasl_password,[]}, {type,membase}, {num_vbuckets,1024}, {db_shards,4}, {ht_size,3079}, {tap_keepalive,0}, {tap_noop_interval,20}, {max_txn_size,1000}, {ht_locks,5}, {servers,['ns_1@10.1.4.243','ns_1@10.1.4.244']}, {map,[['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243'|...], [...]|...]}]}]}] INFO REPORT <0.64.0> 2011-05-07 17:30:34 =============================================================================== ns_node_disco_conf_events config all INFO REPORT <0.85.0> 2011-05-07 17:30:34 =============================================================================== Pushing config INFO REPORT <0.85.0> 2011-05-07 17:30:34 =============================================================================== Pushing config done ERROR REPORT <0.2915.0> 2011-05-07 17:30:34 =============================================================================== ns_1@10.1.4.243:stats_reader:176: Some nodes didn't respond: ['ns_1@10.1.4.244'] ERROR REPORT <0.3024.0> 2011-05-07 17:30:34 =============================================================================== ns_1@10.1.4.243:stats_reader:176: Some nodes didn't respond: ['ns_1@10.1.4.244'] INFO REPORT <0.3035.0> 2011-05-07 17:30:35 =============================================================================== ns_1@10.1.4.243:ns_rebalancer:420: Waiting for ['ns_1@10.1.4.244'] INFO REPORT <0.3035.0> 2011-05-07 17:30:36 =============================================================================== ns_1@10.1.4.243:ns_rebalancer:420: Waiting for ['ns_1@10.1.4.244'] INFO REPORT <0.3035.0> 2011-05-07 17:30:37 =============================================================================== ns_1@10.1.4.243:ns_rebalancer:420: Waiting for ['ns_1@10.1.4.244'] INFO REPORT <0.77.0> 2011-05-07 17:30:37 =============================================================================== Detected a new node (from node 'ns_1@10.1.4.243'). Moving config around. INFO REPORT <0.77.0> 2011-05-07 17:30:37 =============================================================================== ns_node_disco_log: nodes changed: ['ns_1@10.1.4.243','ns_1@10.1.4.244'] INFO REPORT <0.85.0> 2011-05-07 17:30:37 =============================================================================== Pulling config INFO REPORT <0.85.0> 2011-05-07 17:30:37 =============================================================================== Pulling config from: 'ns_1@10.1.4.244' INFO REPORT <0.85.0> 2011-05-07 17:30:37 =============================================================================== Pulling config done INFO REPORT <0.3035.0> 2011-05-07 17:30:38 =============================================================================== ns_1@10.1.4.243:ns_rebalancer:420: Waiting for ['ns_1@10.1.4.244'] INFO REPORT <0.3035.0> 2011-05-07 17:30:39 =============================================================================== ns_1@10.1.4.243:ns_rebalancer:420: Waiting for ['ns_1@10.1.4.244'] INFO REPORT <0.3035.0> 2011-05-07 17:30:40 =============================================================================== ns_1@10.1.4.243:ns_rebalancer:420: Waiting for ['ns_1@10.1.4.244'] INFO REPORT <0.3035.0> 2011-05-07 17:30:41 =============================================================================== ns_1@10.1.4.243:ns_rebalancer:420: Waiting for ['ns_1@10.1.4.244'] INFO REPORT <0.3098.0> 2011-05-07 17:30:42 =============================================================================== vbucketmigrator<0.3098.0>: Connecting to {Sock 10.1.4.244:11210} vbucketmigrator<0.3098.0>: Authenticating towards: {Sock 10.1.4.244:11210} vbucketmigrator<0.3098.0>: Authenticated towards: {Sock 10.1.4.244:11210} vbucketmigrator<0.3098.0>: Connecting to {Sock 10.1.4.243:11210} vbucketmigrator<0.3098.0>: Authenticating towards: {Sock 10.1.4.243:11210} vbucketmigrator<0.3098.0>: Authenticated towards: {Sock 10.1.4.243:11210} vbucketmigrator<0.3098.0>: Starting to move bucket 0 INFO REPORT <0.104.0> 2011-05-07 17:30:42 =============================================================================== memcached<0.104.0>: Vbucket is going dead. INFO REPORT <0.109.0> 2011-05-07 17:30:43 =============================================================================== ns_1@10.1.4.243:ns_orchestrator:164: Skipping janitor in state rebalancing: {rebalancing_state, <0.3035.0>, {dict, 2, 16, 16, 8, 80, 48, {[], [], [], [], [], [], [], [], [], [], [], [], [], [], [], []}, {{[], [], [], [['ns_1@10.1.4.243'| 0.0]], [['ns_1@10.1.4.244'| 0.0]], [], [], [], [], [], [], [], [], [], [], []}}}} INFO REPORT <0.2211.0> 2011-05-07 17:30:43 =============================================================================== ns_1@10.1.4.243:ns_doctor:82: Current node statuses: [{'ns_1@10.1.4.243', [{last_heard,{1304,814643,60449}}, {active_buckets,["default"]}, {memory, [{total,22952304}, {processes,12050636}, {processes_used,12047228}, {system,10901668}, {atom,536281}, {atom_used,521961}, {binary,162840}, {code,4290757}, {ets,953948}]}, {cluster_compatibility_version,1}, {version, [{kernel,"2.13.4"}, {mnesia,"4.4.12"}, {stdlib,"1.16.4"}, {os_mon,"2.2.4"}, {inets,"5.2"}, {ns_server,"1.6.5.4r"}, {menelaus,"1.6.5.4r"}, {sasl,"2.1.8"}]}, {system_arch,"i686-pc-linux-gnu"}, {wall_clock,1540}, {memory_data,{1050726400,789106688,{<0.2218.0>,2542932}}}, {disk_data, [{"/",7583436,18}, {"/dev",508460,1}, {"/dev/shm",513048,0}, {"/var/run",513048,1}, {"/var/lock",513048,0}, {"/lib/init/rw",513048,0}, {"/boot",233191,8}]}, {meminfo, <<"MemTotal: 1026100 kB\nMemFree: 253372 kB\nBuffers: 152768 kB\nCached: 450968 kB\nSwapCached: 0 kB\nActive: 456492 kB\nInactive: 277824 kB\nActive(anon): 126408 kB\nInactive(anon): 4412 kB\nActive(file): 330084 kB\nInactive(file): 273412 kB\nUnevictable: 0 kB\nMlocked: 0 kB\nHighTotal: 137160 kB\nHighFree: 252 kB\nLowTotal: 888940 kB\nLowFree: 253120 kB\nSwapTotal: 397304 kB\nSwapFree: 397304 kB\nDirty: 128 kB\nWriteback: 0 kB\nAnonPages: 130596 kB\nMapped: 10348 kB\nShmem: 240 kB\nSlab: 28196 kB\nSReclaimable: 23284 kB\nSUnreclaim: 4912 kB\nKernelStack: 968 kB\nPageTables: 1088 kB\nNFS_Unstable: 0 kB\nBounce: 0 kB\nWritebackTmp: 0 kB\nCommitLimit: 910352 kB\nCommitted_AS: 336984 kB\nVmallocTotal: 122880 kB\nVmallocUsed: 3540 kB\nVmallocChunk: 114424 kB\nHardwareCorrupted: 0 kB\nHugePages_Total: 0\nHugePages_Free: 0\nHugePages_Rsvd: 0\nHugePages_Surp: 0\nHugepagesize: 2048 kB\nDirectMap4k: 10232 kB\nDirectMap2M: 901120 kB\n">>}, {replication,[{"default",1.0}]}, {system_memory_data, [{system_total_memory,1050726400}, {free_swap,406839296}, {total_swap,406839296}, {cached_memory,461791232}, {buffered_memory,156434432}, {free_memory,259452928}, {total_memory,1050726400}]}, {statistics, [{wall_clock,{1530410,1}}, {context_switches,{312834,0}}, {garbage_collection,{68491,384935062,0}}, {io,{{input,17969183},{output,21323289}}}, {reductions,{258445843,173968}}, {run_queue,0}, {runtime,{25790,10}}]}]}, {'ns_1@10.1.4.244', [{last_heard,{1304,814643,110201}}, {active_buckets,["default"]}, {memory, [{total,13615008}, {processes,3949764}, {processes_used,3945940}, {system,9665244}, {atom,505565}, {atom_used,474698}, {binary,113000}, {code,3843102}, {ets,302300}]}, {cluster_compatibility_version,1}, {version, [{kernel,"2.13.4"}, {mnesia,"4.4.12"}, {stdlib,"1.16.4"}, {os_mon,"2.2.4"}, {ns_server,"1.6.5.4r"}, {menelaus,"1.6.5.4r"}, {sasl,"2.1.8"}]}, {system_arch,"i686-pc-linux-gnu"}, {wall_clock,43}, {memory_data,{1050726400,900386816,{<6517.21.0>,81564}}}, {disk_data, [{"/",7583436,16}, {"/dev",508460,1}, {"/dev/shm",513048,0}, {"/var/run",513048,1}, {"/var/lock",513048,0}, {"/lib/init/rw",513048,0}, {"/boot",233191,8}]}, {meminfo, <<"MemTotal: 1026100 kB\nMemFree: 171940 kB\nBuffers: 20516 kB\nCached: 243016 kB\nSwapCached: 0 kB\nActive: 122068 kB\nInactive: 162108 kB\nActive(anon): 14724 kB\nInactive(anon): 6152 kB\nActive(file): 107344 kB\nInactive(file): 155956 kB\nUnevictable: 0 kB\nMlocked: 0 kB\nHighTotal: 137160 kB\nHighFree: 2240 kB\nLowTotal: 888940 kB\nLowFree: 169700 kB\nSwapTotal: 397304 kB\nSwapFree: 397304 kB\nDirty: 116 kB\nWriteback: 0 kB\nAnonPages: 20652 kB\nMapped: 8388 kB\nShmem: 232 kB\nSlab: 21560 kB\nSReclaimable: 16296 kB\nSUnreclaim: 5264 kB\nKernelStack: 840 kB\nPageTables: 724 kB\nNFS_Unstable: 0 kB\nBounce: 0 kB\nWritebackTmp: 0 kB\nCommitLimit: 910352 kB\nCommitted_AS: 127880 kB\nVmallocTotal: 122880 kB\nVmallocUsed: 3540 kB\nVmallocChunk: 114844 kB\nHardwareCorrupted: 0 kB\nHugePages_Total: 0\nHugePages_Free: 0\nHugePages_Rsvd: 0\nHugePages_Surp: 0\nHugepagesize: 2048 kB\nDirectMap4k: 10232 kB\nDirectMap2M: 901120 kB\n">>}, {replication,[{"default",1.0}]}, {system_memory_data, [{system_total_memory,1050726400}, {free_swap,406839296}, {total_swap,406839296}, {cached_memory,248848384}, {buffered_memory,21008384}, {free_memory,176095232}, {total_memory,1050726400}]}, {statistics, [{wall_clock,{33023,211}}, {context_switches,{12676,0}}, {garbage_collection,{2577,5036116,0}}, {io,{{input,4813252},{output,1193506}}}, {reductions,{2278358,74634}}, {run_queue,3}, {runtime,{590,10}}]}]}] INFO REPORT <0.3098.0> 2011-05-07 17:30:48 =============================================================================== vbucketmigrator<0.3098.0>: Bucket 0 moved to the next server INFO REPORT <0.109.0> 2011-05-07 17:30:53 =============================================================================== ns_1@10.1.4.243:ns_orchestrator:164: Skipping janitor in state rebalancing: {rebalancing_state, <0.3035.0>, {dict, 2, 16, 16, 8, 80, 48, {[], [], [], [], [], [], [], [], [], [], [], [], [], [], [], []}, {{[], [], [], [['ns_1@10.1.4.243'| 0.0]], [['ns_1@10.1.4.244'| 0.0]], [], [], [], [], [], [], [], [], [], [], []}}}} INFO REPORT <0.109.0> 2011-05-07 17:31:03 =============================================================================== ns_1@10.1.4.243:ns_orchestrator:164: Skipping janitor in state rebalancing: {rebalancing_state, <0.3035.0>, {dict, 2, 16, 16, 8, 80, 48, {[], [], [], [], [], [], [], [], [], [], [], [], [], [], [], []}, {{[], [], [], [['ns_1@10.1.4.243'| 0.0]], [['ns_1@10.1.4.244'| 0.0]], [], [], [], [], [], [], [], [], [], [], []}}}} INFO REPORT <0.85.0> 2011-05-07 17:31:03 =============================================================================== Pulling config from: 'ns_1@10.1.4.244' INFO REPORT <0.2204.0> 2011-05-07 17:31:04 =============================================================================== ns_1@10.1.4.243:stats_collector:71: Stats for bucket "default": auth_cmds 9 auth_errors 0 bucket_conns 10 bytes_read 39277808 bytes_written 56128182 cas_badval 0 cas_hits 0 cas_misses 0 cmd_flush 0 cmd_get 204800 cmd_set 215040 conn_yields 0 connection_structures 20 curr_connections 20 curr_items 204600 curr_items_tot 204800 daemon_connections 10 decr_hits 0 decr_misses 0 delete_hits 0 delete_misses 0 ep_bg_fetched 0 ep_commit_num 457 ep_commit_time 0 ep_commit_time_total 5 ep_data_age 1 ep_data_age_highwat 5 ep_db_cleaner_status complete ep_db_strategy multiMTVBDB ep_dbinit 1 ep_dbname /var/opt/membase/1.6.5.4r/data/ns_1/default ep_dbshards 4 ep_expired 0 ep_flush_duration 0 ep_flush_duration_highwat 1 ep_flush_duration_total 6 ep_flush_preempts 0 ep_flusher_state running ep_flusher_todo 0 ep_io_num_read 200 ep_io_num_write 215040 ep_io_read_bytes 22800 ep_io_write_bytes 24514560 ep_item_begin_failed 0 ep_item_commit_failed 0 ep_item_flush_expired 0 ep_item_flush_failed 0 ep_kv_size 35840000 ep_max_data_size 839909376 ep_max_txn_size 1000 ep_mem_high_wat 629932032 ep_mem_low_wat 503945625 ep_min_data_age 0 ep_num_active_non_resident 0 ep_num_eject_failures 0 ep_num_eject_replicas 0 ep_num_expiry_pager_runs 0 ep_num_non_resident 0 ep_num_not_my_vbuckets 0 ep_num_pager_runs 0 ep_num_value_ejects 0 ep_oom_errors 0 ep_overhead 12960032 ep_pending_ops 0 ep_pending_ops_max 0 ep_pending_ops_max_duration 0 ep_pending_ops_total 0 ep_queue_age_cap 900 ep_queue_size 0 ep_storage_age 0 ep_storage_age_highwat 2 ep_storage_type featured ep_store_max_concurrency 10 ep_store_max_readers 9 ep_store_max_readwrite 1 ep_tap_bg_fetch_requeued 0 ep_tap_bg_fetched 200 ep_tap_keepalive 300 ep_tmp_oom_errors 0 ep_too_old 0 ep_too_young 0 ep_total_cache_size 35840000 ep_total_del_items 0 ep_total_enqueued 215239 ep_total_new_items 204800 ep_total_persisted 215040 ep_vbucket_del 0 ep_vbucket_del_fail 0 ep_version 1.6.5.3.1_8_gce31baa ep_warmed_up 0 ep_warmup true ep_warmup_dups 0 ep_warmup_oom 0 ep_warmup_thread complete ep_warmup_time 29657 get_hits 204800 get_misses 0 incr_hits 0 incr_misses 0 libevent 2.0.7-rc limit_maxbytes 67108864 mem_used 48800032 pid 5769 pointer_size 32 rejected_conns 0 rusage_system 2.492155 rusage_user 4.488280 tap_connect_received 1 tap_mutation_sent 400 tap_opaque_sent 1 tap_vbucket_set_sent 2 threads 4 time 1304814663 total_connections 20 uptime 1562 version 1.4.4_382_g9df3289 INFO REPORT <0.85.0> 2011-05-07 17:31:12 =============================================================================== Pulling config from: 'ns_1@10.1.4.244' INFO REPORT <0.109.0> 2011-05-07 17:31:13 =============================================================================== ns_1@10.1.4.243:ns_orchestrator:164: Skipping janitor in state rebalancing: {rebalancing_state, <0.3035.0>, {dict, 2, 16, 16, 8, 80, 48, {[], [], [], [], [], [], [], [], [], [], [], [], [], [], [], []}, {{[], [], [], [['ns_1@10.1.4.243'| 0.0]], [['ns_1@10.1.4.244'| 0.0]], [], [], [], [], [], [], [], [], [], [], []}}}} INFO REPORT <0.109.0> 2011-05-07 17:31:23 =============================================================================== ns_1@10.1.4.243:ns_orchestrator:164: Skipping janitor in state rebalancing: {rebalancing_state, <0.3035.0>, {dict, 2, 16, 16, 8, 80, 48, {[], [], [], [], [], [], [], [], [], [], [], [], [], [], [], []}, {{[], [], [], [['ns_1@10.1.4.243'| 0.0]], [['ns_1@10.1.4.244'| 0.0]], [], [], [], [], [], [], [], [], [], [], []}}}} INFO REPORT <0.109.0> 2011-05-07 17:31:33 =============================================================================== ns_1@10.1.4.243:ns_orchestrator:164: Skipping janitor in state rebalancing: {rebalancing_state, <0.3035.0>, {dict, 2, 16, 16, 8, 80, 48, {[], [], [], [], [], [], [], [], [], [], [], [], [], [], [], []}, {{[], [], [], [['ns_1@10.1.4.243'| 0.0]], [['ns_1@10.1.4.244'| 0.0]], [], [], [], [], [], [], [], [], [], [], []}}}} ERROR REPORT <0.3098.0> 2011-05-07 17:31:41 =============================================================================== ** Generic server <0.3098.0> terminating ** Last message in was {'EXIT',<0.3097.0>,stopped} ** When Server state == {state,#Port<0.4320>,vbucketmigrator, {["Bucket 0 moved to the next server", "Starting to move bucket 0"], ["Authenticated towards: {Sock 10.1.4.243:11210}"]}, undefined,[],0} ** Reason for termination == ** stopped CRASH REPORT <0.3098.0> 2011-05-07 17:31:41 =============================================================================== Crashing process initial_call {ns_port_server,init,['Argument__1']} pid <0.3098.0> registered_name [] error_info {exit,stopped, [{gen_server,terminate,6},{proc_lib,init_p_do_apply,3}]} ancestors [<0.3097.0>] messages [{'EXIT',#Port<0.4320>,normal}] links [] dictionary [] trap_exit true status running heap_size 1597 stack_size 24 reductions 574 INFO REPORT <0.109.0> 2011-05-07 17:31:41 =============================================================================== ns_log: logging ns_orchestrator:2:Rebalance exited with reason stopped INFO REPORT <0.64.0> 2011-05-07 17:31:41 =============================================================================== config change: buckets -> [{configs,[{"default", [{num_replicas,1}, {ram_quota,839909376}, {auth_type,sasl}, {sasl_password,[]}, {type,membase}, {num_vbuckets,1024}, {db_shards,4}, {ht_size,3079}, {tap_keepalive,0}, {tap_noop_interval,20}, {max_txn_size,1000}, {ht_locks,5}, {servers,['ns_1@10.1.4.243','ns_1@10.1.4.244']}, {map,[['ns_1@10.1.4.243','ns_1@10.1.4.244'], ['ns_1@10.1.4.243','ns_1@10.1.4.244'], ['ns_1@10.1.4.243','ns_1@10.1.4.244'], ['ns_1@10.1.4.243','ns_1@10.1.4.244'], ['ns_1@10.1.4.243','ns_1@10.1.4.244'], ['ns_1@10.1.4.243','ns_1@10.1.4.244'], ['ns_1@10.1.4.243','ns_1@10.1.4.244'], ['ns_1@10.1.4.243','ns_1@10.1.4.244'], ['ns_1@10.1.4.243','ns_1@10.1.4.244'], ['ns_1@10.1.4.243','ns_1@10.1.4.244'], ['ns_1@10.1.4.243','ns_1@10.1.4.244'], ['ns_1@10.1.4.243','ns_1@10.1.4.244'], ['ns_1@10.1.4.243','ns_1@10.1.4.244'], ['ns_1@10.1.4.243','ns_1@10.1.4.244'], ['ns_1@10.1.4.243','ns_1@10.1.4.244'], ['ns_1@10.1.4.243','ns_1@10.1.4.244'], ['ns_1@10.1.4.243','ns_1@10.1.4.244'], ['ns_1@10.1.4.243','ns_1@10.1.4.244'], ['ns_1@10.1.4.243','ns_1@10.1.4.244'], ['ns_1@10.1.4.243','ns_1@10.1.4.244'], ['ns_1@10.1.4.243','ns_1@10.1.4.244'], ['ns_1@10.1.4.243','ns_1@10.1.4.244'], ['ns_1@10.1.4.243','ns_1@10.1.4.244'], ['ns_1@10.1.4.243','ns_1@10.1.4.244'], ['ns_1@10.1.4.243','ns_1@10.1.4.244'], ['ns_1@10.1.4.243','ns_1@10.1.4.244'], ['ns_1@10.1.4.243','ns_1@10.1.4.244'], ['ns_1@10.1.4.243','ns_1@10.1.4.244'], ['ns_1@10.1.4.243','ns_1@10.1.4.244'], ['ns_1@10.1.4.243','ns_1@10.1.4.244'], ['ns_1@10.1.4.243','ns_1@10.1.4.244'], ['ns_1@10.1.4.243','ns_1@10.1.4.244'], ['ns_1@10.1.4.243','ns_1@10.1.4.244'], ['ns_1@10.1.4.243','ns_1@10.1.4.244'], ['ns_1@10.1.4.243','ns_1@10.1.4.244'], ['ns_1@10.1.4.243','ns_1@10.1.4.244'], ['ns_1@10.1.4.243','ns_1@10.1.4.244'], ['ns_1@10.1.4.243','ns_1@10.1.4.244'], ['ns_1@10.1.4.243','ns_1@10.1.4.244'], ['ns_1@10.1.4.243','ns_1@10.1.4.244'], ['ns_1@10.1.4.243','ns_1@10.1.4.244'], ['ns_1@10.1.4.243','ns_1@10.1.4.244'], ['ns_1@10.1.4.243','ns_1@10.1.4.244'], ['ns_1@10.1.4.243','ns_1@10.1.4.244'], ['ns_1@10.1.4.243','ns_1@10.1.4.244'], ['ns_1@10.1.4.243','ns_1@10.1.4.244'], ['ns_1@10.1.4.243','ns_1@10.1.4.244'], ['ns_1@10.1.4.243','ns_1@10.1.4.244'], ['ns_1@10.1.4.243','ns_1@10.1.4.244'], ['ns_1@10.1.4.243','ns_1@10.1.4.244'], ['ns_1@10.1.4.243','ns_1@10.1.4.244'], ['ns_1@10.1.4.243','ns_1@10.1.4.244'], ['ns_1@10.1.4.243','ns_1@10.1.4.244'], ['ns_1@10.1.4.243','ns_1@10.1.4.244'], ['ns_1@10.1.4.243','ns_1@10.1.4.244'], ['ns_1@10.1.4.243','ns_1@10.1.4.244'], ['ns_1@10.1.4.243','ns_1@10.1.4.244'], ['ns_1@10.1.4.243','ns_1@10.1.4.244'], ['ns_1@10.1.4.243','ns_1@10.1.4.244'], ['ns_1@10.1.4.243','ns_1@10.1.4.244'], ['ns_1@10.1.4.243','ns_1@10.1.4.244'], ['ns_1@10.1.4.243','ns_1@10.1.4.244'], ['ns_1@10.1.4.243','ns_1@10.1.4.244'], ['ns_1@10.1.4.243','ns_1@10.1.4.244'], ['ns_1@10.1.4.243','ns_1@10.1.4.244'], ['ns_1@10.1.4.243','ns_1@10.1.4.244'], ['ns_1@10.1.4.243','ns_1@10.1.4.244'], ['ns_1@10.1.4.243','ns_1@10.1.4.244'], ['ns_1@10.1.4.243','ns_1@10.1.4.244'], ['ns_1@10.1.4.243','ns_1@10.1.4.244'], ['ns_1@10.1.4.243','ns_1@10.1.4.244'], ['ns_1@10.1.4.243','ns_1@10.1.4.244'], ['ns_1@10.1.4.243','ns_1@10.1.4.244'], ['ns_1@10.1.4.243','ns_1@10.1.4.244'], ['ns_1@10.1.4.243','ns_1@10.1.4.244'], ['ns_1@10.1.4.243'|...], [...]|...]}]}]}] INFO REPORT <0.64.0> 2011-05-07 17:31:41 =============================================================================== ns_node_disco_conf_events config all INFO REPORT <0.85.0> 2011-05-07 17:31:41 =============================================================================== Pushing config INFO REPORT <0.85.0> 2011-05-07 17:31:41 =============================================================================== Pushing config done INFO REPORT <0.64.0> 2011-05-07 17:31:41 =============================================================================== config change: rebalance_status -> {none,<<"Rebalance failed. See logs for detailed reason. You can try rebalance again.">>} INFO REPORT <0.64.0> 2011-05-07 17:31:41 =============================================================================== ns_node_disco_conf_events config all INFO REPORT <0.85.0> 2011-05-07 17:31:41 =============================================================================== Pushing config INFO REPORT <0.85.0> 2011-05-07 17:31:41 =============================================================================== Pushing config done INFO REPORT <0.3386.0> 2011-05-07 17:31:43 =============================================================================== ns_1@10.1.4.243:ns_vbm_sup:267: Starting replicator for vbuckets [1023,1022,1021,1020,1019,1018,1017,1016,1015,1014,1013,1012,1011,1010,1009,1008,1007,1006,1005,1004,1003,1002,1001,1000,999,998,997,996,995,994,993,992,991,990,989,988,987,986,985,984,983,982,981,980,979,978,977,976,975,974,973,972,971,970,969,968,967,966,965,964,963,962,961,960,959,958,957,956,955,954,953,952,951,950,949,948,947,946,945,944,943,942,941,940,939,938,937,936,935,934,933,932,931,930,929,928,927,926,925,924,923,922,921,920,919,918,917,916,915,914,913,912,911,910,909,908,907,906,905,904,903,902,901,900,899,898,897,896,895,894,893,892,891,890,889,888,887,886,885,884,883,882,881,880,879,878,877,876,875,874,873,872,871,870,869,868,867,866,865,864,863,862,861,860,859,858,857,856,855,854,853,852,851,850,849,848,847,846,845,844,843,842,841,840,839,838,837,836,835,834,833,832,831,830,829,828,827,826,825,824,823,822,821,820,819,818,817,816,815,814,813,812,811,810,809,808,807,806,805,804,803,802,801,800,799,798,797,796,795,794,793,792,791,790,789,788,787,786,785,784,783,782,781,780,779,778,777,776,775,774,773,772,771,770,769,768,767,766,765,764,763,762,761,760,759,758,757,756,755,754,753,752,751,750,749,748,747,746,745,744,743,742,741,740,739,738,737,736,735,734,733,732,731,730,729,728,727,726,725,724,723,722,721,720,719,718,717,716,715,714,713,712,711,710,709,708,707,706,705,704,703,702,701,700,699,698,697,696,695,694,693,692,691,690,689,688,687,686,685,684,683,682,681,680,679,678,677,676,675,674,673,672,671,670,669,668,667,666,665,664,663,662,661,660,659,658,657,656,655,654,653,652,651,650,649,648,647,646,645,644,643,642,641,640,639,638,637,636,635,634,633,632,631,630,629,628,627,626,625,624,623,622,621,620,619,618,617,616,615,614,613,612,611,610,609,608,607,606,605,604,603,602,601,600,599,598,597,596,595,594,593,592,591,590,589,588,587,586,585,584,583,582,581,580,579,578,577,576,575,574,573,572,571,570,569,568,567,566,565,564,563,562,561,560,559,558,557,556,555,554,553,552,551,550,549,548,547,546,545,544,543,542,541,540,539,538,537,536,535,534,533,532,531,530,529,528,527,526,525,524,523,522,521,520,519,518,517,516,515,514,513,512,511,510,509,508,507,506,505,504,503,502,501,500,499,498,497,496,495,494,493,492,491,490,489,488,487,486,485,484,483,482,481,480,479,478,477,476,475,474,473,472,471,470,469,468,467,466,465,464,463,462,461,460,459,458,457,456,455,454,453,452,451,450,449,448,447,446,445,444,443,442,441,440,439,438,437,436,435,434,433,432,431,430,429,428,427,426,425,424,423,422,421,420,419,418,417,416,415,414,413,412,411,410,409,408,407,406,405,404,403,402,401,400,399,398,397,396,395,394,393,392,391,390,389,388,387,386,385,384,383,382,381,380,379,378,377,376,375,374,373,372,371,370,369,368,367,366,365,364,363,362,361,360,359,358,357,356,355,354,353,352,351,350,349,348,347,346,345,344,343,342,341,340,339,338,337,336,335,334,333,332,331,330,329,328,327,326,325,324,323,322,321,320,319,318,317,316,315,314,313,312,311,310,309,308,307,306,305,304,303,302,301,300,299,298,297,296,295,294,293,292,291,290,289,288,287,286,285,284,283,282,281,280,279,278,277,276,275,274,273,272,271,270,269,268,267,266,265,264,263,262,261,260,259,258,257,256,255,254,253,252,251,250,249,248,247,246,245,244,243,242,241,240,239,238,237,236,235,234,233,232,231,230,229,228,227,226,225,224,223,222,221,220,219,218,217,216,215,214,213,212,211,210,209,208,207,206,205,204,203,202,201,200,199,198,197,196,195,194,193,192,191,190,189,188,187,186,185,184,183,182,181,180,179,178,177,176,175,174,173,172,171,170,169,168,167,166,165,164,163,162,161,160,159,158,157,156,155,154,153,152,151,150,149,148,147,146,145,144,143,142,141,140,139,138,137,136,135,134,133,132,131,130,129,128,127,126,125,124,123,122,121,120,119,118,117,116,115,114,113,112,111,110,109,108,107,106,105,104,103,102,101,100,99,98,97,96,95,94,93,92,91,90,89,88,87,86,85,84,83,82,81,80,79,78,77,76,75,74,73,72,71,70,69,68,67,66,65,64,63,62,61,60,59,58,57,56,55,54,53,52,51,50,49,48,47,46,45,44,43,42,41,40,39,38,37,36,35,34,33,32,31,30,29,28,27,26,25,24,23,22,21,20,19,18,17,16,15,14,13,12,11,10,9,8,7,6,5,4,3,2,1] in bucket "default" from node 'ns_1@10.1.4.243' to node 'ns_1@10.1.4.244' INFO REPORT <0.64.0> 2011-05-07 17:31:43 =============================================================================== config change: buckets -> [{configs,[{"default", [{num_replicas,1}, {ram_quota,839909376}, {auth_type,sasl}, {sasl_password,[]}, {type,membase}, {num_vbuckets,1024}, {db_shards,4}, {ht_size,3079}, {tap_keepalive,0}, {tap_noop_interval,20}, {max_txn_size,1000}, {ht_locks,5}, {servers,['ns_1@10.1.4.243','ns_1@10.1.4.244']}, {map,[['ns_1@10.1.4.244',undefined], ['ns_1@10.1.4.243','ns_1@10.1.4.244'], ['ns_1@10.1.4.243','ns_1@10.1.4.244'], ['ns_1@10.1.4.243','ns_1@10.1.4.244'], ['ns_1@10.1.4.243','ns_1@10.1.4.244'], ['ns_1@10.1.4.243','ns_1@10.1.4.244'], ['ns_1@10.1.4.243','ns_1@10.1.4.244'], ['ns_1@10.1.4.243','ns_1@10.1.4.244'], ['ns_1@10.1.4.243','ns_1@10.1.4.244'], ['ns_1@10.1.4.243','ns_1@10.1.4.244'], ['ns_1@10.1.4.243','ns_1@10.1.4.244'], ['ns_1@10.1.4.243','ns_1@10.1.4.244'], ['ns_1@10.1.4.243','ns_1@10.1.4.244'], ['ns_1@10.1.4.243','ns_1@10.1.4.244'], ['ns_1@10.1.4.243','ns_1@10.1.4.244'], ['ns_1@10.1.4.243','ns_1@10.1.4.244'], ['ns_1@10.1.4.243','ns_1@10.1.4.244'], ['ns_1@10.1.4.243','ns_1@10.1.4.244'], ['ns_1@10.1.4.243','ns_1@10.1.4.244'], ['ns_1@10.1.4.243','ns_1@10.1.4.244'], ['ns_1@10.1.4.243','ns_1@10.1.4.244'], ['ns_1@10.1.4.243','ns_1@10.1.4.244'], ['ns_1@10.1.4.243','ns_1@10.1.4.244'], ['ns_1@10.1.4.243','ns_1@10.1.4.244'], ['ns_1@10.1.4.243','ns_1@10.1.4.244'], ['ns_1@10.1.4.243','ns_1@10.1.4.244'], ['ns_1@10.1.4.243','ns_1@10.1.4.244'], ['ns_1@10.1.4.243','ns_1@10.1.4.244'], ['ns_1@10.1.4.243','ns_1@10.1.4.244'], ['ns_1@10.1.4.243','ns_1@10.1.4.244'], ['ns_1@10.1.4.243','ns_1@10.1.4.244'], ['ns_1@10.1.4.243','ns_1@10.1.4.244'], ['ns_1@10.1.4.243','ns_1@10.1.4.244'], ['ns_1@10.1.4.243','ns_1@10.1.4.244'], ['ns_1@10.1.4.243','ns_1@10.1.4.244'], ['ns_1@10.1.4.243','ns_1@10.1.4.244'], ['ns_1@10.1.4.243','ns_1@10.1.4.244'], ['ns_1@10.1.4.243','ns_1@10.1.4.244'], ['ns_1@10.1.4.243','ns_1@10.1.4.244'], ['ns_1@10.1.4.243','ns_1@10.1.4.244'], ['ns_1@10.1.4.243','ns_1@10.1.4.244'], ['ns_1@10.1.4.243','ns_1@10.1.4.244'], ['ns_1@10.1.4.243','ns_1@10.1.4.244'], ['ns_1@10.1.4.243','ns_1@10.1.4.244'], ['ns_1@10.1.4.243','ns_1@10.1.4.244'], ['ns_1@10.1.4.243','ns_1@10.1.4.244'], ['ns_1@10.1.4.243','ns_1@10.1.4.244'], ['ns_1@10.1.4.243','ns_1@10.1.4.244'], ['ns_1@10.1.4.243','ns_1@10.1.4.244'], ['ns_1@10.1.4.243','ns_1@10.1.4.244'], ['ns_1@10.1.4.243','ns_1@10.1.4.244'], ['ns_1@10.1.4.243','ns_1@10.1.4.244'], ['ns_1@10.1.4.243','ns_1@10.1.4.244'], ['ns_1@10.1.4.243','ns_1@10.1.4.244'], ['ns_1@10.1.4.243','ns_1@10.1.4.244'], ['ns_1@10.1.4.243','ns_1@10.1.4.244'], ['ns_1@10.1.4.243','ns_1@10.1.4.244'], ['ns_1@10.1.4.243','ns_1@10.1.4.244'], ['ns_1@10.1.4.243','ns_1@10.1.4.244'], ['ns_1@10.1.4.243','ns_1@10.1.4.244'], ['ns_1@10.1.4.243','ns_1@10.1.4.244'], ['ns_1@10.1.4.243','ns_1@10.1.4.244'], ['ns_1@10.1.4.243','ns_1@10.1.4.244'], ['ns_1@10.1.4.243','ns_1@10.1.4.244'], ['ns_1@10.1.4.243','ns_1@10.1.4.244'], ['ns_1@10.1.4.243','ns_1@10.1.4.244'], ['ns_1@10.1.4.243','ns_1@10.1.4.244'], ['ns_1@10.1.4.243','ns_1@10.1.4.244'], ['ns_1@10.1.4.243','ns_1@10.1.4.244'], ['ns_1@10.1.4.243','ns_1@10.1.4.244'], ['ns_1@10.1.4.243','ns_1@10.1.4.244'], ['ns_1@10.1.4.243','ns_1@10.1.4.244'], ['ns_1@10.1.4.243','ns_1@10.1.4.244'], ['ns_1@10.1.4.243','ns_1@10.1.4.244'], ['ns_1@10.1.4.243','ns_1@10.1.4.244'], ['ns_1@10.1.4.243'|...], [...]|...]}]}]}] INFO REPORT <0.64.0> 2011-05-07 17:31:43 =============================================================================== ns_node_disco_conf_events config all INFO REPORT <0.85.0> 2011-05-07 17:31:43 =============================================================================== Pushing config INFO REPORT <0.85.0> 2011-05-07 17:31:43 =============================================================================== Pushing config done INFO REPORT <0.2211.0> 2011-05-07 17:31:43 =============================================================================== ns_1@10.1.4.243:ns_doctor:82: Current node statuses: [{'ns_1@10.1.4.243', [{last_heard,{1304,814703,60369}}, {active_buckets,["default"]}, {memory, [{total,25538312}, {processes,14589692}, {processes_used,14575068}, {system,10948620}, {atom,536281}, {atom_used,521993}, {binary,175064}, {code,4290757}, {ets,988500}]}, {cluster_compatibility_version,1}, {version, [{kernel,"2.13.4"}, {mnesia,"4.4.12"}, {stdlib,"1.16.4"}, {os_mon,"2.2.4"}, {inets,"5.2"}, {ns_server,"1.6.5.4r"}, {menelaus,"1.6.5.4r"}, {sasl,"2.1.8"}]}, {system_arch,"i686-pc-linux-gnu"}, {wall_clock,1600}, {memory_data,{1050726400,792924160,{<0.3052.0>,3328660}}}, {disk_data, [{"/",7583436,18}, {"/dev",508460,1}, {"/dev/shm",513048,0}, {"/var/run",513048,1}, {"/var/lock",513048,0}, {"/lib/init/rw",513048,0}, {"/boot",233191,8}]}, {meminfo, <<"MemTotal: 1026100 kB\nMemFree: 254124 kB\nBuffers: 152772 kB\nCached: 451272 kB\nSwapCached: 0 kB\nActive: 455268 kB\nInactive: 278164 kB\nActive(anon): 125216 kB\nInactive(anon): 4412 kB\nActive(file): 330052 kB\nInactive(file): 273752 kB\nUnevictable: 0 kB\nMlocked: 0 kB\nHighTotal: 137160 kB\nHighFree: 252 kB\nLowTotal: 888940 kB\nLowFree: 253872 kB\nSwapTotal: 397304 kB\nSwapFree: 397304 kB\nDirty: 84 kB\nWriteback: 0 kB\nAnonPages: 129388 kB\nMapped: 10532 kB\nShmem: 240 kB\nSlab: 28196 kB\nSReclaimable: 23292 kB\nSUnreclaim: 4904 kB\nKernelStack: 976 kB\nPageTables: 1112 kB\nNFS_Unstable: 0 kB\nBounce: 0 kB\nWritebackTmp: 0 kB\nCommitLimit: 910352 kB\nCommitted_AS: 349160 kB\nVmallocTotal: 122880 kB\nVmallocUsed: 3540 kB\nVmallocChunk: 114424 kB\nHardwareCorrupted: 0 kB\nHugePages_Total: 0\nHugePages_Free: 0\nHugePages_Rsvd: 0\nHugePages_Surp: 0\nHugepagesize: 2048 kB\nDirectMap4k: 10232 kB\nDirectMap2M: 901120 kB\n">>}, {replication,[{"default",1.0}]}, {system_memory_data, [{system_total_memory,1050726400}, {free_swap,406839296}, {total_swap,406839296}, {cached_memory,462102528}, {buffered_memory,156438528}, {free_memory,260222976}, {total_memory,1050726400}]}, {statistics, [{wall_clock,{1590410,1}}, {context_switches,{333239,0}}, {garbage_collection,{73480,430116665,0}}, {io,{{input,19062005},{output,23009269}}}, {reductions,{268125063,1041584}}, {run_queue,0}, {runtime,{26930,170}}]}]}, {'ns_1@10.1.4.244', [{last_heard,{1304,814703,110170}}, {active_buckets,["default"]}, {memory, [{total,14290192}, {processes,4236916}, {processes_used,4232676}, {system,10053276}, {atom,515245}, {atom_used,490164}, {binary,274136}, {code,3969328}, {ets,391972}]}, {cluster_compatibility_version,1}, {version, [{kernel,"2.13.4"}, {mnesia,"4.4.12"}, {stdlib,"1.16.4"}, {os_mon,"2.2.4"}, {ns_server,"1.6.5.4r"}, {menelaus,"1.6.5.4r"}, {sasl,"2.1.8"}]}, {system_arch,"i686-pc-linux-gnu"}, {wall_clock,103}, {memory_data,{1050726400,877510656,{<6517.183.0>,901104}}}, {disk_data, [{"/",7583436,16}, {"/dev",508460,1}, {"/dev/shm",513048,0}, {"/var/run",513048,1}, {"/var/lock",513048,0}, {"/lib/init/rw",513048,0}, {"/boot",233191,8}]}, {meminfo, <<"MemTotal: 1026100 kB\nMemFree: 204036 kB\nBuffers: 21012 kB\nCached: 244832 kB\nSwapCached: 0 kB\nActive: 157492 kB\nInactive: 162900 kB\nActive(anon): 48628 kB\nInactive(anon): 6152 kB\nActive(file): 108864 kB\nInactive(file): 156748 kB\nUnevictable: 0 kB\nMlocked: 0 kB\nHighTotal: 137160 kB\nHighFree: 756 kB\nLowTotal: 888940 kB\nLowFree: 203280 kB\nSwapTotal: 397304 kB\nSwapFree: 397304 kB\nDirty: 112 kB\nWriteback: 0 kB\nAnonPages: 54604 kB\nMapped: 10028 kB\nShmem: 232 kB\nSlab: 21512 kB\nSReclaimable: 16300 kB\nSUnreclaim: 5212 kB\nKernelStack: 928 kB\nPageTables: 852 kB\nNFS_Unstable: 0 kB\nBounce: 0 kB\nWritebackTmp: 0 kB\nCommitLimit: 910352 kB\nCommitted_AS: 255368 kB\nVmallocTotal: 122880 kB\nVmallocUsed: 3540 kB\nVmallocChunk: 114844 kB\nHardwareCorrupted: 0 kB\nHugePages_Total: 0\nHugePages_Free: 0\nHugePages_Rsvd: 0\nHugePages_Surp: 0\nHugepagesize: 2048 kB\nDirectMap4k: 10232 kB\nDirectMap2M: 901120 kB\n">>}, {replication,[{"default",1.0}]}, {system_memory_data, [{system_total_memory,1050726400}, {free_swap,406839296}, {total_swap,406839296}, {cached_memory,250707968}, {buffered_memory,21516288}, {free_memory,208932864}, {total_memory,1050726400}]}, {statistics, [{wall_clock,{93021,1}}, {context_switches,{21355,0}}, {garbage_collection,{4903,11630509,0}}, {io,{{input,5473300},{output,1697750}}}, {reductions,{4335640,246083}}, {run_queue,0}, {runtime,{890,50}}]}]}] INFO REPORT <0.3386.0> 2011-05-07 17:31:44 =============================================================================== ns_1@10.1.4.243:ns_vbm_sup:255: Args = [vbucketmigrator,"./bin/vbucketmigrator/vbucketmigrator", ["-e","-a","default","-h","10.1.4.243:11210","-d","10.1.4.244:11210","-A", "-N","r-ns_1@10.1.4.244-ns_1@10.1.4.243-1304814704.103784","-v","-b","511", "-b","510","-b","509","-b","508","-b","507","-b","506","-b","505","-b", "504","-b","503","-b","502","-b","501","-b","500","-b","499","-b","498", "-b","497","-b","496","-b","495","-b","494","-b","493","-b","492","-b", "491","-b","490","-b","489","-b","488","-b","487","-b","486","-b","485", "-b","484","-b","483","-b","482","-b","481","-b","480","-b","479","-b", "478","-b","477","-b","476","-b","475","-b","474","-b","473","-b","472", "-b","471","-b","470","-b","469","-b","468","-b","467","-b","466","-b", "465","-b","464","-b","463","-b","462","-b","461","-b","460","-b","459", "-b","458","-b","457","-b","456","-b","455","-b","454","-b","453","-b", "452","-b","451","-b","450","-b","449","-b","448","-b","447","-b","446", "-b","445","-b","444","-b","443","-b","442","-b","441","-b","440","-b", "439","-b","438","-b","437","-b","436","-b","435","-b","434","-b","433", "-b","432","-b","431","-b","430","-b","429","-b","428","-b","427","-b", "426","-b","425","-b","424","-b","423","-b","422","-b","421","-b","420", "-b","419","-b","418","-b","417","-b","416","-b","415","-b","414","-b", "413","-b","412","-b","411","-b","410","-b","409","-b","408","-b","407", "-b","406","-b","405","-b","404","-b","403","-b","402","-b","401","-b", "400","-b","399","-b","398","-b","397","-b","396","-b","395","-b","394", "-b","393","-b","392","-b","391","-b","390","-b","389","-b","388","-b", "387","-b","386","-b","385","-b","384","-b","383","-b","382","-b","381", "-b","380","-b","379","-b","378","-b","377","-b","376","-b","375","-b", "374","-b","373","-b","372","-b","371","-b","370","-b","369","-b","368", "-b","367","-b","366","-b","365","-b","364","-b","363","-b","362","-b", "361","-b","360","-b","359","-b","358","-b","357","-b","356","-b","355", "-b","354","-b","353","-b","352","-b","351","-b","350","-b","349","-b", "348","-b","347","-b","346","-b","345","-b","344","-b","343","-b","342", "-b","341","-b","340","-b","339","-b","338","-b","337","-b","336","-b", "335","-b","334","-b","333","-b","332","-b","331","-b","330","-b","329", "-b","328","-b","327","-b","326","-b","325","-b","324","-b","323","-b", "322","-b","321","-b","320","-b","319","-b","318","-b","317","-b","316", "-b","315","-b","314","-b","313","-b","312","-b","311","-b","310","-b", "309","-b","308","-b","307","-b","306","-b","305","-b","304","-b","303", "-b","302","-b","301","-b","300","-b","299","-b","298","-b","297","-b", "296","-b","295","-b","294","-b","293","-b","292","-b","291","-b","290", "-b","289","-b","288","-b","287","-b","286","-b","285","-b","284","-b", "283","-b","282","-b","281","-b","280","-b","279","-b","278","-b","277", "-b","276","-b","275","-b","274","-b","273","-b","272","-b","271","-b", "270","-b","269","-b","268","-b","267","-b","266","-b","265","-b","264", "-b","263","-b","262","-b","261","-b","260","-b","259","-b","258","-b", "257","-b","256","-b","255","-b","254","-b","253","-b","252","-b","251", "-b","250","-b","249","-b","248","-b","247","-b","246","-b","245","-b", "244","-b","243","-b","242","-b","241","-b","240","-b","239","-b","238", "-b","237","-b","236","-b","235","-b","234","-b","233","-b","232","-b", "231","-b","230","-b","229","-b","228","-b","227","-b","226","-b","225", "-b","224","-b","223","-b","222","-b","221","-b","220","-b","219","-b", "218","-b","217","-b","216","-b","215","-b","214","-b","213","-b","212", "-b","211","-b","210","-b","209","-b","208","-b","207","-b","206","-b", "205","-b","204","-b","203","-b","202","-b","201","-b","200","-b","199", "-b","198","-b","197","-b","196","-b","195","-b","194","-b","193","-b", "192","-b","191","-b","190","-b","189","-b","188","-b","187","-b","186", "-b","185","-b","184","-b","183","-b","182","-b","181","-b","180","-b", "179","-b","178","-b","177","-b","176","-b","175","-b","174","-b","173", "-b","172","-b","171","-b","170","-b","169","-b","168","-b","167","-b", "166","-b","165","-b","164","-b","163","-b","162","-b","161","-b","160", "-b","159","-b","158","-b","157","-b","156","-b","155","-b","154","-b", "153","-b","152","-b","151","-b","150","-b","149","-b","148","-b","147", "-b","146","-b","145","-b","144","-b","143","-b","142","-b","141","-b", "140","-b","139","-b","138","-b","137","-b","136","-b","135","-b","134", "-b","133","-b","132","-b","131","-b","130","-b","129","-b","128","-b", "127","-b","126","-b","125","-b","124","-b","123","-b","122","-b","121", "-b","120","-b","119","-b","118","-b","117","-b","116","-b","115","-b", "114","-b","113","-b","112","-b","111","-b","110","-b","109","-b","108", "-b","107","-b","106","-b","105","-b","104","-b","103","-b","102","-b", "101","-b","100","-b","99","-b","98","-b","97","-b","96","-b","95","-b", "94","-b","93","-b","92","-b","91","-b","90","-b","89","-b","88","-b","87", "-b","86","-b","85","-b","84","-b","83","-b","82","-b","81","-b","80","-b", "79","-b","78","-b","77","-b","76","-b","75","-b","74","-b","73","-b","72", "-b","71","-b","70","-b","69","-b","68","-b","67","-b","66","-b","65","-b", "64","-b","63","-b","62","-b","61","-b","60","-b","59","-b","58","-b","57", "-b","56","-b","55","-b","54","-b","53","-b","52","-b","51","-b","50","-b", "49","-b","48","-b","47","-b","46","-b","45","-b","44","-b","43","-b","42", "-b","41","-b","40","-b","39","-b","38","-b","37","-b","36","-b","35","-b", "34","-b","33","-b","32","-b","31","-b","30","-b","29","-b","28","-b","27", "-b","26","-b","25","-b","24","-b","23","-b","22","-b","21","-b","20","-b", "19","-b","18","-b","17","-b","16","-b","15","-b","14","-b","13","-b","12", "-b","11","-b","10","-b","9","-b","8","-b","7","-b","6","-b","5","-b","4", "-b","3","-b","2","-b","1"], [use_stdio,stderr_to_stdout,{write_data,[[],"\n"]}]] INFO REPORT <0.3408.0> 2011-05-07 17:31:44 =============================================================================== starting ns_port_server with delay of 5000 PROGRESS REPORT <0.226.0> 2011-05-07 17:31:44 =============================================================================== supervisor {local,'ns_vbm_sup-default'} started [{pid,<0.3408.0>}, {name, {child_id, [511,510,509,508,507,506,505,504,503,502,501,500,499,498, 497,496,495,494,493,492,491,490,489,488,487,486,485,484, 483,482,481,480,479,478,477,476,475,474,473,472,471,470, 469,468,467,466,465,464,463,462,461,460,459,458,457,456, 455,454,453,452,451,450,449,448,447,446,445,444,443,442, 441,440,439,438,437,436,435,434,433,432,431,430,429,428, 427,426,425,424,423,422,421,420,419,418,417,416,415,414, 413,412,411,410,409,408,407,406,405,404,403,402,401,400, 399,398,397,396,395,394,393,392,391,390,389,388,387,386, 385,384,383,382,381,380,379,378,377,376,375,374,373,372, 371,370,369,368,367,366,365,364,363,362,361,360,359,358, 357,356,355,354,353,352,351,350,349,348,347,346,345,344, 343,342,341,340,339,338,337,336,335,334,333,332,331,330, 329,328,327,326,325,324,323,322,321,320,319,318,317,316, 315,314,313,312,311,310,309,308,307,306,305,304,303,302, 301,300,299,298,297,296,295,294,293,292,291,290,289,288, 287,286,285,284,283,282,281,280,279,278,277,276,275,274, 273,272,271,270,269,268,267,266,265,264,263,262,261,260, 259,258,257,256,255,254,253,252,251,250,249,248,247,246, 245,244,243,242,241,240,239,238,237,236,235,234,233,232, 231,230,229,228,227,226,225,224,223,222,221,220,219,218, 217,216,215,214,213,212,211,210,209,208,207,206,205,204, 203,202,201,200,199,198,197,196,195,194,193,192,191,190, 189,188,187,186,185,184,183,182,181,180,179,178,177,176, 175,174,173,172,171,170,169,168,167,166,165,164,163,162, 161,160,159,158,157,156,155,154,153,152,151,150,149,148, 147,146,145,144,143,142,141,140,139,138,137,136,135,134, 133,132,131,130,129,128,127,126,125,124,123,122,121,120, 119,118,117,116,115,114,113,112,111,110,109,108,107,106, 105,104,103,102,101,100,99,98,97,96,95,94,93,92,91,90,89, 88,87,86,85,84,83,82,81,80,79,78,77,76,75,74,73,72,71,70, 69,68,67,66,65,64,63,62,61,60,59,58,57,56,55,54,53,52,51, 50,49,48,47,46,45,44,43,42,41,40,39,38,37,36,35,34,33,32, 31,30,29,28,27,26,25,24,23,22,21,20,19,18,17,16,15,14,13, 12,11,10,9,8,7,6,5,4,3,2,1], 'ns_1@10.1.4.244'}}, {mfa, {supervisor_cushion,start_link, [{child_id, [511,510,509,508,507,506,505,504,503,502,501,500,499, 498,497,496,495,494,493,492,491,490,489,488,487,486, 485,484,483,482,481,480,479,478,477,476,475,474,473, 472,471,470,469,468,467,466,465,464,463,462,461,460, 459,458,457,456,455,454,453,452,451,450,449,448,447, 446,445,444,443,442,441,440,439,438,437,436,435,434, 433,432,431,430,429,428,427,426,425,424,423,422,421, 420,419,418,417,416,415,414,413,412,411,410,409,408, 407,406,405,404,403,402,401,400,399,398,397,396,395, 394,393,392,391,390,389,388,387,386,385,384,383,382, 381,380,379,378,377,376,375,374,373,372,371,370,369, 368,367,366,365,364,363,362,361,360,359,358,357,356, 355,354,353,352,351,350,349,348,347,346,345,344,343, 342,341,340,339,338,337,336,335,334,333,332,331,330, 329,328,327,326,325,324,323,322,321,320,319,318,317, 316,315,314,313,312,311,310,309,308,307,306,305,304, 303,302,301,300,299,298,297,296,295,294,293,292,291, 290,289,288,287,286,285,284,283,282,281,280,279,278, 277,276,275,274,273,272,271,270,269,268,267,266,265, 264,263,262,261,260,259,258,257,256,255,254,253,252, 251,250,249,248,247,246,245,244,243,242,241,240,239, 238,237,236,235,234,233,232,231,230,229,228,227,226, 225,224,223,222,221,220,219,218,217,216,215,214,213, 212,211,210,209,208,207,206,205,204,203,202,201,200, 199,198,197,196,195,194,193,192,191,190,189,188,187, 186,185,184,183,182,181,180,179,178,177,176,175,174, 173,172,171,170,169,168,167,166,165,164,163,162,161, 160,159,158,157,156,155,154,153,152,151,150,149,148, 147,146,145,144,143,142,141,140,139,138,137,136,135, 134,133,132,131,130,129,128,127,126,125,124,123,122, 121,120,119,118,117,116,115,114,113,112,111,110,109, 108,107,106,105,104,103,102,101,100,99,98,97,96,95,94, 93,92,91,90,89,88,87,86,85,84,83,82,81,80,79,78,77,76, 75,74,73,72,71,70,69,68,67,66,65,64,63,62,61,60,59,58, 57,56,55,54,53,52,51,50,49,48,47,46,45,44,43,42,41,40, 39,38,37,36,35,34,33,32,31,30,29,28,27,26,25,24,23,22, 21,20,19,18,17,16,15,14,13,12,11,10,9,8,7,6,5,4,3,2,1], 'ns_1@10.1.4.244'}, 5000,ns_port_server,start_link, [vbucketmigrator,"./bin/vbucketmigrator/vbucketmigrator", ["-e","-a","default","-h","10.1.4.243:11210","-d", "10.1.4.244:11210","-A","-N", "r-ns_1@10.1.4.244-ns_1@10.1.4.243-1304814704.103784", "-v","-b","511","-b","510","-b","509","-b","508","-b", "507","-b","506","-b","505","-b","504","-b","503","-b", "502","-b","501","-b","500","-b","499","-b","498","-b", "497","-b","496","-b","495","-b","494","-b","493","-b", "492","-b","491","-b","490","-b","489","-b","488","-b", "487","-b","486","-b","485","-b","484","-b","483","-b", "482","-b","481","-b","480","-b","479","-b","478","-b", "477","-b","476","-b","475","-b","474","-b","473","-b", "472","-b","471","-b","470","-b","469","-b","468","-b", "467","-b","466","-b","465","-b","464","-b","463","-b", "462","-b","461","-b","460","-b","459","-b","458","-b", "457","-b","456","-b","455","-b","454","-b","453","-b", "452","-b","451","-b","450","-b","449","-b","448","-b", "447","-b","446","-b","445","-b","444","-b","443","-b", "442","-b","441","-b","440","-b","439","-b","438","-b", "437","-b","436","-b","435","-b","434","-b","433","-b", "432","-b","431","-b","430","-b","429","-b","428","-b", "427","-b","426","-b","425","-b","424","-b","423","-b", "422","-b","421","-b","420","-b","419","-b","418","-b", "417","-b","416","-b","415","-b","414","-b","413","-b", "412","-b","411","-b","410","-b","409","-b","408","-b", "407","-b","406","-b","405","-b","404","-b","403","-b", "402","-b","401","-b","400","-b","399","-b","398","-b", "397","-b","396","-b","395","-b","394","-b","393","-b", "392","-b","391","-b","390","-b","389","-b","388","-b", "387","-b","386","-b","385","-b","384","-b","383","-b", "382","-b","381","-b","380","-b","379","-b","378","-b", "377","-b","376","-b","375","-b","374","-b","373","-b", "372","-b","371","-b","370","-b","369","-b","368","-b", "367","-b","366","-b","365","-b","364","-b","363","-b", "362","-b","361","-b","360","-b","359","-b","358","-b", "357","-b","356","-b","355","-b","354","-b","353","-b", "352","-b","351","-b","350","-b","349","-b","348","-b", "347","-b","346","-b","345","-b","344","-b","343","-b", "342","-b","341","-b","340","-b","339","-b","338","-b", "337","-b","336","-b","335","-b","334","-b","333","-b", "332","-b","331","-b","330","-b","329","-b","328","-b", "327","-b","326","-b","325","-b","324","-b","323","-b", "322","-b","321","-b","320","-b","319","-b","318","-b", "317","-b","316","-b","315","-b","314","-b","313","-b", "312","-b","311","-b","310","-b","309","-b","308","-b", "307","-b","306","-b","305","-b","304","-b","303","-b", "302","-b","301","-b","300","-b","299","-b","298","-b", "297","-b","296","-b","295","-b","294","-b","293","-b", "292","-b","291","-b","290","-b","289","-b","288","-b", "287","-b","286","-b","285","-b","284","-b","283","-b", "282","-b","281","-b","280","-b","279","-b","278","-b", "277","-b","276","-b","275","-b","274","-b","273","-b", "272","-b","271","-b","270","-b","269","-b","268","-b", "267","-b","266","-b","265","-b","264","-b","263","-b", "262","-b","261","-b","260","-b","259","-b","258","-b", "257","-b","256","-b","255","-b","254","-b","253","-b", "252","-b","251","-b","250","-b","249","-b","248","-b", "247","-b","246","-b","245","-b","244","-b","243","-b", "242","-b","241","-b","240","-b","239","-b","238","-b", "237","-b","236","-b","235","-b","234","-b","233","-b", "232","-b","231","-b","230","-b","229","-b","228","-b", "227","-b","226","-b","225","-b","224","-b","223","-b", "222","-b","221","-b","220","-b","219","-b","218","-b", "217","-b","216","-b","215","-b","214","-b","213","-b", "212","-b","211","-b","210","-b","209","-b","208","-b", "207","-b","206","-b","205","-b","204","-b","203","-b", "202","-b","201","-b","200","-b","199","-b","198","-b", "197","-b","196","-b","195","-b","194","-b","193","-b", "192","-b","191","-b","190","-b","189","-b","188","-b", "187","-b","186","-b","185","-b","184","-b","183","-b", "182","-b","181","-b","180","-b","179","-b","178","-b", "177","-b","176","-b","175","-b","174","-b","173","-b", "172","-b","171","-b","170","-b","169","-b","168","-b", "167","-b","166","-b","165","-b","164","-b","163","-b", "162","-b","161","-b","160","-b","159","-b","158","-b", "157","-b","156","-b","155","-b","154","-b","153","-b", "152","-b","151","-b","150","-b","149","-b","148","-b", "147","-b","146","-b","145","-b","144","-b","143","-b", "142","-b","141","-b","140","-b","139","-b","138","-b", "137","-b","136","-b","135","-b","134","-b","133","-b", "132","-b","131","-b","130","-b","129","-b","128","-b", "127","-b","126","-b","125","-b","124","-b","123","-b", "122","-b","121","-b","120","-b","119","-b","118","-b", "117","-b","116","-b","115","-b","114","-b","113","-b", "112","-b","111","-b","110","-b","109","-b","108","-b", "107","-b","106","-b","105","-b","104","-b","103","-b", "102","-b","101","-b","100","-b","99","-b","98","-b", "97","-b","96","-b","95","-b","94","-b","93","-b","92", "-b","91","-b","90","-b","89","-b","88","-b","87","-b", "86","-b","85","-b","84","-b","83","-b","82","-b","81", "-b","80","-b","79","-b","78","-b","77","-b","76","-b", "75","-b","74","-b","73","-b","72","-b","71","-b","70", "-b","69","-b","68","-b","67","-b","66","-b","65","-b", "64","-b","63","-b","62","-b","61","-b","60","-b","59", "-b","58","-b","57","-b","56","-b","55","-b","54","-b", "53","-b","52","-b","51","-b","50","-b","49","-b","48", "-b","47","-b","46","-b","45","-b","44","-b","43","-b", "42","-b","41","-b","40","-b","39","-b","38","-b","37", "-b","36","-b","35","-b","34","-b","33","-b","32","-b", "31","-b","30","-b","29","-b","28","-b","27","-b","26", "-b","25","-b","24","-b","23","-b","22","-b","21","-b", "20","-b","19","-b","18","-b","17","-b","16","-b","15", "-b","14","-b","13","-b","12","-b","11","-b","10","-b", "9","-b","8","-b","7","-b","6","-b","5","-b","4","-b", "3","-b","2","-b","1"], [use_stdio,stderr_to_stdout,{write_data,[[],"\n"]}]]]}}, {restart_type,permanent}, {shutdown,10}, {child_type,worker}] INFO REPORT <0.3386.0> 2011-05-07 17:31:44 =============================================================================== ns_1@10.1.4.243:ns_vbm_sup:255: Args = [vbucketmigrator,"./bin/vbucketmigrator/vbucketmigrator", ["-e","-a","default","-h","10.1.4.243:11210","-d","10.1.4.244:11210","-A", "-N","r-ns_1@10.1.4.244-ns_1@10.1.4.243-1304814704.105247","-v","-b","1023", "-b","1022","-b","1021","-b","1020","-b","1019","-b","1018","-b","1017", "-b","1016","-b","1015","-b","1014","-b","1013","-b","1012","-b","1011", "-b","1010","-b","1009","-b","1008","-b","1007","-b","1006","-b","1005", "-b","1004","-b","1003","-b","1002","-b","1001","-b","1000","-b","999","-b", "998","-b","997","-b","996","-b","995","-b","994","-b","993","-b","992", "-b","991","-b","990","-b","989","-b","988","-b","987","-b","986","-b", "985","-b","984","-b","983","-b","982","-b","981","-b","980","-b","979", "-b","978","-b","977","-b","976","-b","975","-b","974","-b","973","-b", "972","-b","971","-b","970","-b","969","-b","968","-b","967","-b","966", "-b","965","-b","964","-b","963","-b","962","-b","961","-b","960","-b", "959","-b","958","-b","957","-b","956","-b","955","-b","954","-b","953", "-b","952","-b","951","-b","950","-b","949","-b","948","-b","947","-b", "946","-b","945","-b","944","-b","943","-b","942","-b","941","-b","940", "-b","939","-b","938","-b","937","-b","936","-b","935","-b","934","-b", "933","-b","932","-b","931","-b","930","-b","929","-b","928","-b","927", "-b","926","-b","925","-b","924","-b","923","-b","922","-b","921","-b", "920","-b","919","-b","918","-b","917","-b","916","-b","915","-b","914", "-b","913","-b","912","-b","911","-b","910","-b","909","-b","908","-b", "907","-b","906","-b","905","-b","904","-b","903","-b","902","-b","901", "-b","900","-b","899","-b","898","-b","897","-b","896","-b","895","-b", "894","-b","893","-b","892","-b","891","-b","890","-b","889","-b","888", "-b","887","-b","886","-b","885","-b","884","-b","883","-b","882","-b", "881","-b","880","-b","879","-b","878","-b","877","-b","876","-b","875", "-b","874","-b","873","-b","872","-b","871","-b","870","-b","869","-b", "868","-b","867","-b","866","-b","865","-b","864","-b","863","-b","862", "-b","861","-b","860","-b","859","-b","858","-b","857","-b","856","-b", "855","-b","854","-b","853","-b","852","-b","851","-b","850","-b","849", "-b","848","-b","847","-b","846","-b","845","-b","844","-b","843","-b", "842","-b","841","-b","840","-b","839","-b","838","-b","837","-b","836", "-b","835","-b","834","-b","833","-b","832","-b","831","-b","830","-b", "829","-b","828","-b","827","-b","826","-b","825","-b","824","-b","823", "-b","822","-b","821","-b","820","-b","819","-b","818","-b","817","-b", "816","-b","815","-b","814","-b","813","-b","812","-b","811","-b","810", "-b","809","-b","808","-b","807","-b","806","-b","805","-b","804","-b", "803","-b","802","-b","801","-b","800","-b","799","-b","798","-b","797", "-b","796","-b","795","-b","794","-b","793","-b","792","-b","791","-b", "790","-b","789","-b","788","-b","787","-b","786","-b","785","-b","784", "-b","783","-b","782","-b","781","-b","780","-b","779","-b","778","-b", "777","-b","776","-b","775","-b","774","-b","773","-b","772","-b","771", "-b","770","-b","769","-b","768","-b","767","-b","766","-b","765","-b", "764","-b","763","-b","762","-b","761","-b","760","-b","759","-b","758", "-b","757","-b","756","-b","755","-b","754","-b","753","-b","752","-b", "751","-b","750","-b","749","-b","748","-b","747","-b","746","-b","745", "-b","744","-b","743","-b","742","-b","741","-b","740","-b","739","-b", "738","-b","737","-b","736","-b","735","-b","734","-b","733","-b","732", "-b","731","-b","730","-b","729","-b","728","-b","727","-b","726","-b", "725","-b","724","-b","723","-b","722","-b","721","-b","720","-b","719", "-b","718","-b","717","-b","716","-b","715","-b","714","-b","713","-b", "712","-b","711","-b","710","-b","709","-b","708","-b","707","-b","706", "-b","705","-b","704","-b","703","-b","702","-b","701","-b","700","-b", "699","-b","698","-b","697","-b","696","-b","695","-b","694","-b","693", "-b","692","-b","691","-b","690","-b","689","-b","688","-b","687","-b", "686","-b","685","-b","684","-b","683","-b","682","-b","681","-b","680", "-b","679","-b","678","-b","677","-b","676","-b","675","-b","674","-b", "673","-b","672","-b","671","-b","670","-b","669","-b","668","-b","667", "-b","666","-b","665","-b","664","-b","663","-b","662","-b","661","-b", "660","-b","659","-b","658","-b","657","-b","656","-b","655","-b","654", "-b","653","-b","652","-b","651","-b","650","-b","649","-b","648","-b", "647","-b","646","-b","645","-b","644","-b","643","-b","642","-b","641", "-b","640","-b","639","-b","638","-b","637","-b","636","-b","635","-b", "634","-b","633","-b","632","-b","631","-b","630","-b","629","-b","628", "-b","627","-b","626","-b","625","-b","624","-b","623","-b","622","-b", "621","-b","620","-b","619","-b","618","-b","617","-b","616","-b","615", "-b","614","-b","613","-b","612","-b","611","-b","610","-b","609","-b", "608","-b","607","-b","606","-b","605","-b","604","-b","603","-b","602", "-b","601","-b","600","-b","599","-b","598","-b","597","-b","596","-b", "595","-b","594","-b","593","-b","592","-b","591","-b","590","-b","589", "-b","588","-b","587","-b","586","-b","585","-b","584","-b","583","-b", "582","-b","581","-b","580","-b","579","-b","578","-b","577","-b","576", "-b","575","-b","574","-b","573","-b","572","-b","571","-b","570","-b", "569","-b","568","-b","567","-b","566","-b","565","-b","564","-b","563", "-b","562","-b","561","-b","560","-b","559","-b","558","-b","557","-b", "556","-b","555","-b","554","-b","553","-b","552","-b","551","-b","550", "-b","549","-b","548","-b","547","-b","546","-b","545","-b","544","-b", "543","-b","542","-b","541","-b","540","-b","539","-b","538","-b","537", "-b","536","-b","535","-b","534","-b","533","-b","532","-b","531","-b", "530","-b","529","-b","528","-b","527","-b","526","-b","525","-b","524", "-b","523","-b","522","-b","521","-b","520","-b","519","-b","518","-b", "517","-b","516","-b","515","-b","514","-b","513","-b","512"], [use_stdio,stderr_to_stdout,{write_data,[[],"\n"]}]] INFO REPORT <0.3410.0> 2011-05-07 17:31:44 =============================================================================== starting ns_port_server with delay of 5000 PROGRESS REPORT <0.226.0> 2011-05-07 17:31:44 =============================================================================== supervisor {local,'ns_vbm_sup-default'} started [{pid,<0.3410.0>}, {name, {child_id, [1023,1022,1021,1020,1019,1018,1017,1016,1015,1014,1013, 1012,1011,1010,1009,1008,1007,1006,1005,1004,1003,1002, 1001,1000,999,998,997,996,995,994,993,992,991,990,989,988, 987,986,985,984,983,982,981,980,979,978,977,976,975,974, 973,972,971,970,969,968,967,966,965,964,963,962,961,960, 959,958,957,956,955,954,953,952,951,950,949,948,947,946, 945,944,943,942,941,940,939,938,937,936,935,934,933,932, 931,930,929,928,927,926,925,924,923,922,921,920,919,918, 917,916,915,914,913,912,911,910,909,908,907,906,905,904, 903,902,901,900,899,898,897,896,895,894,893,892,891,890, 889,888,887,886,885,884,883,882,881,880,879,878,877,876, 875,874,873,872,871,870,869,868,867,866,865,864,863,862, 861,860,859,858,857,856,855,854,853,852,851,850,849,848, 847,846,845,844,843,842,841,840,839,838,837,836,835,834, 833,832,831,830,829,828,827,826,825,824,823,822,821,820, 819,818,817,816,815,814,813,812,811,810,809,808,807,806, 805,804,803,802,801,800,799,798,797,796,795,794,793,792, 791,790,789,788,787,786,785,784,783,782,781,780,779,778, 777,776,775,774,773,772,771,770,769,768,767,766,765,764, 763,762,761,760,759,758,757,756,755,754,753,752,751,750, 749,748,747,746,745,744,743,742,741,740,739,738,737,736, 735,734,733,732,731,730,729,728,727,726,725,724,723,722, 721,720,719,718,717,716,715,714,713,712,711,710,709,708, 707,706,705,704,703,702,701,700,699,698,697,696,695,694, 693,692,691,690,689,688,687,686,685,684,683,682,681,680, 679,678,677,676,675,674,673,672,671,670,669,668,667,666, 665,664,663,662,661,660,659,658,657,656,655,654,653,652, 651,650,649,648,647,646,645,644,643,642,641,640,639,638, 637,636,635,634,633,632,631,630,629,628,627,626,625,624, 623,622,621,620,619,618,617,616,615,614,613,612,611,610, 609,608,607,606,605,604,603,602,601,600,599,598,597,596, 595,594,593,592,591,590,589,588,587,586,585,584,583,582, 581,580,579,578,577,576,575,574,573,572,571,570,569,568, 567,566,565,564,563,562,561,560,559,558,557,556,555,554, 553,552,551,550,549,548,547,546,545,544,543,542,541,540, 539,538,537,536,535,534,533,532,531,530,529,528,527,526, 525,524,523,522,521,520,519,518,517,516,515,514,513,512], 'ns_1@10.1.4.244'}}, {mfa, {supervisor_cushion,start_link, [{child_id, [1023,1022,1021,1020,1019,1018,1017,1016,1015,1014, 1013,1012,1011,1010,1009,1008,1007,1006,1005,1004, 1003,1002,1001,1000,999,998,997,996,995,994,993,992, 991,990,989,988,987,986,985,984,983,982,981,980,979, 978,977,976,975,974,973,972,971,970,969,968,967,966, 965,964,963,962,961,960,959,958,957,956,955,954,953, 952,951,950,949,948,947,946,945,944,943,942,941,940, 939,938,937,936,935,934,933,932,931,930,929,928,927, 926,925,924,923,922,921,920,919,918,917,916,915,914, 913,912,911,910,909,908,907,906,905,904,903,902,901, 900,899,898,897,896,895,894,893,892,891,890,889,888, 887,886,885,884,883,882,881,880,879,878,877,876,875, 874,873,872,871,870,869,868,867,866,865,864,863,862, 861,860,859,858,857,856,855,854,853,852,851,850,849, 848,847,846,845,844,843,842,841,840,839,838,837,836, 835,834,833,832,831,830,829,828,827,826,825,824,823, 822,821,820,819,818,817,816,815,814,813,812,811,810, 809,808,807,806,805,804,803,802,801,800,799,798,797, 796,795,794,793,792,791,790,789,788,787,786,785,784, 783,782,781,780,779,778,777,776,775,774,773,772,771, 770,769,768,767,766,765,764,763,762,761,760,759,758, 757,756,755,754,753,752,751,750,749,748,747,746,745, 744,743,742,741,740,739,738,737,736,735,734,733,732, 731,730,729,728,727,726,725,724,723,722,721,720,719, 718,717,716,715,714,713,712,711,710,709,708,707,706, 705,704,703,702,701,700,699,698,697,696,695,694,693, 692,691,690,689,688,687,686,685,684,683,682,681,680, 679,678,677,676,675,674,673,672,671,670,669,668,667, 666,665,664,663,662,661,660,659,658,657,656,655,654, 653,652,651,650,649,648,647,646,645,644,643,642,641, 640,639,638,637,636,635,634,633,632,631,630,629,628, 627,626,625,624,623,622,621,620,619,618,617,616,615, 614,613,612,611,610,609,608,607,606,605,604,603,602, 601,600,599,598,597,596,595,594,593,592,591,590,589, 588,587,586,585,584,583,582,581,580,579,578,577,576, 575,574,573,572,571,570,569,568,567,566,565,564,563, 562,561,560,559,558,557,556,555,554,553,552,551,550, 549,548,547,546,545,544,543,542,541,540,539,538,537, 536,535,534,533,532,531,530,529,528,527,526,525,524, 523,522,521,520,519,518,517,516,515,514,513,512], 'ns_1@10.1.4.244'}, 5000,ns_port_server,start_link, [vbucketmigrator,"./bin/vbucketmigrator/vbucketmigrator", ["-e","-a","default","-h","10.1.4.243:11210","-d", "10.1.4.244:11210","-A","-N", "r-ns_1@10.1.4.244-ns_1@10.1.4.243-1304814704.105247", "-v","-b","1023","-b","1022","-b","1021","-b","1020", "-b","1019","-b","1018","-b","1017","-b","1016","-b", "1015","-b","1014","-b","1013","-b","1012","-b","1011", "-b","1010","-b","1009","-b","1008","-b","1007","-b", "1006","-b","1005","-b","1004","-b","1003","-b","1002", "-b","1001","-b","1000","-b","999","-b","998","-b","997", "-b","996","-b","995","-b","994","-b","993","-b","992", "-b","991","-b","990","-b","989","-b","988","-b","987", "-b","986","-b","985","-b","984","-b","983","-b","982", "-b","981","-b","980","-b","979","-b","978","-b","977", "-b","976","-b","975","-b","974","-b","973","-b","972", "-b","971","-b","970","-b","969","-b","968","-b","967", "-b","966","-b","965","-b","964","-b","963","-b","962", "-b","961","-b","960","-b","959","-b","958","-b","957", "-b","956","-b","955","-b","954","-b","953","-b","952", "-b","951","-b","950","-b","949","-b","948","-b","947", "-b","946","-b","945","-b","944","-b","943","-b","942", "-b","941","-b","940","-b","939","-b","938","-b","937", "-b","936","-b","935","-b","934","-b","933","-b","932", "-b","931","-b","930","-b","929","-b","928","-b","927", "-b","926","-b","925","-b","924","-b","923","-b","922", "-b","921","-b","920","-b","919","-b","918","-b","917", "-b","916","-b","915","-b","914","-b","913","-b","912", "-b","911","-b","910","-b","909","-b","908","-b","907", "-b","906","-b","905","-b","904","-b","903","-b","902", "-b","901","-b","900","-b","899","-b","898","-b","897", "-b","896","-b","895","-b","894","-b","893","-b","892", "-b","891","-b","890","-b","889","-b","888","-b","887", "-b","886","-b","885","-b","884","-b","883","-b","882", "-b","881","-b","880","-b","879","-b","878","-b","877", "-b","876","-b","875","-b","874","-b","873","-b","872", "-b","871","-b","870","-b","869","-b","868","-b","867", "-b","866","-b","865","-b","864","-b","863","-b","862", "-b","861","-b","860","-b","859","-b","858","-b","857", "-b","856","-b","855","-b","854","-b","853","-b","852", "-b","851","-b","850","-b","849","-b","848","-b","847", "-b","846","-b","845","-b","844","-b","843","-b","842", "-b","841","-b","840","-b","839","-b","838","-b","837", "-b","836","-b","835","-b","834","-b","833","-b","832", "-b","831","-b","830","-b","829","-b","828","-b","827", "-b","826","-b","825","-b","824","-b","823","-b","822", "-b","821","-b","820","-b","819","-b","818","-b","817", "-b","816","-b","815","-b","814","-b","813","-b","812", "-b","811","-b","810","-b","809","-b","808","-b","807", "-b","806","-b","805","-b","804","-b","803","-b","802", "-b","801","-b","800","-b","799","-b","798","-b","797", "-b","796","-b","795","-b","794","-b","793","-b","792", "-b","791","-b","790","-b","789","-b","788","-b","787", "-b","786","-b","785","-b","784","-b","783","-b","782", "-b","781","-b","780","-b","779","-b","778","-b","777", "-b","776","-b","775","-b","774","-b","773","-b","772", "-b","771","-b","770","-b","769","-b","768","-b","767", "-b","766","-b","765","-b","764","-b","763","-b","762", "-b","761","-b","760","-b","759","-b","758","-b","757", "-b","756","-b","755","-b","754","-b","753","-b","752", "-b","751","-b","750","-b","749","-b","748","-b","747", "-b","746","-b","745","-b","744","-b","743","-b","742", "-b","741","-b","740","-b","739","-b","738","-b","737", "-b","736","-b","735","-b","734","-b","733","-b","732", "-b","731","-b","730","-b","729","-b","728","-b","727", "-b","726","-b","725","-b","724","-b","723","-b","722", "-b","721","-b","720","-b","719","-b","718","-b","717", "-b","716","-b","715","-b","714","-b","713","-b","712", "-b","711","-b","710","-b","709","-b","708","-b","707", "-b","706","-b","705","-b","704","-b","703","-b","702", "-b","701","-b","700","-b","699","-b","698","-b","697", "-b","696","-b","695","-b","694","-b","693","-b","692", "-b","691","-b","690","-b","689","-b","688","-b","687", "-b","686","-b","685","-b","684","-b","683","-b","682", "-b","681","-b","680","-b","679","-b","678","-b","677", "-b","676","-b","675","-b","674","-b","673","-b","672", "-b","671","-b","670","-b","669","-b","668","-b","667", "-b","666","-b","665","-b","664","-b","663","-b","662", "-b","661","-b","660","-b","659","-b","658","-b","657", "-b","656","-b","655","-b","654","-b","653","-b","652", "-b","651","-b","650","-b","649","-b","648","-b","647", "-b","646","-b","645","-b","644","-b","643","-b","642", "-b","641","-b","640","-b","639","-b","638","-b","637", "-b","636","-b","635","-b","634","-b","633","-b","632", "-b","631","-b","630","-b","629","-b","628","-b","627", "-b","626","-b","625","-b","624","-b","623","-b","622", "-b","621","-b","620","-b","619","-b","618","-b","617", "-b","616","-b","615","-b","614","-b","613","-b","612", "-b","611","-b","610","-b","609","-b","608","-b","607", "-b","606","-b","605","-b","604","-b","603","-b","602", "-b","601","-b","600","-b","599","-b","598","-b","597", "-b","596","-b","595","-b","594","-b","593","-b","592", "-b","591","-b","590","-b","589","-b","588","-b","587", "-b","586","-b","585","-b","584","-b","583","-b","582", "-b","581","-b","580","-b","579","-b","578","-b","577", "-b","576","-b","575","-b","574","-b","573","-b","572", "-b","571","-b","570","-b","569","-b","568","-b","567", "-b","566","-b","565","-b","564","-b","563","-b","562", "-b","561","-b","560","-b","559","-b","558","-b","557", "-b","556","-b","555","-b","554","-b","553","-b","552", "-b","551","-b","550","-b","549","-b","548","-b","547", "-b","546","-b","545","-b","544","-b","543","-b","542", "-b","541","-b","540","-b","539","-b","538","-b","537", "-b","536","-b","535","-b","534","-b","533","-b","532", "-b","531","-b","530","-b","529","-b","528","-b","527", "-b","526","-b","525","-b","524","-b","523","-b","522", "-b","521","-b","520","-b","519","-b","518","-b","517", "-b","516","-b","515","-b","514","-b","513","-b","512"], [use_stdio,stderr_to_stdout,{write_data,[[],"\n"]}]]]}}, {restart_type,permanent}, {shutdown,10}, {child_type,worker}] INFO REPORT <0.3409.0> 2011-05-07 17:31:45 =============================================================================== vbucketmigrator<0.3409.0>: Connecting to {Sock 10.1.4.244:11210} vbucketmigrator<0.3409.0>: Authenticating towards: {Sock 10.1.4.244:11210} vbucketmigrator<0.3409.0>: Authenticated towards: {Sock 10.1.4.244:11210} vbucketmigrator<0.3409.0>: Connecting to {Sock 10.1.4.243:11210} vbucketmigrator<0.3409.0>: Authenticating towards: {Sock 10.1.4.243:11210} vbucketmigrator<0.3409.0>: Authenticated towards: {Sock 10.1.4.243:11210} INFO REPORT <0.3411.0> 2011-05-07 17:31:45 =============================================================================== vbucketmigrator<0.3411.0>: Connecting to {Sock 10.1.4.244:11210} vbucketmigrator<0.3411.0>: Authenticating towards: {Sock 10.1.4.244:11210} vbucketmigrator<0.3411.0>: Authenticated towards: {Sock 10.1.4.244:11210} vbucketmigrator<0.3411.0>: Connecting to {Sock 10.1.4.243:11210} vbucketmigrator<0.3411.0>: Authenticating towards: {Sock 10.1.4.243:11210} vbucketmigrator<0.3411.0>: Authenticated towards: {Sock 10.1.4.243:11210} INFO REPORT <0.85.0> 2011-05-07 17:31:50 =============================================================================== Pulling config from: 'ns_1@10.1.4.244' INFO REPORT <0.109.0> 2011-05-07 17:31:50 =============================================================================== ns_log: logging ns_orchestrator:4:Starting rebalance, KeepNodes = ['ns_1@10.1.4.243'], EjectNodes = ['ns_1@10.1.4.244'] INFO REPORT <0.71.0> 2011-05-07 17:31:50 =============================================================================== ns_log: suppressing duplicate log ns_orchestrator:4("Starting rebalance, KeepNodes = ['ns_1@10.1.4.243'], EjectNodes = ['ns_1@10.1.4.244']\n") because it's been seen 1 times in the past 139.352773 secs (last seen 139.352773 secs ago INFO REPORT <0.64.0> 2011-05-07 17:31:50 =============================================================================== config change: {node,'ns_1@10.1.4.243',membership} -> active INFO REPORT <0.64.0> 2011-05-07 17:31:50 =============================================================================== ns_node_disco_conf_events config all INFO REPORT <0.85.0> 2011-05-07 17:31:50 =============================================================================== Pushing config INFO REPORT <0.85.0> 2011-05-07 17:31:50 =============================================================================== Pushing config done INFO REPORT <0.64.0> 2011-05-07 17:31:50 =============================================================================== config change: rebalance_status -> running INFO REPORT <0.64.0> 2011-05-07 17:31:50 =============================================================================== ns_node_disco_conf_events config all INFO REPORT <0.85.0> 2011-05-07 17:31:50 =============================================================================== Pushing config INFO REPORT <0.85.0> 2011-05-07 17:31:50 =============================================================================== Pushing config done INFO REPORT <6517.473.0> 2011-05-07 17:31:51 =============================================================================== vbucketmigrator<0.473.0>: Connecting to {Sock 10.1.4.243:11210} vbucketmigrator<0.473.0>: Authenticating towards: {Sock 10.1.4.243:11210} vbucketmigrator<0.473.0>: Authenticated towards: {Sock 10.1.4.243:11210} vbucketmigrator<0.473.0>: Connecting to {Sock 10.1.4.244:11210} vbucketmigrator<0.473.0>: Authenticating towards: {Sock 10.1.4.244:11210} vbucketmigrator<0.473.0>: Authenticated towards: {Sock 10.1.4.244:11210} vbucketmigrator<0.473.0>: Starting to move bucket 0 vbucketmigrator<0.473.0>: Bucket 0 moved to the next server INFO REPORT <0.109.0> 2011-05-07 17:31:53 =============================================================================== ns_1@10.1.4.243:ns_orchestrator:164: Skipping janitor in state rebalancing: {rebalancing_state, <0.3446.0>, {dict, 2, 16, 16, 8, 80, 48, {[], [], [], [], [], [], [], [], [], [], [], [], [], [], [], []}, {{[], [], [], [['ns_1@10.1.4.243'| 0.0]], [['ns_1@10.1.4.244'| 0.0]], [], [], [], [], [], [], [], [], [], [], []}}}} ERROR REPORT <6517.473.0> 2011-05-07 17:31:54 =============================================================================== ** Generic server <6517.473.0> terminating ** Last message in was {'EXIT',<6517.472.0>,stopped} ** When Server state == {state,#Port<6517.2814>,vbucketmigrator, {["Bucket 0 moved to the next server", "Starting to move bucket 0"], ["Authenticated towards: {Sock 10.1.4.244:11210}"]}, undefined,[],0} ** Reason for termination == ** stopped CRASH REPORT <6517.473.0> 2011-05-07 17:31:54 =============================================================================== Crashing process initial_call {ns_port_server,init,['Argument__1']} pid <6517.473.0> registered_name [] error_info {exit,stopped, [{gen_server,terminate,6},{proc_lib,init_p_do_apply,3}]} ancestors [<6517.472.0>] messages [{'EXIT',#Port<6517.2814>,normal}] links [] dictionary [] trap_exit true status running heap_size 987 stack_size 24 reductions 413 INFO REPORT <0.109.0> 2011-05-07 17:31:54 =============================================================================== ns_log: logging ns_orchestrator:2:Rebalance exited with reason stopped INFO REPORT <0.71.0> 2011-05-07 17:31:54 =============================================================================== ns_log: suppressing duplicate log ns_orchestrator:2("Rebalance exited with reason stopped\n") because it's been seen 1 times in the past 12.959749 secs (last seen 12.959749 secs ago INFO REPORT <0.64.0> 2011-05-07 17:31:54 =============================================================================== config change: buckets -> [{configs,[{"default", [{num_replicas,1}, {ram_quota,839909376}, {auth_type,sasl}, {sasl_password,[]}, {type,membase}, {num_vbuckets,1024}, {db_shards,4}, {ht_size,3079}, {tap_keepalive,0}, {tap_noop_interval,20}, {max_txn_size,1000}, {ht_locks,5}, {servers,['ns_1@10.1.4.243','ns_1@10.1.4.244']}, {map,[['ns_1@10.1.4.244','ns_1@10.1.4.243'], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243'|...], [...]|...]}]}]}] INFO REPORT <0.64.0> 2011-05-07 17:31:54 =============================================================================== ns_node_disco_conf_events config all INFO REPORT <0.85.0> 2011-05-07 17:31:54 =============================================================================== Pushing config INFO REPORT <0.85.0> 2011-05-07 17:31:54 =============================================================================== Pushing config done INFO REPORT <0.64.0> 2011-05-07 17:31:54 =============================================================================== config change: rebalance_status -> {none,<<"Rebalance failed. See logs for detailed reason. You can try rebalance again.">>} INFO REPORT <0.64.0> 2011-05-07 17:31:54 =============================================================================== ns_node_disco_conf_events config all INFO REPORT <0.85.0> 2011-05-07 17:31:54 =============================================================================== Pushing config INFO REPORT <0.85.0> 2011-05-07 17:31:54 =============================================================================== Pushing config done INFO REPORT <0.109.0> 2011-05-07 17:31:56 =============================================================================== ns_log: logging ns_orchestrator:4:Starting rebalance, KeepNodes = ['ns_1@10.1.4.243'], EjectNodes = ['ns_1@10.1.4.244'] INFO REPORT <0.71.0> 2011-05-07 17:31:56 =============================================================================== ns_log: suppressing duplicate log ns_orchestrator:4("Starting rebalance, KeepNodes = ['ns_1@10.1.4.243'], EjectNodes = ['ns_1@10.1.4.244']\n") because it's been seen 2 times in the past 145.77911 secs (last seen 6.426337 secs ago INFO REPORT <0.64.0> 2011-05-07 17:31:56 =============================================================================== config change: {node,'ns_1@10.1.4.243',membership} -> active INFO REPORT <0.64.0> 2011-05-07 17:31:56 =============================================================================== ns_node_disco_conf_events config all INFO REPORT <0.85.0> 2011-05-07 17:31:56 =============================================================================== Pushing config INFO REPORT <0.85.0> 2011-05-07 17:31:56 =============================================================================== Pushing config done INFO REPORT <0.64.0> 2011-05-07 17:31:56 =============================================================================== config change: rebalance_status -> running INFO REPORT <0.64.0> 2011-05-07 17:31:56 =============================================================================== ns_node_disco_conf_events config all INFO REPORT <0.85.0> 2011-05-07 17:31:56 =============================================================================== Pushing config INFO REPORT <0.3531.0> 2011-05-07 17:31:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 1 from replica to dead because we don't have all copies INFO REPORT <0.85.0> 2011-05-07 17:31:56 =============================================================================== Pushing config done INFO REPORT <0.3531.0> 2011-05-07 17:31:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 2 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 3 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 4 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 5 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 6 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 7 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 8 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 9 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 10 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 11 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 12 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 13 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 14 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 15 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 16 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 17 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 18 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 19 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 20 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 21 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 22 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 23 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 24 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 25 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 26 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 27 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 28 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 29 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 30 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 31 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 32 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 33 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 34 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 35 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 36 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 37 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 38 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 39 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 40 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 41 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 42 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 43 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 44 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 45 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 46 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 47 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 48 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 49 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 50 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 51 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 52 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 53 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 54 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 55 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 56 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 57 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 58 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 59 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 60 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 61 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 62 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 63 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 64 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 65 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 66 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 67 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 68 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 69 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 70 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 71 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 72 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 73 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 74 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 75 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 76 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 77 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 78 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 79 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 80 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 81 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 82 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 83 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 84 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 85 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 86 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 87 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 88 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 89 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 90 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 91 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 92 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 93 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 94 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 95 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 96 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 97 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 98 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 99 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 100 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 101 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 102 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 103 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 104 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 105 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 106 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 107 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 108 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 109 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 110 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 111 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 112 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 113 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 114 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 115 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 116 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 117 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 118 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 119 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 120 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 121 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 122 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 123 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 124 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 125 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 126 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 127 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 128 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 129 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 130 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 131 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 132 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 133 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 134 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 135 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 136 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 137 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 138 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 139 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 140 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 141 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 142 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 143 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 144 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 145 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 146 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 147 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 148 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 149 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 150 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 151 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 152 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 153 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 154 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 155 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 156 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 157 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 158 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 159 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 160 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 161 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 162 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 163 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 164 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 165 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 166 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 167 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 168 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 169 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 170 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 171 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 172 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 173 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 174 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 175 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 176 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 177 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 178 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 179 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 180 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 181 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 182 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 183 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 184 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 185 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 186 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 187 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 188 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 189 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 190 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 191 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 192 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 193 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 194 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 195 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 196 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 197 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 198 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 199 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 200 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 201 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 202 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 203 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 204 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 205 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 206 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 207 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 208 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 209 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 210 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 211 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 212 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 213 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 214 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 215 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 216 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 217 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 218 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 219 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 220 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 221 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 222 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 223 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 224 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 225 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 226 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 227 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 228 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 229 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 230 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 231 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 232 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 233 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 234 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 235 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 236 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 237 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 238 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 239 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 240 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 241 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 242 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 243 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 244 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 245 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 246 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 247 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 248 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 249 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 250 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 251 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 252 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 253 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 254 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 255 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 256 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 257 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 258 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 259 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 260 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 261 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 262 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 263 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 264 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 265 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 266 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 267 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 268 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 269 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 270 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 271 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 272 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 273 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 274 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 275 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 276 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 277 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 278 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 279 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 280 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 281 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 282 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 283 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 284 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 285 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 286 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 287 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 288 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 289 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 290 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 291 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 292 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 293 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 294 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 295 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 296 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 297 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 298 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 299 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 300 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 301 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 302 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 303 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 304 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 305 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 306 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 307 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 308 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 309 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 310 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 311 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 312 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 313 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 314 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 315 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 316 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 317 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 318 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 319 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 320 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 321 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 322 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 323 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 324 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 325 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 326 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 327 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 328 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 329 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 330 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 331 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 332 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 333 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 334 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 335 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 336 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 337 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 338 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 339 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 340 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 341 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 342 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 343 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 344 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 345 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 346 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 347 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 348 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 349 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 350 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 351 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 352 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 353 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 354 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 355 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 356 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 357 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 358 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 359 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 360 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 361 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 362 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 363 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 364 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 365 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 366 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 367 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 368 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 369 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 370 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 371 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 372 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 373 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 374 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 375 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 376 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 377 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 378 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 379 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 380 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 381 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 382 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 383 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 384 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 385 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 386 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 387 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 388 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 389 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 390 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 391 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 392 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 393 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 394 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 395 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 396 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 397 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 398 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 399 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 400 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 401 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 402 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 403 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 404 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 405 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 406 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 407 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 408 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 409 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 410 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 411 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 412 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 413 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 414 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 415 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 416 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 417 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 418 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 419 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 420 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 421 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 422 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 423 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 424 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 425 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 426 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 427 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 428 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 429 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 430 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 431 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 432 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 433 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 434 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 435 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 436 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 437 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 438 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 439 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 440 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 441 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 442 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 443 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 444 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 445 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 446 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 447 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 448 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 449 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 450 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 451 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 452 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 453 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 454 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 455 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 456 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 457 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 458 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 459 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 460 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 461 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 462 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 463 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 464 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 465 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 466 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 467 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 468 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 469 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 470 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 471 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 472 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 473 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 474 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 475 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 476 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 477 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 478 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 479 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 480 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 481 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 482 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 483 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 484 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 485 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 486 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 487 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 488 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 489 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 490 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 491 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 492 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 493 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 494 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 495 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 496 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 497 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 498 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 499 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 500 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 501 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 502 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 503 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 504 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 505 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 506 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 507 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 508 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 509 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 510 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 511 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 512 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 513 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 514 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 515 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 516 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 517 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 518 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 519 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 520 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 521 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 522 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 523 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 524 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 525 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 526 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 527 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 528 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 529 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 530 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 531 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 532 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 533 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 534 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 535 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 536 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 537 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 538 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 539 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 540 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 541 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 542 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 543 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 544 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 545 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 546 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 547 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 548 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 549 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 550 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 551 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 552 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 553 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 554 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 555 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 556 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 557 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 558 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 559 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 560 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 561 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 562 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 563 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 564 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 565 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 566 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 567 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 568 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 569 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 570 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 571 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 572 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 573 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 574 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 575 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 576 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 577 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 578 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 579 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 580 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 581 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 582 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 583 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 584 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 585 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 586 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 587 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 588 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 589 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 590 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 591 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 592 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 593 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 594 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 595 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 596 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 597 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 598 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 599 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 600 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 601 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 602 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 603 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 604 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 605 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 606 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 607 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 608 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 609 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 610 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 611 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 612 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 613 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 614 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 615 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 616 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 617 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 618 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 619 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 620 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 621 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 622 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 623 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 624 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 625 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 626 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 627 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 628 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 629 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 630 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 631 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 632 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 633 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 634 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 635 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 636 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 637 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 638 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 639 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 640 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 641 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 642 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 643 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 644 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 645 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 646 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 647 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 648 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 649 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 650 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 651 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 652 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 653 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 654 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 655 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 656 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 657 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 658 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 659 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 660 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 661 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 662 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 663 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 664 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 665 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 666 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 667 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 668 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 669 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 670 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 671 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 672 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 673 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 674 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 675 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 676 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 677 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 678 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 679 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 680 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 681 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 682 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 683 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 684 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 685 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 686 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 687 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 688 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 689 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 690 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 691 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 692 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 693 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 694 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 695 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 696 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 697 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 698 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 699 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 700 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 701 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 702 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 703 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 704 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 705 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 706 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 707 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 708 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 709 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 710 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 711 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 712 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 713 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 714 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 715 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 716 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 717 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 718 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 719 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 720 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 721 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 722 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 723 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 724 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 725 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 726 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 727 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 728 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 729 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 730 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 731 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 732 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 733 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 734 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 735 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 736 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 737 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 738 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 739 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 740 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 741 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 742 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 743 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 744 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 745 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 746 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 747 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 748 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 749 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 750 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 751 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 752 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 753 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 754 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 755 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 756 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 757 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 758 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 759 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 760 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 761 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 762 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 763 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 764 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 765 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 766 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 767 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 768 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 769 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 770 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 771 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 772 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 773 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 774 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 775 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 776 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 777 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 778 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 779 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 780 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 781 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 782 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 783 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 784 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 785 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 786 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 787 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 788 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 789 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 790 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 791 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 792 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 793 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 794 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 795 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 796 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 797 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 798 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 799 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 800 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 801 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 802 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 803 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 804 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 805 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 806 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 807 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 808 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 809 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 810 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 811 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 812 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 813 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 814 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 815 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 816 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 817 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 818 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 819 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 820 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 821 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 822 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 823 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 824 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 825 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 826 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 827 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 828 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 829 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 830 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 831 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 832 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 833 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 834 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 835 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 836 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 837 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 838 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 839 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 840 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 841 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 842 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 843 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 844 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 845 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 846 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 847 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 848 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 849 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 850 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 851 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 852 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 853 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 854 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 855 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 856 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 857 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 858 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 859 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 860 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 861 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 862 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 863 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 864 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 865 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 866 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 867 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 868 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 869 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 870 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 871 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 872 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 873 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 874 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 875 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 876 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 877 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 878 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 879 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 880 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 881 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 882 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 883 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 884 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 885 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 886 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 887 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 888 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 889 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 890 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 891 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 892 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 893 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 894 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 895 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 896 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 897 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 898 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 899 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 900 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 901 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 902 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 903 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 904 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 905 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 906 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 907 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 908 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 909 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 910 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 911 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 912 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 913 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 914 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 915 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 916 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 917 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 918 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 919 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 920 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 921 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 922 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 923 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 924 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 925 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 926 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 927 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 928 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 929 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 930 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 931 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 932 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 933 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 934 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 935 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 936 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 937 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 938 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 939 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 940 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 941 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 942 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 943 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 944 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 945 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 946 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 947 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 948 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 949 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 950 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 951 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 952 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 953 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 954 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 955 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 956 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 957 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 958 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 959 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 960 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 961 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 962 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 963 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 964 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 965 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 966 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 967 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 968 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 969 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 970 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 971 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 972 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 973 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 974 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 975 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 976 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 977 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 978 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 979 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 980 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 981 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 982 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 983 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 984 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 985 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 986 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 987 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 988 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 989 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 990 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 991 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 992 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 993 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 994 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 995 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 996 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 997 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 998 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 999 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 1000 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 1001 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 1002 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 1003 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 1004 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 1005 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 1006 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 1007 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 1008 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 1009 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 1010 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 1011 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 1012 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 1013 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 1014 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 1015 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 1016 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 1017 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 1018 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 1019 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 1020 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 1021 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 1022 from replica to dead because we don't have all copies INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_janitor:220: Setting vbucket 'ns_1@10.1.4.244' in "default" on 1023 from replica to dead because we don't have all copies INFO REPORT <0.64.0> 2011-05-07 17:31:57 =============================================================================== config change: buckets -> [{configs,[{"default", [{num_replicas,1}, {ram_quota,839909376}, {auth_type,sasl}, {sasl_password,[]}, {type,membase}, {num_vbuckets,1024}, {db_shards,4}, {ht_size,3079}, {tap_keepalive,0}, {tap_noop_interval,20}, {max_txn_size,1000}, {ht_locks,5}, {servers,['ns_1@10.1.4.243','ns_1@10.1.4.244']}, {map,[['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243'|...], [...]|...]}]}]}] INFO REPORT <0.64.0> 2011-05-07 17:31:57 =============================================================================== ns_node_disco_conf_events config all INFO REPORT <0.85.0> 2011-05-07 17:31:57 =============================================================================== Pushing config INFO REPORT <0.85.0> 2011-05-07 17:31:57 =============================================================================== Pushing config done INFO REPORT <0.64.0> 2011-05-07 17:31:57 =============================================================================== config change: buckets -> [{configs,[{"default", [{num_replicas,1}, {ram_quota,839909376}, {auth_type,sasl}, {sasl_password,[]}, {type,membase}, {num_vbuckets,1024}, {db_shards,4}, {ht_size,3079}, {tap_keepalive,0}, {tap_noop_interval,20}, {max_txn_size,1000}, {ht_locks,5}, {servers,['ns_1@10.1.4.243']}, {map,[['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243'|...], [...]|...]}]}]}] INFO REPORT <0.64.0> 2011-05-07 17:31:57 =============================================================================== ns_node_disco_conf_events config all INFO REPORT <0.3531.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_cluster:201: Shunning 'ns_1@10.1.4.244' INFO REPORT <0.85.0> 2011-05-07 17:31:57 =============================================================================== Pushing config INFO REPORT <0.109.0> 2011-05-07 17:31:57 =============================================================================== ns_log: logging ns_orchestrator:1:Rebalance completed successfully. INFO REPORT <0.71.0> 2011-05-07 17:31:57 =============================================================================== ns_log: suppressing duplicate log ns_orchestrator:1("Rebalance completed successfully.\n") because it's been seen 1 times in the past 146.581449 secs (last seen 146.581449 secs ago INFO REPORT <0.85.0> 2011-05-07 17:31:57 =============================================================================== Pushing config done INFO REPORT <0.64.0> 2011-05-07 17:31:57 =============================================================================== config change: {node,'ns_1@10.1.4.244',membership} -> inactiveAdded INFO REPORT <0.64.0> 2011-05-07 17:31:57 =============================================================================== ns_node_disco_conf_events config all INFO REPORT <0.85.0> 2011-05-07 17:31:57 =============================================================================== Pushing config INFO REPORT <0.85.0> 2011-05-07 17:31:57 =============================================================================== Pushing config done INFO REPORT <0.64.0> 2011-05-07 17:31:57 =============================================================================== ns_node_disco_conf_events config on nodes_wanted INFO REPORT <0.64.0> 2011-05-07 17:31:57 =============================================================================== config change: nodes_wanted -> ['ns_1@10.1.4.243'] INFO REPORT <0.3581.0> 2011-05-07 17:31:57 =============================================================================== ns_node_disco cookie_sync INFO REPORT <0.3581.0> 2011-05-07 17:31:57 =============================================================================== ns_node_disco: nodes_wanted updated: ['ns_1@10.1.4.243'], with cookie: elyrdmyctqykhepq INFO REPORT <0.3581.0> 2011-05-07 17:31:57 =============================================================================== ns_node_disco: nodes_wanted pong: ['ns_1@10.1.4.243'], with cookie: elyrdmyctqykhepq INFO REPORT <0.64.0> 2011-05-07 17:31:57 =============================================================================== ns_node_disco_conf_events config all INFO REPORT <0.85.0> 2011-05-07 17:31:57 =============================================================================== Pushing config INFO REPORT <0.85.0> 2011-05-07 17:31:57 =============================================================================== Pushing config done ERROR REPORT <0.109.0> 2011-05-07 17:31:57 =============================================================================== ns_1@10.1.4.243:ns_orchestrator:199: Got unexpected message {'EXIT', <6517.240.0>, normal} in state idle with data {idle_state, []} INFO REPORT <0.78.0> 2011-05-07 17:31:57 =============================================================================== ns_log: logging ns_node_disco:5:Node 'ns_1@10.1.4.243' saw that node 'ns_1@10.1.4.244' went down. INFO REPORT <0.71.0> 2011-05-07 17:31:57 =============================================================================== ns_log: suppressing duplicate log ns_node_disco:5("Node 'ns_1@10.1.4.243' saw that node 'ns_1@10.1.4.244' went down.") because it's been seen 1 times in the past 146.557451 secs (last seen 146.557451 secs ago INFO REPORT <0.64.0> 2011-05-07 17:31:57 =============================================================================== config change: rebalance_status -> none INFO REPORT <0.64.0> 2011-05-07 17:31:57 =============================================================================== ns_node_disco_conf_events config all INFO REPORT <0.85.0> 2011-05-07 17:31:57 =============================================================================== Pushing config INFO REPORT <0.85.0> 2011-05-07 17:31:57 =============================================================================== Pushing config done INFO REPORT <0.77.0> 2011-05-07 17:32:02 =============================================================================== Detected a new node (from node 'ns_1@10.1.4.243'). Moving config around. INFO REPORT <0.77.0> 2011-05-07 17:32:02 =============================================================================== ns_node_disco_log: nodes changed: ['ns_1@10.1.4.243'] INFO REPORT <0.85.0> 2011-05-07 17:32:02 =============================================================================== Pulling config INFO REPORT <0.85.0> 2011-05-07 17:32:02 =============================================================================== Pulling config done INFO REPORT <0.2211.0> 2011-05-07 17:32:43 =============================================================================== ns_1@10.1.4.243:ns_doctor:82: Current node statuses: [{'ns_1@10.1.4.243', [{last_heard,{1304,814763,60277}}, {active_buckets,["default"]}, {memory, [{total,25298536}, {processes,14335132}, {processes_used,14318956}, {system,10963404}, {atom,536281}, {atom_used,522284}, {binary,171544}, {code,4290757}, {ets,1022660}]}, {cluster_compatibility_version,1}, {version, [{kernel,"2.13.4"}, {mnesia,"4.4.12"}, {stdlib,"1.16.4"}, {os_mon,"2.2.4"}, {inets,"5.2"}, {ns_server,"1.6.5.4r"}, {menelaus,"1.6.5.4r"}, {sasl,"2.1.8"}]}, {system_arch,"i686-pc-linux-gnu"}, {wall_clock,1660}, {memory_data,{1050726400,792068096,{<0.2218.0>,2057360}}}, {disk_data, [{"/",7583436,18}, {"/dev",508460,1}, {"/dev/shm",513048,0}, {"/var/run",513048,1}, {"/var/lock",513048,0}, {"/lib/init/rw",513048,0}, {"/boot",233191,8}]}, {meminfo, <<"MemTotal: 1026100 kB\nMemFree: 251132 kB\nBuffers: 152780 kB\nCached: 452136 kB\nSwapCached: 0 kB\nActive: 457784 kB\nInactive: 278680 kB\nActive(anon): 127376 kB\nInactive(anon): 4412 kB\nActive(file): 330408 kB\nInactive(file): 274268 kB\nUnevictable: 0 kB\nMlocked: 0 kB\nHighTotal: 137160 kB\nHighFree: 252 kB\nLowTotal: 888940 kB\nLowFree: 250880 kB\nSwapTotal: 397304 kB\nSwapFree: 397304 kB\nDirty: 100 kB\nWriteback: 0 kB\nAnonPages: 131548 kB\nMapped: 10412 kB\nShmem: 240 kB\nSlab: 28216 kB\nSReclaimable: 23296 kB\nSUnreclaim: 4920 kB\nKernelStack: 976 kB\nPageTables: 1100 kB\nNFS_Unstable: 0 kB\nBounce: 0 kB\nWritebackTmp: 0 kB\nCommitLimit: 910352 kB\nCommitted_AS: 352472 kB\nVmallocTotal: 122880 kB\nVmallocUsed: 3540 kB\nVmallocChunk: 114424 kB\nHardwareCorrupted: 0 kB\nHugePages_Total: 0\nHugePages_Free: 0\nHugePages_Rsvd: 0\nHugePages_Surp: 0\nHugepagesize: 2048 kB\nDirectMap4k: 10232 kB\nDirectMap2M: 901120 kB\n">>}, {replication,[{"default",1.0}]}, {system_memory_data, [{system_total_memory,1050726400}, {free_swap,406839296}, {total_swap,406839296}, {cached_memory,462987264}, {buffered_memory,156446720}, {free_memory,257159168}, {total_memory,1050726400}]}, {statistics, [{wall_clock,{1650410,1}}, {context_switches,{364084,0}}, {garbage_collection,{79893,489106123,0}}, {io,{{input,19494241},{output,24952065}}}, {reductions,{287410627,2200863}}, {run_queue,0}, {runtime,{28560,150}}]}]}, {'ns_1@10.1.4.244', [{last_heard,{1304,814717,110014}}, {active_buckets,["default"]}, {memory, [{total,14419808}, {processes,4361060}, {processes_used,4356820}, {system,10058748}, {atom,515729}, {atom_used,491094}, {binary,261632}, {code,3977696}, {ets,401668}]}, {cluster_compatibility_version,1}, {version, [{kernel,"2.13.4"}, {mnesia,"4.4.12"}, {stdlib,"1.16.4"}, {os_mon,"2.2.4"}, {ns_server,"1.6.5.4r"}, {menelaus,"1.6.5.4r"}, {sasl,"2.1.8"}]}, {system_arch,"i686-pc-linux-gnu"}, {wall_clock,117}, {memory_data,{1050726400,877510656,{<6517.183.0>,901104}}}, {disk_data, [{"/",7583436,16}, {"/dev",508460,1}, {"/dev/shm",513048,0}, {"/var/run",513048,1}, {"/var/lock",513048,0}, {"/lib/init/rw",513048,0}, {"/boot",233191,8}]}, {meminfo, <<"MemTotal: 1026100 kB\nMemFree: 155248 kB\nBuffers: 24376 kB\nCached: 261032 kB\nSwapCached: 0 kB\nActive: 209908 kB\nInactive: 177556 kB\nActive(anon): 96156 kB\nInactive(anon): 6152 kB\nActive(file): 113752 kB\nInactive(file): 171404 kB\nUnevictable: 0 kB\nMlocked: 0 kB\nHighTotal: 137160 kB\nHighFree: 252 kB\nLowTotal: 888940 kB\nLowFree: 154996 kB\nSwapTotal: 397304 kB\nSwapFree: 397304 kB\nDirty: 304 kB\nWriteback: 0 kB\nAnonPages: 102080 kB\nMapped: 10224 kB\nShmem: 232 kB\nSlab: 21548 kB\nSReclaimable: 16300 kB\nSUnreclaim: 5248 kB\nKernelStack: 944 kB\nPageTables: 964 kB\nNFS_Unstable: 0 kB\nBounce: 0 kB\nWritebackTmp: 0 kB\nCommitLimit: 910352 kB\nCommitted_AS: 318372 kB\nVmallocTotal: 122880 kB\nVmallocUsed: 3540 kB\nVmallocChunk: 114844 kB\nHardwareCorrupted: 0 kB\nHugePages_Total: 0\nHugePages_Free: 0\nHugePages_Rsvd: 0\nHugePages_Surp: 0\nHugepagesize: 2048 kB\nDirectMap4k: 10232 kB\nDirectMap2M: 901120 kB\n">>}, {replication,[{"default",1.0}]}, {system_memory_data, [{system_total_memory,1050726400}, {free_swap,406839296}, {total_swap,406839296}, {cached_memory,267296768}, {buffered_memory,24961024}, {free_memory,158973952}, {total_memory,1050726400}]}, {statistics, [{wall_clock,{113022,2}}, {context_switches,{25943,0}}, {garbage_collection,{5681,15361914,0}}, {io,{{input,5786525},{output,2145861}}}, {reductions,{5258163,537480}}, {run_queue,0}, {runtime,{1070,120}}]}]}] INFO REPORT <0.2204.0> 2011-05-07 17:32:44 =============================================================================== ns_1@10.1.4.243:stats_collector:71: Stats for bucket "default": auth_cmds 12 auth_errors 0 bucket_conns 9 bytes_read 39346980 bytes_written 57683616 cas_badval 0 cas_hits 0 cas_misses 0 cmd_flush 0 cmd_get 204800 cmd_set 215040 conn_yields 13 connection_structures 22 curr_connections 19 curr_items 204800 curr_items_tot 204800 daemon_connections 10 decr_hits 0 decr_misses 0 delete_hits 0 delete_misses 0 ep_bg_fetched 0 ep_commit_num 458 ep_commit_time 0 ep_commit_time_total 5 ep_data_age 1 ep_data_age_highwat 5 ep_db_cleaner_status complete ep_db_strategy multiMTVBDB ep_dbinit 1 ep_dbname /var/opt/membase/1.6.5.4r/data/ns_1/default ep_dbshards 4 ep_expired 0 ep_flush_duration 0 ep_flush_duration_highwat 1 ep_flush_duration_total 6 ep_flush_preempts 0 ep_flusher_state running ep_flusher_todo 0 ep_io_num_read 200 ep_io_num_write 215242 ep_io_read_bytes 22800 ep_io_write_bytes 24537360 ep_item_begin_failed 0 ep_item_commit_failed 0 ep_item_flush_expired 0 ep_item_flush_failed 0 ep_kv_size 35840000 ep_max_data_size 839909376 ep_max_txn_size 1000 ep_mem_high_wat 629932032 ep_mem_low_wat 503945625 ep_min_data_age 0 ep_num_active_non_resident 0 ep_num_eject_failures 0 ep_num_eject_replicas 0 ep_num_expiry_pager_runs 0 ep_num_non_resident 0 ep_num_not_my_vbuckets 0 ep_num_pager_runs 0 ep_num_value_ejects 0 ep_oom_errors 0 ep_overhead 12960032 ep_pending_ops 0 ep_pending_ops_max 0 ep_pending_ops_max_duration 0 ep_pending_ops_total 0 ep_queue_age_cap 900 ep_queue_size 0 ep_storage_age 0 ep_storage_age_highwat 2 ep_storage_type featured ep_store_max_concurrency 10 ep_store_max_readers 9 ep_store_max_readwrite 1 ep_tap_bg_fetch_requeued 0 ep_tap_bg_fetched 200 ep_tap_keepalive 300 ep_tmp_oom_errors 0 ep_too_old 0 ep_too_young 0 ep_total_cache_size 35840000 ep_total_del_items 0 ep_total_enqueued 215439 ep_total_new_items 205000 ep_total_persisted 215240 ep_vbucket_del 1 ep_vbucket_del_avg_walltime 109375 ep_vbucket_del_fail 0 ep_vbucket_del_max_walltime 109375 ep_vbucket_del_total_walltime 109375 ep_version 1.6.5.3.1_8_gce31baa ep_warmed_up 0 ep_warmup true ep_warmup_dups 0 ep_warmup_oom 0 ep_warmup_thread complete ep_warmup_time 29657 get_hits 204800 get_misses 0 incr_hits 0 incr_misses 0 libevent 2.0.7-rc limit_maxbytes 67108864 mem_used 48800032 pid 5769 pointer_size 32 rejected_conns 0 rusage_system 2.616163 rusage_user 5.012313 tap_connect_received 3 tap_mutation_received 400 tap_mutation_sent 91026 tap_opaque_received 1 tap_opaque_sent 1026 tap_vbucket_set_received 2 tap_vbucket_set_sent 2 threads 4 time 1304814764 total_connections 23 uptime 1663 version 1.4.4_382_g9df3289 INFO REPORT <0.2211.0> 2011-05-07 17:33:43 =============================================================================== ns_1@10.1.4.243:ns_doctor:82: Current node statuses: [{'ns_1@10.1.4.243', [{last_heard,{1304,814823,60251}}, {active_buckets,["default"]}, {memory, [{total,21717456}, {processes,10701420}, {processes_used,10684732}, {system,11016036}, {atom,536281}, {atom_used,522284}, {binary,190960}, {code,4290757}, {ets,1055964}]}, {cluster_compatibility_version,1}, {version, [{kernel,"2.13.4"}, {mnesia,"4.4.12"}, {stdlib,"1.16.4"}, {os_mon,"2.2.4"}, {inets,"5.2"}, {ns_server,"1.6.5.4r"}, {menelaus,"1.6.5.4r"}, {sasl,"2.1.8"}]}, {system_arch,"i686-pc-linux-gnu"}, {wall_clock,1720}, {memory_data,{1050726400,793821184,{<0.3548.0>,3328660}}}, {disk_data, [{"/",7583436,18}, {"/dev",508460,1}, {"/dev/shm",513048,0}, {"/var/run",513048,1}, {"/var/lock",513048,0}, {"/lib/init/rw",513048,0}, {"/boot",233191,8}]}, {meminfo, <<"MemTotal: 1026100 kB\nMemFree: 252992 kB\nBuffers: 152780 kB\nCached: 452396 kB\nSwapCached: 0 kB\nActive: 455724 kB\nInactive: 278820 kB\nActive(anon): 125196 kB\nInactive(anon): 4412 kB\nActive(file): 330528 kB\nInactive(file): 274408 kB\nUnevictable: 0 kB\nMlocked: 0 kB\nHighTotal: 137160 kB\nHighFree: 252 kB\nLowTotal: 888940 kB\nLowFree: 252740 kB\nSwapTotal: 397304 kB\nSwapFree: 397304 kB\nDirty: 96 kB\nWriteback: 0 kB\nAnonPages: 129368 kB\nMapped: 10412 kB\nShmem: 240 kB\nSlab: 28220 kB\nSReclaimable: 23300 kB\nSUnreclaim: 4920 kB\nKernelStack: 976 kB\nPageTables: 1096 kB\nNFS_Unstable: 0 kB\nBounce: 0 kB\nWritebackTmp: 0 kB\nCommitLimit: 910352 kB\nCommitted_AS: 348204 kB\nVmallocTotal: 122880 kB\nVmallocUsed: 3540 kB\nVmallocChunk: 114424 kB\nHardwareCorrupted: 0 kB\nHugePages_Total: 0\nHugePages_Free: 0\nHugePages_Rsvd: 0\nHugePages_Surp: 0\nHugepagesize: 2048 kB\nDirectMap4k: 10232 kB\nDirectMap2M: 901120 kB\n">>}, {replication,[{"default",1.0}]}, {system_memory_data, [{system_total_memory,1050726400}, {free_swap,406839296}, {total_swap,406839296}, {cached_memory,463253504}, {buffered_memory,156446720}, {free_memory,259063808}, {total_memory,1050726400}]}, {statistics, [{wall_clock,{1710410,1}}, {context_switches,{381329,0}}, {garbage_collection,{83642,520422978,0}}, {io,{{input,19542335},{output,25335984}}}, {reductions,{301018316,2192315}}, {run_queue,0}, {runtime,{29660,200}}]}]}, {'ns_1@10.1.4.244', [{last_heard,{1304,814717,110014}}, {active_buckets,["default"]}, {memory, [{total,14419808}, {processes,4361060}, {processes_used,4356820}, {system,10058748}, {atom,515729}, {atom_used,491094}, {binary,261632}, {code,3977696}, {ets,401668}]}, {cluster_compatibility_version,1}, {version, [{kernel,"2.13.4"}, {mnesia,"4.4.12"}, {stdlib,"1.16.4"}, {os_mon,"2.2.4"}, {ns_server,"1.6.5.4r"}, {menelaus,"1.6.5.4r"}, {sasl,"2.1.8"}]}, {system_arch,"i686-pc-linux-gnu"}, {wall_clock,117}, {memory_data,{1050726400,877510656,{<6517.183.0>,901104}}}, {disk_data, [{"/",7583436,16}, {"/dev",508460,1}, {"/dev/shm",513048,0}, {"/var/run",513048,1}, {"/var/lock",513048,0}, {"/lib/init/rw",513048,0}, {"/boot",233191,8}]}, {meminfo, <<"MemTotal: 1026100 kB\nMemFree: 155248 kB\nBuffers: 24376 kB\nCached: 261032 kB\nSwapCached: 0 kB\nActive: 209908 kB\nInactive: 177556 kB\nActive(anon): 96156 kB\nInactive(anon): 6152 kB\nActive(file): 113752 kB\nInactive(file): 171404 kB\nUnevictable: 0 kB\nMlocked: 0 kB\nHighTotal: 137160 kB\nHighFree: 252 kB\nLowTotal: 888940 kB\nLowFree: 154996 kB\nSwapTotal: 397304 kB\nSwapFree: 397304 kB\nDirty: 304 kB\nWriteback: 0 kB\nAnonPages: 102080 kB\nMapped: 10224 kB\nShmem: 232 kB\nSlab: 21548 kB\nSReclaimable: 16300 kB\nSUnreclaim: 5248 kB\nKernelStack: 944 kB\nPageTables: 964 kB\nNFS_Unstable: 0 kB\nBounce: 0 kB\nWritebackTmp: 0 kB\nCommitLimit: 910352 kB\nCommitted_AS: 318372 kB\nVmallocTotal: 122880 kB\nVmallocUsed: 3540 kB\nVmallocChunk: 114844 kB\nHardwareCorrupted: 0 kB\nHugePages_Total: 0\nHugePages_Free: 0\nHugePages_Rsvd: 0\nHugePages_Surp: 0\nHugepagesize: 2048 kB\nDirectMap4k: 10232 kB\nDirectMap2M: 901120 kB\n">>}, {replication,[{"default",1.0}]}, {system_memory_data, [{system_total_memory,1050726400}, {free_swap,406839296}, {total_swap,406839296}, {cached_memory,267296768}, {buffered_memory,24961024}, {free_memory,158973952}, {total_memory,1050726400}]}, {statistics, [{wall_clock,{113022,2}}, {context_switches,{25943,0}}, {garbage_collection,{5681,15361914,0}}, {io,{{input,5786525},{output,2145861}}}, {reductions,{5258163,537480}}, {run_queue,0}, {runtime,{1070,120}}]}]}] INFO REPORT <0.2204.0> 2011-05-07 17:34:24 =============================================================================== ns_1@10.1.4.243:stats_collector:71: Stats for bucket "default": auth_cmds 12 auth_errors 0 bucket_conns 9 bytes_read 39352036 bytes_written 59336219 cas_badval 0 cas_hits 0 cas_misses 0 cmd_flush 0 cmd_get 204800 cmd_set 215040 conn_yields 13 connection_structures 22 curr_connections 19 curr_items 204800 curr_items_tot 204800 daemon_connections 10 decr_hits 0 decr_misses 0 delete_hits 0 delete_misses 0 ep_bg_fetched 0 ep_commit_num 458 ep_commit_time 0 ep_commit_time_total 5 ep_data_age 1 ep_data_age_highwat 5 ep_db_cleaner_status complete ep_db_strategy multiMTVBDB ep_dbinit 1 ep_dbname /var/opt/membase/1.6.5.4r/data/ns_1/default ep_dbshards 4 ep_expired 0 ep_flush_duration 0 ep_flush_duration_highwat 1 ep_flush_duration_total 6 ep_flush_preempts 0 ep_flusher_state running ep_flusher_todo 0 ep_io_num_read 200 ep_io_num_write 215242 ep_io_read_bytes 22800 ep_io_write_bytes 24537360 ep_item_begin_failed 0 ep_item_commit_failed 0 ep_item_flush_expired 0 ep_item_flush_failed 0 ep_kv_size 35840000 ep_max_data_size 839909376 ep_max_txn_size 1000 ep_mem_high_wat 629932032 ep_mem_low_wat 503945625 ep_min_data_age 0 ep_num_active_non_resident 0 ep_num_eject_failures 0 ep_num_eject_replicas 0 ep_num_expiry_pager_runs 0 ep_num_non_resident 0 ep_num_not_my_vbuckets 0 ep_num_pager_runs 0 ep_num_value_ejects 0 ep_oom_errors 0 ep_overhead 12960032 ep_pending_ops 0 ep_pending_ops_max 0 ep_pending_ops_max_duration 0 ep_pending_ops_total 0 ep_queue_age_cap 900 ep_queue_size 0 ep_storage_age 0 ep_storage_age_highwat 2 ep_storage_type featured ep_store_max_concurrency 10 ep_store_max_readers 9 ep_store_max_readwrite 1 ep_tap_bg_fetch_requeued 0 ep_tap_bg_fetched 200 ep_tap_keepalive 300 ep_tmp_oom_errors 0 ep_too_old 0 ep_too_young 0 ep_total_cache_size 35840000 ep_total_del_items 0 ep_total_enqueued 215439 ep_total_new_items 205000 ep_total_persisted 215240 ep_vbucket_del 1 ep_vbucket_del_avg_walltime 109375 ep_vbucket_del_fail 0 ep_vbucket_del_max_walltime 109375 ep_vbucket_del_total_walltime 109375 ep_version 1.6.5.3.1_8_gce31baa ep_warmed_up 0 ep_warmup true ep_warmup_dups 0 ep_warmup_oom 0 ep_warmup_thread complete ep_warmup_time 29657 get_hits 204800 get_misses 0 incr_hits 0 incr_misses 0 libevent 2.0.7-rc limit_maxbytes 67108864 mem_used 48800032 pid 5769 pointer_size 32 rejected_conns 0 rusage_system 2.620163 rusage_user 5.072317 tap_connect_received 3 tap_mutation_received 400 tap_mutation_sent 91026 tap_opaque_received 1 tap_opaque_sent 1026 tap_vbucket_set_received 2 tap_vbucket_set_sent 2 threads 4 time 1304814864 total_connections 23 uptime 1763 version 1.4.4_382_g9df3289 INFO REPORT <0.2211.0> 2011-05-07 17:34:43 =============================================================================== ns_1@10.1.4.243:ns_doctor:82: Current node statuses: [{'ns_1@10.1.4.243', [{last_heard,{1304,814883,60257}}, {active_buckets,["default"]}, {memory, [{total,21741448}, {processes,10698108}, {processes_used,10681420}, {system,11043340}, {atom,536281}, {atom_used,522284}, {binary,180960}, {code,4290757}, {ets,1092508}]}, {cluster_compatibility_version,1}, {version, [{kernel,"2.13.4"}, {mnesia,"4.4.12"}, {stdlib,"1.16.4"}, {os_mon,"2.2.4"}, {inets,"5.2"}, {ns_server,"1.6.5.4r"}, {menelaus,"1.6.5.4r"}, {sasl,"2.1.8"}]}, {system_arch,"i686-pc-linux-gnu"}, {wall_clock,1780}, {memory_data,{1050726400,790773760,{<0.3990.0>,3328660}}}, {disk_data, [{"/",7583436,18}, {"/dev",508460,1}, {"/dev/shm",513048,0}, {"/var/run",513048,1}, {"/var/lock",513048,0}, {"/lib/init/rw",513048,0}, {"/boot",233191,8}]}, {meminfo, <<"MemTotal: 1026100 kB\nMemFree: 253736 kB\nBuffers: 152780 kB\nCached: 451440 kB\nSwapCached: 0 kB\nActive: 455680 kB\nInactive: 278108 kB\nActive(anon): 125396 kB\nInactive(anon): 4412 kB\nActive(file): 330284 kB\nInactive(file): 273696 kB\nUnevictable: 0 kB\nMlocked: 0 kB\nHighTotal: 137160 kB\nHighFree: 252 kB\nLowTotal: 888940 kB\nLowFree: 253484 kB\nSwapTotal: 397304 kB\nSwapFree: 397304 kB\nDirty: 80 kB\nWriteback: 0 kB\nAnonPages: 129568 kB\nMapped: 10412 kB\nShmem: 240 kB\nSlab: 28228 kB\nSReclaimable: 23312 kB\nSUnreclaim: 4916 kB\nKernelStack: 976 kB\nPageTables: 1096 kB\nNFS_Unstable: 0 kB\nBounce: 0 kB\nWritebackTmp: 0 kB\nCommitLimit: 910352 kB\nCommitted_AS: 348204 kB\nVmallocTotal: 122880 kB\nVmallocUsed: 3540 kB\nVmallocChunk: 114424 kB\nHardwareCorrupted: 0 kB\nHugePages_Total: 0\nHugePages_Free: 0\nHugePages_Rsvd: 0\nHugePages_Surp: 0\nHugepagesize: 2048 kB\nDirectMap4k: 10232 kB\nDirectMap2M: 901120 kB\n">>}, {replication,[{"default",1.0}]}, {system_memory_data, [{system_total_memory,1050726400}, {free_swap,406839296}, {total_swap,406839296}, {cached_memory,462274560}, {buffered_memory,156446720}, {free_memory,259825664}, {total_memory,1050726400}]}, {statistics, [{wall_clock,{1770410,1}}, {context_switches,{398932,0}}, {garbage_collection,{87512,552260406,0}}, {io,{{input,20339777},{output,26698971}}}, {reductions,{314762782,2286702}}, {run_queue,0}, {runtime,{30690,200}}]}]}, {'ns_1@10.1.4.244', [{last_heard,{1304,814717,110014}}, {active_buckets,["default"]}, {memory, [{total,14419808}, {processes,4361060}, {processes_used,4356820}, {system,10058748}, {atom,515729}, {atom_used,491094}, {binary,261632}, {code,3977696}, {ets,401668}]}, {cluster_compatibility_version,1}, {version, [{kernel,"2.13.4"}, {mnesia,"4.4.12"}, {stdlib,"1.16.4"}, {os_mon,"2.2.4"}, {ns_server,"1.6.5.4r"}, {menelaus,"1.6.5.4r"}, {sasl,"2.1.8"}]}, {system_arch,"i686-pc-linux-gnu"}, {wall_clock,117}, {memory_data,{1050726400,877510656,{<6517.183.0>,901104}}}, {disk_data, [{"/",7583436,16}, {"/dev",508460,1}, {"/dev/shm",513048,0}, {"/var/run",513048,1}, {"/var/lock",513048,0}, {"/lib/init/rw",513048,0}, {"/boot",233191,8}]}, {meminfo, <<"MemTotal: 1026100 kB\nMemFree: 155248 kB\nBuffers: 24376 kB\nCached: 261032 kB\nSwapCached: 0 kB\nActive: 209908 kB\nInactive: 177556 kB\nActive(anon): 96156 kB\nInactive(anon): 6152 kB\nActive(file): 113752 kB\nInactive(file): 171404 kB\nUnevictable: 0 kB\nMlocked: 0 kB\nHighTotal: 137160 kB\nHighFree: 252 kB\nLowTotal: 888940 kB\nLowFree: 154996 kB\nSwapTotal: 397304 kB\nSwapFree: 397304 kB\nDirty: 304 kB\nWriteback: 0 kB\nAnonPages: 102080 kB\nMapped: 10224 kB\nShmem: 232 kB\nSlab: 21548 kB\nSReclaimable: 16300 kB\nSUnreclaim: 5248 kB\nKernelStack: 944 kB\nPageTables: 964 kB\nNFS_Unstable: 0 kB\nBounce: 0 kB\nWritebackTmp: 0 kB\nCommitLimit: 910352 kB\nCommitted_AS: 318372 kB\nVmallocTotal: 122880 kB\nVmallocUsed: 3540 kB\nVmallocChunk: 114844 kB\nHardwareCorrupted: 0 kB\nHugePages_Total: 0\nHugePages_Free: 0\nHugePages_Rsvd: 0\nHugePages_Surp: 0\nHugepagesize: 2048 kB\nDirectMap4k: 10232 kB\nDirectMap2M: 901120 kB\n">>}, {replication,[{"default",1.0}]}, {system_memory_data, [{system_total_memory,1050726400}, {free_swap,406839296}, {total_swap,406839296}, {cached_memory,267296768}, {buffered_memory,24961024}, {free_memory,158973952}, {total_memory,1050726400}]}, {statistics, [{wall_clock,{113022,2}}, {context_switches,{25943,0}}, {garbage_collection,{5681,15361914,0}}, {io,{{input,5786525},{output,2145861}}}, {reductions,{5258163,537480}}, {run_queue,0}, {runtime,{1070,120}}]}]}] INFO REPORT <3.54.0> 2011-05-07 17:35:45 =============================================================================== nonode@nohost:log_os_info:25: OS type: {unix,linux} Version: {2,6,32} Runtime info: [{otp_release,"R13B03"}, {erl_version,"5.7.4"}, {erl_version_long,"Erlang R13B03 (erts-5.7.4) [source] [rq:1] [async-threads:16] [hipe] [kernel-poll:false]\n"}, {system_arch_raw,"i686-pc-linux-gnu"}, {system_arch,"i686-pc-linux-gnu"}, {localtime,{{2011,5,7},{17,35,45}}}, {memory,[{total,7706696}, {processes,649420}, {processes_used,642092}, {system,7057276}, {atom,255161}, {atom_used,252399}, {binary,41272}, {code,1775796}, {ets,165092}]}, {loaded,[ns_info,log_os_info,misc,ns_log_mf_h, ns_server_cluster_sup,ns_server,timer,io_lib_fread, cpu_sup,memsup,disksup,os_mon,io_lib_pretty,unicode, io_lib_format,io_lib,io,sasl_report,release_handler, calendar,overload,alarm_handler,sasl_report_tty_h, sasl,ns_bootstrap,file_io_server,orddict,erl_eval, file,c,error_logger_tty_h,queue,kernel_config,user, user_sup,supervisor_bridge,standard_error,gb_sets, hipe_unified_loader,packages,code_server,code, file_server,net_kernel,global_group,erl_distribution, string,ets,filename,inet_gethost_native,inet_parse, inet,inet_udp,os,inet_config,inet_db,global,gb_trees, rpc,dict,supervisor,kernel,application_master,sys, application,gen_server,erl_parse,proplists,erl_scan, lists,application_controller,proc_lib,gen,gen_event, error_logger,heart,error_handler,erlang, erl_prim_loader,prim_zip,zlib,prim_file,prim_inet, init,otp_ring0]}, {applications,[{kernel,"ERTS CXC 138 10","2.13.4"}, {stdlib,"ERTS CXC 138 10","1.16.4"}, {os_mon,"CPO CXC 138 46","2.2.4"}, {ns_server,"Membase server","1.6.5.4r"}, {menelaus,"Membase menelaus","1.6.5.4r"}, {sasl,"SASL CXC 138 11","2.1.8"}]}, {pre_loaded,[erlang,erl_prim_loader,prim_zip,zlib,prim_file, prim_inet,init,otp_ring0]}, {process_count,44}, {node,nonode@nohost}, {nodes,[]}, {registered,[standard_error_sup,rex,user,init,kernel_sup, inet_db,global_name_server,release_handler, code_server,file_server_2,error_logger,overload, ns_server_cluster_sup,application_controller, kernel_safe_sup,standard_error,alarm_handler, global_group,timer_server,sasl_sup,os_mon_sup, cpu_sup,erl_prim_loader,memsup,sasl_safe_sup, disksup]}, {cookie,nocookie}, {wordsize,4}, {wall_clock,0}] INFO REPORT <3.55.0> 2011-05-07 17:35:45 =============================================================================== reading ip config from "/etc/opt/membase/1.6.5.4r/ip" PROGRESS REPORT <3.56.0> 2011-05-07 17:35:45 =============================================================================== supervisor {local,inet_gethost_native_sup} started [{pid,<3.57.0>},{mfa,{inet_gethost_native,init,[[]]}}] PROGRESS REPORT <3.27.0> 2011-05-07 17:35:45 =============================================================================== supervisor {local,kernel_safe_sup} started [{pid,<3.56.0>}, {name,inet_gethost_native_sup}, {mfa,{inet_gethost_native,start_link,[]}}, {restart_type,temporary}, {shutdown,1000}, {child_type,worker}] PROGRESS REPORT <3.58.0> 2011-05-07 17:35:45 =============================================================================== supervisor {local,net_sup} started [{pid,<3.59.0>}, {name,erl_epmd}, {mfa,{erl_epmd,start_link,[]}}, {restart_type,permanent}, {shutdown,2000}, {child_type,worker}] PROGRESS REPORT <3.58.0> 2011-05-07 17:35:45 =============================================================================== supervisor {local,net_sup} started [{pid,<3.60.0>}, {name,auth}, {mfa,{auth,start_link,[]}}, {restart_type,permanent}, {shutdown,2000}, {child_type,worker}] PROGRESS REPORT <0.58.0> 2011-05-07 17:35:45 =============================================================================== supervisor {local,net_sup} started [{pid,<0.61.0>}, {name,net_kernel}, {mfa,{net_kernel,start_link,[['ns_1@10.1.4.243',longnames]]}}, {restart_type,permanent}, {shutdown,2000}, {child_type,worker}] PROGRESS REPORT <0.11.0> 2011-05-07 17:35:45 =============================================================================== supervisor {local,kernel_sup} started [{pid,<0.58.0>}, {name,net_sup_dynamic}, {mfa,{erl_distribution,start_link,[['ns_1@10.1.4.243',longnames]]}}, {restart_type,permanent}, {shutdown,1000}, {child_type,supervisor}] PROGRESS REPORT <0.54.0> 2011-05-07 17:35:45 =============================================================================== supervisor {local,ns_server_cluster_sup} started [{pid,<0.55.0>}, {name,dist_manager}, {mfa,{dist_manager,start_link,[]}}, {restart_type,permanent}, {shutdown,10}, {child_type,worker}] PROGRESS REPORT <0.54.0> 2011-05-07 17:35:45 =============================================================================== supervisor {local,ns_server_cluster_sup} started [{pid,<0.64.0>}, {name,ns_cluster}, {mfa,{ns_cluster,start_link,[]}}, {restart_type,permanent}, {shutdown,5000}, {child_type,worker}] INFO REPORT <0.65.0> 2011-05-07 17:35:45 =============================================================================== loading config from "/etc/opt/membase/1.6.5.4r/config" PROGRESS REPORT <0.65.0> 2011-05-07 17:35:45 =============================================================================== supervisor {local,ns_config_sup} started [{pid,<0.66.0>}, {name,ns_config_events}, {mfa,{gen_event,start_link,[{local,ns_config_events}]}}, {restart_type,permanent}, {shutdown,10}, {child_type,worker}] PROGRESS REPORT <0.65.0> 2011-05-07 17:35:45 =============================================================================== supervisor {local,ns_config_sup} started [{pid,<0.67.0>}, {name,ns_config}, {mfa,{ns_config,start_link, ["/etc/opt/membase/1.6.5.4r/config", ns_config_default]}}, {restart_type,permanent}, {shutdown,10}, {child_type,worker}] PROGRESS REPORT <0.65.0> 2011-05-07 17:35:45 =============================================================================== supervisor {local,ns_config_sup} started [{pid,<0.69.0>}, {name,ns_config_isasl_sync}, {mfa,{ns_config_isasl_sync,start_link,[]}}, {restart_type,transient}, {shutdown,10}, {child_type,worker}] INFO REPORT <0.66.0> 2011-05-07 17:35:45 =============================================================================== isasl_sync init: ["/var/opt/membase/1.6.5.4r/data/ns_1/isasl.pw","_admin", "_admin"] INFO REPORT <0.66.0> 2011-05-07 17:35:45 =============================================================================== isasl_sync init buckets: [] INFO REPORT <0.66.0> 2011-05-07 17:35:45 =============================================================================== Writing isasl passwd file: "/var/opt/membase/1.6.5.4r/data/ns_1/isasl.pw" PROGRESS REPORT <0.65.0> 2011-05-07 17:35:45 =============================================================================== supervisor {local,ns_config_sup} started [{pid,<0.71.0>}, {name,ns_config_log}, {mfa,{ns_config_log,start_link,[]}}, {restart_type,transient}, {shutdown,10}, {child_type,worker}] PROGRESS REPORT <0.54.0> 2011-05-07 17:35:45 =============================================================================== supervisor {local,ns_server_cluster_sup} started [{pid,<0.65.0>}, {name,ns_config_sup}, {mfa,{ns_config_sup,start_link,[]}}, {restart_type,permanent}, {shutdown,infinity}, {child_type,supervisor}] PROGRESS REPORT <0.72.0> 2011-05-07 17:35:45 =============================================================================== supervisor {local,ns_server_sup} started [{pid,<0.73.0>}, {name,ns_log}, {mfa,{ns_log,start_link,[]}}, {restart_type,permanent}, {shutdown,10}, {child_type,worker}] PROGRESS REPORT <0.72.0> 2011-05-07 17:35:45 =============================================================================== supervisor {local,ns_server_sup} started [{pid,<0.74.0>}, {name,ns_log_events}, {mfa,{gen_event,start_link,[{local,ns_log_events}]}}, {restart_type,permanent}, {shutdown,10}, {child_type,worker}] PROGRESS REPORT <0.75.0> 2011-05-07 17:35:45 =============================================================================== supervisor {local,ns_mail_sup} started [{pid,<0.76.0>}, {name,ns_mail}, {mfa,{ns_mail,start_link,[]}}, {restart_type,permanent}, {shutdown,10}, {child_type,worker}] PROGRESS REPORT <0.75.0> 2011-05-07 17:35:45 =============================================================================== supervisor {local,ns_mail_sup} started [{pid,<0.77.0>}, {name,ns_mail_log}, {mfa,{ns_mail_log,start_link,[]}}, {restart_type,transient}, {shutdown,10}, {child_type,worker}] PROGRESS REPORT <0.72.0> 2011-05-07 17:35:45 =============================================================================== supervisor {local,ns_server_sup} started [{pid,<0.75.0>}, {name,ns_mail_sup}, {mfa,{ns_mail_sup,start_link,[]}}, {restart_type,permanent}, {shutdown,infinity}, {child_type,supervisor}] PROGRESS REPORT <0.78.0> 2011-05-07 17:35:45 =============================================================================== supervisor {local,ns_node_disco_sup} started [{pid,<0.79.0>}, {name,ns_node_disco_events}, {mfa,{gen_event,start_link,[{local,ns_node_disco_events}]}}, {restart_type,permanent}, {shutdown,10}, {child_type,worker}] INFO REPORT <0.80.0> 2011-05-07 17:35:45 =============================================================================== Initting ns_node_disco with [] INFO REPORT <0.81.0> 2011-05-07 17:35:45 =============================================================================== ns_node_disco cookie_sync INFO REPORT <0.81.0> 2011-05-07 17:35:45 =============================================================================== ns_log: logging ns_node_disco:2:Node 'ns_1@10.1.4.243' synchronized otp cookie elyrdmyctqykhepq from cluster INFO REPORT <0.81.0> 2011-05-07 17:35:45 =============================================================================== ns_node_disco: nodes_wanted updated: ['ns_1@10.1.4.243'], with cookie: elyrdmyctqykhepq INFO REPORT <0.81.0> 2011-05-07 17:35:45 =============================================================================== ns_node_disco: nodes_wanted pong: ['ns_1@10.1.4.243'], with cookie: elyrdmyctqykhepq PROGRESS REPORT <0.78.0> 2011-05-07 17:35:45 =============================================================================== supervisor {local,ns_node_disco_sup} started [{pid,<0.80.0>}, {name,ns_node_disco}, {mfa,{ns_node_disco,start_link,[]}}, {restart_type,permanent}, {shutdown,10}, {child_type,worker}] PROGRESS REPORT <0.78.0> 2011-05-07 17:35:45 =============================================================================== supervisor {local,ns_node_disco_sup} started [{pid,<0.84.0>}, {name,ns_node_disco_log}, {mfa,{ns_node_disco_log,start_link,[]}}, {restart_type,transient}, {shutdown,10}, {child_type,worker}] PROGRESS REPORT <0.78.0> 2011-05-07 17:35:45 =============================================================================== supervisor {local,ns_node_disco_sup} started [{pid,<0.85.0>}, {name,ns_node_disco_conf_events}, {mfa,{ns_node_disco_conf_events,start_link,[]}}, {restart_type,transient}, {shutdown,10}, {child_type,worker}] INFO REPORT <0.86.0> 2011-05-07 17:35:45 =============================================================================== ns_config_rep init pulling INFO REPORT <0.86.0> 2011-05-07 17:35:45 =============================================================================== ns_config_rep init pushing INFO REPORT <0.86.0> 2011-05-07 17:35:45 =============================================================================== ns_config_rep init reannouncing INFO REPORT <0.66.0> 2011-05-07 17:35:45 =============================================================================== config change: alerts -> [{email,[]}, {email_alerts,false}, [{user,undefined}, {pass,"********"}, {addr,undefined}, {port,undefined}, {encrypt,false}], {alerts,[server_down,server_unresponsive,server_up,server_joined,server_left, bucket_created,bucket_deleted,bucket_auth_failed]}] INFO REPORT <0.66.0> 2011-05-07 17:35:45 =============================================================================== config change: buckets -> [{configs,[{"default", [{num_replicas,1}, {ram_quota,839909376}, {auth_type,sasl}, {sasl_password,[]}, {type,membase}, {num_vbuckets,1024}, {db_shards,4}, {ht_size,3079}, {tap_keepalive,0}, {tap_noop_interval,20}, {max_txn_size,1000}, {ht_locks,5}, {servers,['ns_1@10.1.4.243']}, {map,[['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243'|...], [...]|...]}]}]}] INFO REPORT <0.66.0> 2011-05-07 17:35:45 =============================================================================== Writing isasl passwd file: "/var/opt/membase/1.6.5.4r/data/ns_1/isasl.pw" INFO REPORT <0.66.0> 2011-05-07 17:35:45 =============================================================================== config change: memory_quota -> 801 INFO REPORT <0.66.0> 2011-05-07 17:35:45 =============================================================================== config change: moxi -> [{port,11211},{verbosity,[]}] INFO REPORT <0.66.0> 2011-05-07 17:35:45 =============================================================================== ns_node_disco_conf_events config on nodes_wanted INFO REPORT <0.66.0> 2011-05-07 17:35:45 =============================================================================== config change: nodes_wanted -> ['ns_1@10.1.4.243'] INFO REPORT <0.66.0> 2011-05-07 17:35:45 =============================================================================== ns_node_disco_conf_events config on otp INFO REPORT <0.66.0> 2011-05-07 17:35:45 =============================================================================== config change: otp -> [{cookie,elyrdmyctqykhepq}] INFO REPORT <0.89.0> 2011-05-07 17:35:45 =============================================================================== ns_node_disco cookie_sync INFO REPORT <0.66.0> 2011-05-07 17:35:45 =============================================================================== config change: port_servers -> [{moxi,"./bin/moxi/moxi", ["-Z", {"port_listen=~B,default_bucket_name=default,downstream_max=1024,downstream_conn_max=4,connect_max_errors=5,connect_retry_interval=30000,connect_timeout=400,auth_timeout=100,cycle=200,downstream_conn_queue_timeout=200,downstream_timeout=5000,wait_queue_timeout=200", [port]}, "-z", {"url=http://127.0.0.1:~B/pools/default/saslBucketsStreaming", [{rest,port}]}, "-p","0","-Y","y","-O","stderr", {"~s",[verbosity]}], [{env,[{"EVENT_NOSELECT","1"}, {"MOXI_SASL_PLAIN_USR",{"~s",[{ns_moxi_sup,rest_user,[]}]}}, {"MOXI_SASL_PLAIN_PWD",{"~s",[{ns_moxi_sup,rest_pass,[]}]}}]}, use_stdio,stderr_to_stdout,stream]}, {memcached,"./bin/memcached/memcached", ["-X","./bin/memcached/stdin_term_handler.so","-p", {"~B",[port]}, "-E","./bin/bucket_engine/bucket_engine.so","-B","binary","-r", "-c","10000","-e", {"admin=~s;default_bucket_name=default;auto_create=false", [admin_user]}, {"~s",[verbosity]}], [{env,[{"EVENT_NOSELECT","1"}, {"MEMCACHED_TOP_KEYS","100"}, {"ISASL_PWFILE",{"~s",[{isasl,path}]}}, {"ISASL_DB_CHECK_TIME","1"}]}, use_stdio,stderr_to_stdout,stream]}] INFO REPORT <0.66.0> 2011-05-07 17:35:45 =============================================================================== config change: rebalance_status -> none INFO REPORT <0.90.0> 2011-05-07 17:35:45 =============================================================================== ns_node_disco cookie_sync INFO REPORT <0.66.0> 2011-05-07 17:35:45 =============================================================================== config change: replication -> [{enabled,true}] INFO REPORT <0.66.0> 2011-05-07 17:35:45 =============================================================================== config change: rest -> [{port,8091}] INFO REPORT <0.66.0> 2011-05-07 17:35:45 =============================================================================== config change: rest_creds -> ******** INFO REPORT <0.89.0> 2011-05-07 17:35:45 =============================================================================== ns_node_disco: nodes_wanted updated: ['ns_1@10.1.4.243'], with cookie: elyrdmyctqykhepq INFO REPORT <0.66.0> 2011-05-07 17:35:45 =============================================================================== config change: {node,'ns_1@10.1.4.243',isasl} -> [{path,"/var/opt/membase/1.6.5.4r/data/ns_1/isasl.pw"}] INFO REPORT <0.66.0> 2011-05-07 17:35:45 =============================================================================== config change: {node,'ns_1@10.1.4.243',membership} -> active INFO REPORT <0.90.0> 2011-05-07 17:35:45 =============================================================================== ns_node_disco: nodes_wanted updated: ['ns_1@10.1.4.243'], with cookie: elyrdmyctqykhepq INFO REPORT <0.66.0> 2011-05-07 17:35:45 =============================================================================== config change: {node,'ns_1@10.1.4.243',memcached} -> [{dbdir,"/var/opt/membase/1.6.5.4r/data/ns_1"}, {port,11210}, {admin_user,"_admin"}, {admin_pass,"_admin"}, {bucket_engine,"./bin/bucket_engine/bucket_engine.so"}, {engines,[{membase,[{engine,"bin/ep_engine/ep.so"}, {initfile,"/etc/opt/membase/1.6.5.4r/init.sql"}]}, {memcached,[{engine,"bin/memcached/default_engine.so"}]}]}, {verbosity,[]}] INFO REPORT <0.66.0> 2011-05-07 17:35:45 =============================================================================== config change: {node,'ns_1@10.1.4.243',ns_log} -> [{filename,"/var/opt/membase/1.6.5.4r/data/ns_1/ns_log"}] INFO REPORT <0.89.0> 2011-05-07 17:35:45 =============================================================================== ns_node_disco: nodes_wanted pong: ['ns_1@10.1.4.243'], with cookie: elyrdmyctqykhepq INFO REPORT <0.66.0> 2011-05-07 17:35:45 =============================================================================== config change: {node,'ns_1@10.1.4.244',isasl} -> [{path,"/var/opt/membase/1.6.5.4r/data/ns_1/isasl.pw"}] INFO REPORT <0.66.0> 2011-05-07 17:35:45 =============================================================================== config change: {node,'ns_1@10.1.4.244',membership} -> inactiveAdded INFO REPORT <0.66.0> 2011-05-07 17:35:45 =============================================================================== config change: {node,'ns_1@10.1.4.244',memcached} -> [{dbdir,"/var/opt/membase/1.6.5.4r/data/ns_1"}, {port,11210}, {admin_user,"_admin"}, {admin_pass,"_admin"}, {bucket_engine,"./bin/bucket_engine/bucket_engine.so"}, {engines,[{membase,[{engine,"bin/ep_engine/ep.so"}, {initfile,"/etc/opt/membase/1.6.5.4r/init.sql"}]}, {memcached,[{engine,"bin/memcached/default_engine.so"}]}]}, {verbosity,[]}] INFO REPORT <0.90.0> 2011-05-07 17:35:45 =============================================================================== ns_node_disco: nodes_wanted pong: ['ns_1@10.1.4.243'], with cookie: elyrdmyctqykhepq INFO REPORT <0.66.0> 2011-05-07 17:35:45 =============================================================================== config change: {node,'ns_1@10.1.4.244',ns_log} -> [{filename,"/var/opt/membase/1.6.5.4r/data/ns_1/ns_log"}] INFO REPORT <0.66.0> 2011-05-07 17:35:45 =============================================================================== config change: {node,'ns_1@127.0.0.1',isasl} -> [{path,"/var/opt/membase/1.6.5.4r/data/ns_1/isasl.pw"}] INFO REPORT <0.66.0> 2011-05-07 17:35:45 =============================================================================== config change: {node,'ns_1@127.0.0.1',membership} -> active INFO REPORT <0.66.0> 2011-05-07 17:35:45 =============================================================================== config change: {node,'ns_1@127.0.0.1',memcached} -> [{port,11210}, {dbdir,"/var/opt/membase/1.6.5.4r/data/ns_1"}, {admin_user,"_admin"}, {admin_pass,"_admin"}, {bucket_engine,"./bin/bucket_engine/bucket_engine.so"}, {engines,[{membase,[{engine,"bin/ep_engine/ep.so"}, {initfile,"/etc/opt/membase/1.6.5.4r/init.sql"}]}, {memcached,[{engine,"bin/memcached/default_engine.so"}]}]}, {verbosity,[]}] INFO REPORT <0.66.0> 2011-05-07 17:35:45 =============================================================================== config change: {node,'ns_1@127.0.0.1',ns_log} -> [{filename,"/var/opt/membase/1.6.5.4r/data/ns_1/ns_log"}] INFO REPORT <0.66.0> 2011-05-07 17:35:45 =============================================================================== ns_node_disco_conf_events config all INFO REPORT <0.86.0> 2011-05-07 17:35:45 =============================================================================== Pushing config PROGRESS REPORT <0.78.0> 2011-05-07 17:35:45 =============================================================================== supervisor {local,ns_node_disco_sup} started [{pid,<0.86.0>}, {name,ns_config_rep}, {mfa,{ns_config_rep,start_link,[]}}, {restart_type,permanent}, {shutdown,10}, {child_type,worker}] PROGRESS REPORT <0.72.0> 2011-05-07 17:35:45 =============================================================================== supervisor {local,ns_server_sup} started [{pid,<0.78.0>}, {name,ns_node_disco_sup}, {mfa,{ns_node_disco_sup,start_link,[]}}, {restart_type,permanent}, {shutdown,infinity}, {child_type,supervisor}] INFO REPORT <0.86.0> 2011-05-07 17:35:45 =============================================================================== Pushing config done PROGRESS REPORT <0.72.0> 2011-05-07 17:35:45 =============================================================================== supervisor {local,ns_server_sup} started [{pid,<0.92.0>}, {name,ns_heart}, {mfa,{ns_heart,start_link,[]}}, {restart_type,permanent}, {shutdown,10}, {child_type,worker}] PROGRESS REPORT <0.72.0> 2011-05-07 17:35:45 =============================================================================== supervisor {local,ns_server_sup} started [{pid,<0.94.0>}, {name,ns_doctor}, {mfa,{ns_doctor,start_link,[]}}, {restart_type,permanent}, {shutdown,10}, {child_type,worker}] INFO REPORT <0.94.0> 2011-05-07 17:35:45 =============================================================================== ns_1@10.1.4.243:ns_doctor:78: Got initial status [{'ns_1@10.1.4.243', [{last_heard, {1304,814945,879987}}, {active_buckets,[]}, {memory, [{total,9210720}, {processes,1650232}, {processes_used,1645160}, {system,7560488}, {atom,313325}, {atom_used,291661}, {binary,86664}, {code,2134018}, {ets,192956}]}, {cluster_compatibility_version, 1}, {version, [{kernel,"2.13.4"}, {stdlib,"1.16.4"}, {os_mon,"2.2.4"}, {ns_server,"1.6.5.4r"}, {menelaus,"1.6.5.4r"}, {sasl,"2.1.8"}]}, {system_arch, "i686-pc-linux-gnu"}, {wall_clock,0}, {memory_data, {1050726400,674267136, {<0.21.0>,81564}}}, {disk_data, [{"/",7583436,18}, {"/dev",508460,1}, {"/dev/shm",513048,0}, {"/var/run",513048,1}, {"/var/lock",513048,0}, {"/lib/init/rw",513048, 0}, {"/boot",233191,8}]}, {meminfo, <<"MemTotal: 1026100 kB\nMemFree: 366264 kB\nBuffers: 152780 kB\nCached: 451748 kB\nSwapCached: 0 kB\nActive: 345900 kB\nInactive: 276716 kB\nActive(anon): 13916 kB\nInactive(anon): 4412 kB\nActive(file): 331984 kB\nInactive(file): 272304 kB\nUnevictable: 0 kB\nMlocked: 0 kB\nHighTotal: 137160 kB\nHighFree: 252 kB\nLowTotal: 888940 kB\nLowFree: 366012 kB\nSwapTotal: 397304 kB\nSwapFree: 397304 kB\nDirty: 136 kB\nWriteback: 0 kB\nAnonPages: 18160 kB\nMapped: 8108 kB\nShmem: 240 kB\nSlab: 28180 kB\nSReclaimable: 23288 kB\nSUnreclaim: 4892 kB\nKernelStack: 816 kB\nPageTables: 756 kB\nNFS_Unstable: 0 kB\nBounce: 0 kB\nWritebackTmp: 0 kB\nCommitLimit: 910352 kB\nCommitted_AS: 71652 kB\nVmallocTotal: 122880 kB\nVmallocUsed: 3540 kB\nVmallocChunk: 114424 kB\nHardwareCorrupted: 0 kB\nHugePages_Total: 0\nHugePages_Free: 0\nHugePages_Rsvd: 0\nHugePages_Surp: 0\nHugepagesize: 2048 kB\nDirectMap4k: 10232 kB\nDirectMap2M: 901120 kB\n">>}, {replication, [{"default",1.0}]}, {system_memory_data, [{system_total_memory, 1050726400}, {free_swap,406839296}, {total_swap,406839296}, {cached_memory, 462589952}, {buffered_memory, 156446720}, {free_memory,375054336}, {total_memory, 1050726400}]}, {statistics, [{wall_clock,{315,65}}, {context_switches, {4767,0}}, {garbage_collection, {1096,2127257,0}}, {io, {{input,2237537}, {output,118610}}}, {reductions, {939024,939024}}, {run_queue,0}, {runtime,{260,260}}]}]}] PROGRESS REPORT <0.96.0> 2011-05-07 17:35:45 =============================================================================== supervisor {local,menelaus_sup} started [{pid,<0.97.0>}, {name,menelaus_web}, {mfa,{menelaus_web,start_link,[]}}, {restart_type,permanent}, {shutdown,5000}, {child_type,worker}] PROGRESS REPORT <0.96.0> 2011-05-07 17:35:45 =============================================================================== supervisor {local,menelaus_sup} started [{pid,<0.99.0>}, {name,menelaus_event}, {mfa,{menelaus_event,start_link,[]}}, {restart_type,transient}, {shutdown,5000}, {child_type,worker}] PROGRESS REPORT <0.96.0> 2011-05-07 17:35:45 =============================================================================== supervisor {local,menelaus_sup} started [{pid,<0.100.0>}, {name,hot_keys_keeper}, {mfa,{hot_keys_keeper,start_link,[]}}, {restart_type,permanent}, {shutdown,5000}, {child_type,worker}] INFO REPORT <0.72.0> 2011-05-07 17:35:45 =============================================================================== ns_log: logging menelaus_app:1:Membase Server has started on web port 8091 on node 'ns_1@10.1.4.243'. PROGRESS REPORT <0.72.0> 2011-05-07 17:35:45 =============================================================================== supervisor {local,ns_server_sup} started [{pid,<0.96.0>}, {name,menelaus}, {mfa,{menelaus_app,start_subapp,[]}}, {restart_type,permanent}, {shutdown,infinity}, {child_type,supervisor}] PROGRESS REPORT <0.101.0> 2011-05-07 17:35:45 =============================================================================== supervisor {local,ns_port_sup} started [{pid,<0.102.0>}, {name,ns_port_init}, {mfa,{ns_port_init,start_link,[]}}, {restart_type,transient}, {shutdown,10}, {child_type,worker}] INFO REPORT <0.103.0> 2011-05-07 17:35:45 =============================================================================== starting ns_port_server with delay of 5000 PROGRESS REPORT <0.101.0> 2011-05-07 17:35:45 =============================================================================== supervisor {local,ns_port_sup} started [{pid,<0.103.0>}, {name, {moxi,"./bin/moxi/moxi", ["-Z", "port_listen=11211,default_bucket_name=default,downstream_max=1024,downstream_conn_max=4,connect_max_errors=5,connect_retry_interval=30000,connect_timeout=400,auth_timeout=100,cycle=200,downstream_conn_queue_timeout=200,downstream_timeout=5000,wait_queue_timeout=200", "-z", "url=http://127.0.0.1:8091/pools/default/saslBucketsStreaming", "-p","0","-Y","y","-O","stderr",[]], [{env, [{"EVENT_NOSELECT","1"}, {"MOXI_SASL_PLAIN_USR","Administrator"}, {"MOXI_SASL_PLAIN_PWD","111111"}]}, use_stdio,stderr_to_stdout,stream]}}, {mfa, {supervisor_cushion,start_link, [moxi,5000,ns_port_server,start_link, [moxi,"./bin/moxi/moxi", ["-Z", "port_listen=11211,default_bucket_name=default,downstream_max=1024,downstream_conn_max=4,connect_max_errors=5,connect_retry_interval=30000,connect_timeout=400,auth_timeout=100,cycle=200,downstream_conn_queue_timeout=200,downstream_timeout=5000,wait_queue_timeout=200", "-z", "url=http://127.0.0.1:8091/pools/default/saslBucketsStreaming", "-p","0","-Y","y","-O","stderr",[]], [{env, [{"EVENT_NOSELECT","1"}, {"MOXI_SASL_PLAIN_USR","Administrator"}, {"MOXI_SASL_PLAIN_PWD","111111"}]}, use_stdio,stderr_to_stdout,stream]]]}}, {restart_type,permanent}, {shutdown,10}, {child_type,worker}] INFO REPORT <0.105.0> 2011-05-07 17:35:45 =============================================================================== starting ns_port_server with delay of 5000 PROGRESS REPORT <0.101.0> 2011-05-07 17:35:45 =============================================================================== supervisor {local,ns_port_sup} started [{pid,<0.105.0>}, {name, {memcached,"./bin/memcached/memcached", ["-X","./bin/memcached/stdin_term_handler.so","-p","11210", "-E","./bin/bucket_engine/bucket_engine.so","-B","binary", "-r","-c","10000","-e", "admin=_admin;default_bucket_name=default;auto_create=false", []], [{env, [{"EVENT_NOSELECT","1"}, {"MEMCACHED_TOP_KEYS","100"}, {"ISASL_PWFILE", "/var/opt/membase/1.6.5.4r/data/ns_1/isasl.pw"}, {"ISASL_DB_CHECK_TIME","1"}]}, use_stdio,stderr_to_stdout,stream]}}, {mfa, {supervisor_cushion,start_link, [memcached,5000,ns_port_server,start_link, [memcached,"./bin/memcached/memcached", ["-X","./bin/memcached/stdin_term_handler.so","-p", "11210","-E","./bin/bucket_engine/bucket_engine.so","-B", "binary","-r","-c","10000","-e", "admin=_admin;default_bucket_name=default;auto_create=false", []], [{env, [{"EVENT_NOSELECT","1"}, {"MEMCACHED_TOP_KEYS","100"}, {"ISASL_PWFILE", "/var/opt/membase/1.6.5.4r/data/ns_1/isasl.pw"}, {"ISASL_DB_CHECK_TIME","1"}]}, use_stdio,stderr_to_stdout,stream]]]}}, {restart_type,permanent}, {shutdown,10}, {child_type,worker}] PROGRESS REPORT <0.72.0> 2011-05-07 17:35:45 =============================================================================== supervisor {local,ns_server_sup} started [{pid,<0.101.0>}, {name,ns_port_sup}, {mfa,{ns_port_sup,start_link,[]}}, {restart_type,permanent}, {shutdown,10}, {child_type,worker}] PROGRESS REPORT <0.72.0> 2011-05-07 17:35:45 =============================================================================== supervisor {local,ns_server_sup} started [{pid,<0.107.0>}, {name,ns_tick_event}, {mfa,{gen_event,start_link,[{local,ns_tick_event}]}}, {restart_type,permanent}, {shutdown,10}, {child_type,worker}] PROGRESS REPORT <0.72.0> 2011-05-07 17:35:45 =============================================================================== supervisor {local,ns_server_sup} started [{pid,<0.108.0>}, {name,ns_stats_event}, {mfa,{gen_event,start_link,[{local,ns_stats_event}]}}, {restart_type,permanent}, {shutdown,10}, {child_type,worker}] PROGRESS REPORT <0.72.0> 2011-05-07 17:35:45 =============================================================================== supervisor {local,ns_server_sup} started [{pid,<0.109.0>}, {name,ns_good_bucket_worker}, {mfa,{work_queue,start_link,[ns_good_bucket_worker]}}, {restart_type,permanent}, {shutdown,10}, {child_type,worker}] PROGRESS REPORT <0.110.0> 2011-05-07 17:35:45 =============================================================================== supervisor {local,ns_good_bucket_sup} started [{pid,<0.111.0>}, {name,{ns_vbm_sup,"default"}}, {mfa,{ns_vbm_sup,start_link,["default"]}}, {restart_type,permanent}, {shutdown,1000}, {child_type,worker}] PROGRESS REPORT <0.110.0> 2011-05-07 17:35:45 =============================================================================== supervisor {local,ns_good_bucket_sup} started [{pid,<0.112.0>}, {name,{ns_memcached,"default"}}, {mfa,{ns_memcached,start_link,["default"]}}, {restart_type,permanent}, {shutdown,86400000}, {child_type,worker}] PROGRESS REPORT <0.72.0> 2011-05-07 17:35:45 =============================================================================== supervisor {local,ns_server_sup} started [{pid,<0.110.0>}, {name,ns_good_bucket_sup}, {mfa,{ns_bucket_sup,start_link, [ns_good_bucket_sup, #Fun, ns_good_bucket_worker]}}, {restart_type,permanent}, {shutdown,infinity}, {child_type,supervisor}] ERROR REPORT <0.112.0> 2011-05-07 17:35:45 =============================================================================== ns_1@10.1.4.243:ns_memcached:374: Unable to connect: {error, {badmatch, {error,econnrefused}}}, retrying. INFO REPORT <0.72.0> 2011-05-07 17:35:45 =============================================================================== ns_1@10.1.4.243:misc:725: start_singleton(gen_fsm, ns_orchestrator, [], []): started as <0.113.0> on 'ns_1@10.1.4.243' PROGRESS REPORT <0.72.0> 2011-05-07 17:35:45 =============================================================================== supervisor {local,ns_server_sup} started [{pid,<0.113.0>}, {name,ns_orchestrator}, {mfa,{ns_orchestrator,start_link,[]}}, {restart_type,permanent}, {shutdown,20}, {child_type,worker}] PROGRESS REPORT <0.119.0> 2011-05-07 17:35:46 =============================================================================== supervisor {local,mnesia_sup} started [{pid,<0.120.0>}, {name,mnesia_event}, {mfa,{mnesia_sup,start_event,[]}}, {restart_type,permanent}, {shutdown,30000}, {child_type,worker}] PROGRESS REPORT <0.121.0> 2011-05-07 17:35:46 =============================================================================== supervisor {local,mnesia_kernel_sup} started [{pid,<0.122.0>}, {name,mnesia_monitor}, {mfa,{mnesia_monitor,start,[]}}, {restart_type,permanent}, {shutdown,3000}, {child_type,worker}] PROGRESS REPORT <0.121.0> 2011-05-07 17:35:46 =============================================================================== supervisor {local,mnesia_kernel_sup} started [{pid,<0.123.0>}, {name,mnesia_subscr}, {mfa,{mnesia_subscr,start,[]}}, {restart_type,permanent}, {shutdown,3000}, {child_type,worker}] PROGRESS REPORT <0.121.0> 2011-05-07 17:35:46 =============================================================================== supervisor {local,mnesia_kernel_sup} started [{pid,<0.124.0>}, {name,mnesia_locker}, {mfa,{mnesia_locker,start,[]}}, {restart_type,permanent}, {shutdown,3000}, {child_type,worker}] PROGRESS REPORT <0.121.0> 2011-05-07 17:35:46 =============================================================================== supervisor {local,mnesia_kernel_sup} started [{pid,<0.125.0>}, {name,mnesia_recover}, {mfa,{mnesia_recover,start,[]}}, {restart_type,permanent}, {shutdown,180000}, {child_type,worker}] PROGRESS REPORT <0.27.0> 2011-05-07 17:35:46 =============================================================================== supervisor {local,kernel_safe_sup} started [{pid,<0.127.0>}, {name,dets_sup}, {mfa,{dets_sup,start_link,[]}}, {restart_type,permanent}, {shutdown,1000}, {child_type,supervisor}] PROGRESS REPORT <0.27.0> 2011-05-07 17:35:46 =============================================================================== supervisor {local,kernel_safe_sup} started [{pid,<0.128.0>}, {name,dets}, {mfa,{dets_server,start_link,[]}}, {restart_type,permanent}, {shutdown,2000}, {child_type,worker}] PROGRESS REPORT <0.27.0> 2011-05-07 17:35:46 =============================================================================== supervisor {local,kernel_safe_sup} started [{pid,<0.132.0>}, {name,disk_log_sup}, {mfa,{disk_log_sup,start_link,[]}}, {restart_type,permanent}, {shutdown,1000}, {child_type,supervisor}] PROGRESS REPORT <0.27.0> 2011-05-07 17:35:46 =============================================================================== supervisor {local,kernel_safe_sup} started [{pid,<0.133.0>}, {name,disk_log_server}, {mfa,{disk_log_server,start_link,[]}}, {restart_type,permanent}, {shutdown,2000}, {child_type,worker}] PROGRESS REPORT <0.121.0> 2011-05-07 17:35:46 =============================================================================== supervisor {local,mnesia_kernel_sup} started [{pid,<0.126.0>}, {name,mnesia_tm}, {mfa,{mnesia_tm,start,[]}}, {restart_type,permanent}, {shutdown,30000}, {child_type,worker}] PROGRESS REPORT <0.121.0> 2011-05-07 17:35:46 =============================================================================== supervisor {local,mnesia_kernel_sup} started [{pid,<0.154.0>}, {name,mnesia_checkpoint_sup}, {mfa,{mnesia_checkpoint_sup,start,[]}}, {restart_type,permanent}, {shutdown,infinity}, {child_type,supervisor}] PROGRESS REPORT <0.121.0> 2011-05-07 17:35:46 =============================================================================== supervisor {local,mnesia_kernel_sup} started [{pid,<0.155.0>}, {name,mnesia_snmp_sup}, {mfa,{mnesia_snmp_sup,start,[]}}, {restart_type,permanent}, {shutdown,infinity}, {child_type,supervisor}] PROGRESS REPORT <0.121.0> 2011-05-07 17:35:46 =============================================================================== supervisor {local,mnesia_kernel_sup} started [{pid,<0.156.0>}, {name,mnesia_controller}, {mfa,{mnesia_controller,start,[]}}, {restart_type,permanent}, {shutdown,3000}, {child_type,worker}] PROGRESS REPORT <0.121.0> 2011-05-07 17:35:46 =============================================================================== supervisor {local,mnesia_kernel_sup} started [{pid,<0.157.0>}, {name,mnesia_late_loader}, {mfa,{mnesia_late_loader,start,[]}}, {restart_type,permanent}, {shutdown,3000}, {child_type,worker}] PROGRESS REPORT <0.119.0> 2011-05-07 17:35:46 =============================================================================== supervisor {local,mnesia_sup} started [{pid,<0.121.0>}, {name,mnesia_kernel_sup}, {mfa,{mnesia_kernel_sup,start,[]}}, {restart_type,permanent}, {shutdown,infinity}, {child_type,supervisor}] PROGRESS REPORT <0.7.0> 2011-05-07 17:35:46 =============================================================================== application mnesia started_at 'ns_1@10.1.4.243' INFO REPORT <0.115.0> 2011-05-07 17:35:46 =============================================================================== ns_1@10.1.4.243:ns_mnesia:258: Using existing disk schema on ['ns_1@10.1.4.243']. INFO REPORT <0.115.0> 2011-05-07 17:35:46 =============================================================================== ns_1@10.1.4.243:ns_mnesia:196: Current config: [{access_module,mnesia}, {auto_repair,true}, {backup_module,mnesia_backup}, {checkpoints,[]}, {db_nodes,['ns_1@10.1.4.243']}, {debug,verbose}, {directory, "/var/opt/membase/1.6.5.4r/mnesia"}, {dump_log_load_regulation, false}, {dump_log_time_threshold, 180000}, {dump_log_update_in_place, true}, {dump_log_write_threshold, 1000}, {embedded_mnemosyne,false}, {event_module,mnesia_event}, {extra_db_nodes,[]}, {fallback_activated,false}, {held_locks,[]}, {ignore_fallback_at_startup, false}, {fallback_error_function, {mnesia,lkill}}, {is_running,yes}, {local_tables, ['stats_archiver-default-day', 'stats_archiver-default-month', 'stats_archiver-default-week', 'stats_archiver-default-year', schema, 'stats_archiver-default-minute', 'stats_archiver-default-hour']}, {lock_queue,[]}, {log_version,"4.3"}, {master_node_tables,[]}, {max_wait_for_decision,10000}, {protocol_version,{7,6}}, {running_db_nodes, ['ns_1@10.1.4.243']}, {schema_location,opt_disc}, {schema_version,{3,0}}, {subscribers, [<0.120.0>,<0.115.0>]}, {tables, ['stats_archiver-default-day', 'stats_archiver-default-month', 'stats_archiver-default-week', 'stats_archiver-default-year', schema, 'stats_archiver-default-minute', 'stats_archiver-default-hour']}, {transaction_commits,2}, {transaction_failures,0}, {transaction_log_writes,0}, {transaction_restarts,0}, {transactions,[]}, {use_dir,true}, {core_dir,false}, {no_table_loaders,2}, {dc_dump_limit,4}, {version,"4.4.12"}] PROGRESS REPORT <0.72.0> 2011-05-07 17:35:46 =============================================================================== supervisor {local,ns_server_sup} started [{pid,<0.115.0>}, {name,ns_mnesia}, {mfa,{ns_mnesia,start_link,[]}}, {restart_type,permanent}, {shutdown,10000}, {child_type,worker}] PROGRESS REPORT <0.72.0> 2011-05-07 17:35:46 =============================================================================== supervisor {local,ns_server_sup} started [{pid,<0.161.0>}, {name,ns_bad_bucket_worker}, {mfa,{work_queue,start_link,[ns_bad_bucket_worker]}}, {restart_type,permanent}, {shutdown,10}, {child_type,worker}] PROGRESS REPORT <0.162.0> 2011-05-07 17:35:46 =============================================================================== supervisor {local,ns_bad_bucket_sup} started [{pid,<0.167.0>}, {name,{stats_collector,"default"}}, {mfa,{stats_collector,start_link,["default"]}}, {restart_type,permanent}, {shutdown,10}, {child_type,worker}] PROGRESS REPORT <0.162.0> 2011-05-07 17:35:46 =============================================================================== supervisor {local,ns_bad_bucket_sup} started [{pid,<0.168.0>}, {name,{stats_archiver,"default"}}, {mfa,{stats_archiver,start_link,["default"]}}, {restart_type,permanent}, {shutdown,10}, {child_type,worker}] PROGRESS REPORT <0.162.0> 2011-05-07 17:35:46 =============================================================================== supervisor {local,ns_bad_bucket_sup} started [{pid,<0.171.0>}, {name,{stats_reader,"default"}}, {mfa,{stats_reader,start_link,["default"]}}, {restart_type,permanent}, {shutdown,10}, {child_type,worker}] PROGRESS REPORT <0.72.0> 2011-05-07 17:35:46 =============================================================================== supervisor {local,ns_server_sup} started [{pid,<0.162.0>}, {name,ns_bad_bucket_sup}, {mfa,{ns_bucket_sup,start_link, [ns_bad_bucket_sup, #Fun, ns_bad_bucket_worker]}}, {restart_type,permanent}, {shutdown,infinity}, {child_type,supervisor}] PROGRESS REPORT <0.72.0> 2011-05-07 17:35:46 =============================================================================== supervisor {local,ns_server_sup} started [{pid,<0.174.0>}, {name,ns_moxi_sup}, {mfa,{ns_moxi_sup,start_link,[]}}, {restart_type,permanent}, {shutdown,infinity}, {child_type,supervisor}] INFO REPORT <0.72.0> 2011-05-07 17:35:46 =============================================================================== ns_1@10.1.4.243:misc:725: start_singleton(gen_server, ns_tick, [], []): started as <0.177.0> on 'ns_1@10.1.4.243' PROGRESS REPORT <0.72.0> 2011-05-07 17:35:46 =============================================================================== supervisor {local,ns_server_sup} started [{pid,<0.177.0>}, {name,ns_tick}, {mfa,{ns_tick,start_link,[]}}, {restart_type,permanent}, {shutdown,10}, {child_type,worker}] PROGRESS REPORT <0.54.0> 2011-05-07 17:35:46 =============================================================================== supervisor {local,ns_server_cluster_sup} started [{pid,<0.72.0>}, {name,ns_server_sup}, {mfa,{ns_server_sup,start_link,[]}}, {restart_type,permanent}, {shutdown,infinity}, {child_type,supervisor}] PROGRESS REPORT <0.7.0> 2011-05-07 17:35:46 =============================================================================== application ns_server started_at 'ns_1@10.1.4.243' INFO REPORT <0.104.0> 2011-05-07 17:35:46 =============================================================================== moxi<0.104.0>: 2011-05-07 17:35:45: (cproxy_config.c.316) env: MOXI_SASL_PLAIN_USR (13) moxi<0.104.0>: 2011-05-07 17:35:45: (cproxy_config.c.325) env: MOXI_SASL_PLAIN_PWD (6) INFO REPORT <0.112.0> 2011-05-07 17:35:47 =============================================================================== ns_1@10.1.4.243:ns_memcached:390: Created bucket "default" with config string "vb0=false;waitforwarmup=false;ht_size=3079;ht_locks=5;failpartialwarmup=false;db_shards=4;shardpattern=%d/%b-%i.mb;db_strategy=multiMTVBDB;tap_keepalive=0;tap_noop_interval=20;max_txn_size=1000;max_size=839909376;initfile=/etc/opt/membase/1.6.5.4r/init.sql;tap_keepalive=300;dbname=/var/opt/membase/1.6.5.4r/data/ns_1/default;" INFO REPORT <0.106.0> 2011-05-07 17:35:47 =============================================================================== memcached<0.106.0>: WARNING: Found duplicate entry for "tap_keepalive" INFO REPORT <0.112.0> 2011-05-07 17:35:48 =============================================================================== ns_log: logging ns_memcached:1:Bucket "default" loaded on node 'ns_1@10.1.4.243' in 1 seconds. INFO REPORT <0.167.0> 2011-05-07 17:35:49 =============================================================================== ns_1@10.1.4.243:stats_collector:71: Stats for bucket "default": auth_cmds 1 auth_errors 0 bucket_conns 2 bytes_read 134 bytes_written 14054 cas_badval 0 cas_hits 0 cas_misses 0 cmd_flush 0 cmd_get 0 cmd_set 0 conn_yields 0 connection_structures 12 curr_connections 12 curr_items 0 curr_items_tot 204800 daemon_connections 10 decr_hits 0 decr_misses 0 delete_hits 0 delete_misses 0 ep_bg_fetched 0 ep_commit_num 0 ep_commit_time 0 ep_commit_time_total 0 ep_data_age 0 ep_data_age_highwat 0 ep_db_cleaner_status complete ep_db_strategy multiMTVBDB ep_dbinit 1 ep_dbname /var/opt/membase/1.6.5.4r/data/ns_1/default ep_dbshards 4 ep_expired 0 ep_flush_duration 0 ep_flush_duration_highwat 0 ep_flush_duration_total 0 ep_flush_preempts 0 ep_flusher_state running ep_flusher_todo 0 ep_io_num_read 204800 ep_io_num_write 0 ep_io_read_bytes 23347200 ep_io_write_bytes 0 ep_item_begin_failed 0 ep_item_commit_failed 0 ep_item_flush_expired 0 ep_item_flush_failed 0 ep_kv_size 35840000 ep_max_data_size 839909376 ep_max_txn_size 1000 ep_mem_high_wat 629932032 ep_mem_low_wat 503945625 ep_min_data_age 0 ep_num_active_non_resident 0 ep_num_eject_failures 0 ep_num_eject_replicas 0 ep_num_expiry_pager_runs 0 ep_num_non_resident 0 ep_num_not_my_vbuckets 0 ep_num_pager_runs 0 ep_num_value_ejects 0 ep_oom_errors 0 ep_overhead 12960032 ep_pending_ops 0 ep_pending_ops_max 0 ep_pending_ops_max_duration 0 ep_pending_ops_total 0 ep_queue_age_cap 900 ep_queue_size 0 ep_storage_age 0 ep_storage_age_highwat 0 ep_storage_type featured ep_store_max_concurrency 10 ep_store_max_readers 9 ep_store_max_readwrite 1 ep_tap_bg_fetch_requeued 0 ep_tap_bg_fetched 0 ep_tap_keepalive 300 ep_tmp_oom_errors 0 ep_too_old 0 ep_too_young 0 ep_total_cache_size 35840000 ep_total_del_items 0 ep_total_enqueued 0 ep_total_new_items 0 ep_total_persisted 0 ep_vbucket_del 0 ep_vbucket_del_fail 0 ep_version 1.6.5.3.1_8_gce31baa ep_warmed_up 204800 ep_warmup true ep_warmup_dups 0 ep_warmup_oom 0 ep_warmup_thread complete ep_warmup_time 809268 get_hits 0 get_misses 0 incr_hits 0 incr_misses 0 libevent 2.0.7-rc limit_maxbytes 67108864 mem_used 48800032 pid 8362 pointer_size 32 rejected_conns 0 rusage_system 0.220013 rusage_user 1.004062 threads 4 time 1304814948 total_connections 12 uptime 5 version 1.4.4_382_g9df3289 INFO REPORT <0.259.0> 2011-05-07 17:35:55 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 0 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:55 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 1 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:55 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 2 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:55 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 3 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:55 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 4 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:55 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 5 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:55 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 6 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:55 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 7 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:55 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 8 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:55 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 9 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:55 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 10 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:55 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 11 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:55 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 12 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:55 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 13 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:55 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 14 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:55 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 15 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:55 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 16 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:55 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 17 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:55 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 18 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 19 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 20 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 21 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 22 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 23 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 24 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 25 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 26 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 27 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 28 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 29 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 30 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 31 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 32 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 33 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 34 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 35 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 36 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 37 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 38 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 39 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 40 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 41 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 42 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 43 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 44 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 45 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 46 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 47 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 48 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 49 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 50 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 51 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 52 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 53 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 54 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 55 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 56 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 57 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 58 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 59 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 60 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 61 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 62 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 63 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 64 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 65 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 66 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 67 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 68 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 69 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 70 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 71 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 72 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 73 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 74 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 75 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 76 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 77 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 78 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 79 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 80 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 81 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 82 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 83 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 84 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 85 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 86 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 87 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 88 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 89 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 90 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 91 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 92 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 93 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 94 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 95 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 96 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 97 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 98 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 99 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 100 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 101 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 102 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 103 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 104 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 105 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 106 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 107 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 108 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 109 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 110 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 111 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 112 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 113 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 114 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 115 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 116 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 117 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 118 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 119 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 120 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 121 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 122 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 123 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 124 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 125 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 126 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 127 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 128 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 129 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 130 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 131 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 132 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 133 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 134 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 135 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 136 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 137 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 138 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 139 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 140 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 141 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 142 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 143 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 144 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 145 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 146 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 147 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 148 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 149 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 150 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 151 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 152 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 153 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 154 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 155 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 156 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 157 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 158 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 159 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 160 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 161 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 162 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 163 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 164 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 165 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 166 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 167 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 168 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 169 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 170 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 171 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 172 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 173 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 174 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 175 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 176 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 177 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 178 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 179 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 180 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 181 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 182 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 183 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 184 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 185 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 186 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 187 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 188 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 189 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 190 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 191 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 192 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 193 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 194 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 195 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 196 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 197 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 198 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 199 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 200 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 201 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 202 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 203 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 204 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 205 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 206 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 207 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 208 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 209 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 210 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 211 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 212 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 213 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 214 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 215 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 216 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 217 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 218 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 219 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 220 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 221 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 222 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 223 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 224 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 225 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 226 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 227 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 228 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 229 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 230 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 231 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 232 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 233 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 234 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 235 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 236 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 237 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 238 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 239 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 240 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 241 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 242 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 243 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 244 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 245 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 246 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 247 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 248 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 249 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 250 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 251 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 252 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 253 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 254 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 255 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 256 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 257 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 258 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 259 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 260 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 261 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 262 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 263 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 264 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 265 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 266 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 267 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 268 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 269 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 270 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 271 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 272 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 273 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 274 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 275 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 276 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 277 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 278 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 279 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 280 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 281 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 282 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 283 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 284 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 285 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 286 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 287 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 288 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 289 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 290 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 291 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 292 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 293 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 294 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 295 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 296 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 297 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 298 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 299 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 300 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 301 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 302 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 303 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 304 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 305 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 306 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 307 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 308 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 309 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 310 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 311 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 312 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 313 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 314 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 315 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 316 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 317 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 318 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 319 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 320 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 321 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 322 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 323 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 324 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 325 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 326 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 327 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 328 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 329 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 330 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 331 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 332 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 333 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 334 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 335 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 336 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 337 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 338 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 339 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 340 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 341 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 342 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 343 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 344 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 345 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 346 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 347 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 348 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 349 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 350 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 351 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 352 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 353 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 354 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 355 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 356 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 357 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 358 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 359 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 360 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 361 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 362 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 363 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 364 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 365 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 366 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 367 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 368 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 369 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 370 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 371 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 372 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 373 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 374 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 375 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 376 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 377 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 378 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 379 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 380 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 381 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 382 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 383 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 384 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 385 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 386 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 387 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 388 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 389 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 390 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 391 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 392 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 393 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 394 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 395 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 396 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 397 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 398 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 399 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 400 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 401 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 402 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 403 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 404 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 405 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 406 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 407 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 408 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 409 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 410 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 411 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 412 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 413 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 414 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 415 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 416 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 417 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 418 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 419 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 420 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 421 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 422 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 423 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 424 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 425 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 426 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 427 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 428 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 429 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 430 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 431 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 432 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 433 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 434 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 435 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 436 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 437 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 438 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 439 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 440 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 441 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 442 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 443 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 444 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 445 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 446 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 447 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 448 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 449 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 450 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 451 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 452 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 453 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 454 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 455 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 456 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 457 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 458 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 459 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 460 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 461 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 462 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 463 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 464 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 465 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 466 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 467 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 468 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 469 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 470 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 471 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 472 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 473 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 474 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 475 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 476 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 477 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 478 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 479 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 480 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 481 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 482 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 483 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 484 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 485 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 486 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 487 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 488 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 489 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 490 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 491 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 492 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 493 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 494 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 495 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 496 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 497 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 498 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 499 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 500 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 501 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 502 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 503 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 504 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 505 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 506 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 507 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 508 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 509 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 510 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 511 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 512 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 513 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 514 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 515 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 516 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 517 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 518 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 519 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 520 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 521 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 522 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 523 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 524 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 525 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 526 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 527 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 528 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 529 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 530 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 531 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 532 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 533 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 534 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 535 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 536 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 537 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 538 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 539 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 540 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 541 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 542 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 543 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 544 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 545 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 546 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 547 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 548 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 549 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 550 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 551 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 552 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 553 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 554 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 555 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 556 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 557 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 558 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 559 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 560 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 561 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 562 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 563 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 564 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 565 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 566 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 567 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 568 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 569 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 570 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 571 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 572 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 573 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 574 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 575 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 576 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 577 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 578 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 579 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 580 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 581 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 582 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 583 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 584 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 585 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 586 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 587 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 588 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 589 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 590 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 591 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 592 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 593 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 594 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 595 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 596 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 597 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 598 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 599 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 600 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 601 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 602 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 603 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 604 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 605 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 606 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 607 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 608 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 609 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 610 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 611 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 612 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 613 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 614 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 615 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 616 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 617 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 618 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 619 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 620 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 621 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 622 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 623 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 624 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 625 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 626 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 627 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 628 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 629 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 630 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 631 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 632 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 633 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 634 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 635 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 636 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 637 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 638 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 639 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 640 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 641 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 642 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 643 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 644 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 645 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 646 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 647 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 648 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 649 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 650 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 651 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 652 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 653 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 654 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 655 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 656 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 657 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 658 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 659 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 660 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 661 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 662 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 663 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 664 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 665 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 666 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 667 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 668 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 669 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 670 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 671 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 672 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 673 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 674 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 675 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 676 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 677 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 678 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 679 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 680 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 681 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 682 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 683 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 684 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 685 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 686 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 687 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 688 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 689 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 690 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 691 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 692 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 693 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 694 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 695 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 696 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 697 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 698 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 699 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 700 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 701 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 702 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 703 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 704 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 705 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 706 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 707 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 708 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 709 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 710 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 711 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 712 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 713 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 714 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 715 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 716 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 717 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 718 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 719 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 720 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 721 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 722 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 723 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 724 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 725 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 726 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 727 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 728 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 729 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 730 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 731 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 732 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 733 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 734 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 735 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 736 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 737 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 738 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 739 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 740 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 741 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 742 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 743 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 744 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 745 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 746 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 747 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 748 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 749 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 750 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 751 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 752 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 753 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 754 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 755 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 756 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 757 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 758 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 759 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 760 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 761 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 762 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 763 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 764 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 765 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 766 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 767 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 768 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 769 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 770 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 771 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 772 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 773 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 774 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 775 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 776 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 777 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 778 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 779 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 780 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 781 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 782 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 783 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 784 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 785 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 786 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 787 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 788 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 789 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 790 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 791 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 792 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 793 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 794 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 795 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 796 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 797 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 798 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 799 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 800 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 801 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 802 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 803 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 804 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 805 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 806 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 807 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 808 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 809 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 810 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 811 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 812 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 813 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 814 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 815 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 816 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 817 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 818 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 819 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 820 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 821 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 822 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 823 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 824 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 825 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 826 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 827 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 828 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 829 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 830 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 831 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 832 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 833 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 834 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 835 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 836 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 837 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 838 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 839 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 840 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 841 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 842 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 843 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 844 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 845 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 846 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 847 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 848 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 849 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 850 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 851 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 852 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 853 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 854 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 855 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 856 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 857 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 858 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 859 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 860 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 861 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 862 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 863 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 864 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 865 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 866 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 867 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 868 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 869 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 870 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 871 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 872 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 873 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 874 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 875 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 876 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 877 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 878 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 879 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 880 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 881 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 882 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 883 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 884 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 885 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 886 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 887 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 888 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 889 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 890 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 891 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 892 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 893 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 894 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 895 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 896 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 897 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 898 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 899 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 900 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 901 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 902 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 903 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 904 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 905 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 906 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 907 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 908 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 909 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 910 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 911 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 912 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 913 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 914 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 915 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 916 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 917 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 918 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 919 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 920 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 921 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 922 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 923 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 924 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 925 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 926 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 927 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 928 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 929 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 930 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 931 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 932 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 933 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 934 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 935 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 936 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 937 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 938 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 939 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 940 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 941 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 942 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 943 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 944 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 945 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 946 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 947 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 948 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 949 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 950 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 951 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 952 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 953 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 954 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 955 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 956 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 957 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 958 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 959 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 960 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 961 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 962 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 963 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 964 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 965 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 966 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 967 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 968 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 969 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 970 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 971 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 972 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 973 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 974 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 975 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 976 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 977 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 978 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 979 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 980 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 981 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 982 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 983 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 984 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 985 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 986 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 987 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 988 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 989 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 990 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 991 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 992 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 993 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 994 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 995 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 996 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 997 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 998 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 999 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 1000 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 1001 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 1002 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 1003 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 1004 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 1005 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 1006 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 1007 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 1008 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 1009 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 1010 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 1011 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 1012 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 1013 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 1014 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 1015 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 1016 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 1017 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 1018 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 1019 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 1020 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 1021 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 1022 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.259.0> 2011-05-07 17:35:56 =============================================================================== ns_1@10.1.4.243:ns_janitor:152: Setting vbucket 1023 in "default" on 'ns_1@10.1.4.243' from dead to active. INFO REPORT <0.64.0> 2011-05-07 17:35:58 =============================================================================== ns_1@10.1.4.243:ns_cluster:90: handling add_node("10.1.4.244", 8091, ..) INFO REPORT <0.64.0> 2011-05-07 17:35:58 =============================================================================== ns_1@10.1.4.243:ns_cluster:300: Posting node info to engage_cluster on {"10.1.4.244", 8091}: {struct, [{availableStorage, {struct, [{hdd, [{struct, [{path,<<"/">>}, {sizeKBytes,7583436}, {usagePercent,18}]}, {struct, [{path,<<"/dev">>}, {sizeKBytes,508460}, {usagePercent,1}]}, {struct, [{path,<<"/dev/shm">>}, {sizeKBytes,513048}, {usagePercent,0}]}, {struct, [{path,<<"/var/run">>}, {sizeKBytes,513048}, {usagePercent,1}]}, {struct, [{path,<<"/var/lock">>}, {sizeKBytes,513048}, {usagePercent,0}]}, {struct, [{path,<<"/lib/init/rw">>}, {sizeKBytes,513048}, {usagePercent,0}]}, {struct, [{path,<<"/boot">>}, {sizeKBytes,233191}, {usagePercent,8}]}]}]}}, {memoryQuota,801}, {storageTotals, {struct, [{ram, {struct, [{usedByData,48800032}, {total,1050726400}, {quotaTotal,839909376}, {used,674267136}]}}, {hdd, {struct, [{usedByData,28912640}, {total,7765438464}, {quotaTotal,7765438464}, {used,1397778923}, {free,6367659541}]}}]}}, {storage, {struct, [{ssd,[]}, {hdd, [{struct, [{path,<<"/var/opt/membase/1.6.5.4r/data/ns_1">>}, {quotaMb,none}, {state,ok}]}]}]}}, {uptime,<<"12">>}, {memoryTotal,1050726400}, {memoryFree,376459264}, {mcdMemoryReserved,801}, {mcdMemoryAllocated,801}, {otpNode,<<"ns_1@10.1.4.243">>}, {otpCookie,<<"elyrdmyctqykhepq">>}, {clusterMembership,<<"active">>}, {status,<<"healthy">>}, {hostname,<<"10.1.4.243:8091">>}, {clusterCompatibility,1}, {version,<<"1.6.5.4r">>}, {os,<<"i686-pc-linux-gnu">>}, {ports,{struct,[{proxy,11211},{direct,11210}]}}]} PROGRESS REPORT <0.275.0> 2011-05-07 17:35:58 =============================================================================== supervisor {local,inets_sup} started [{pid,<0.276.0>}, {name,ftp_sup}, {mfa,{ftp_sup,start_link,[]}}, {restart_type,permanent}, {shutdown,infinity}, {child_type,supervisor}] PROGRESS REPORT <0.278.0> 2011-05-07 17:35:58 =============================================================================== supervisor {local,httpc_profile_sup} started [{pid,<0.279.0>}, {name,httpc_manager}, {mfa,{httpc_manager,start_link,[{default,only_session_cookies}]}}, {restart_type,permanent}, {shutdown,4000}, {child_type,worker}] PROGRESS REPORT <0.277.0> 2011-05-07 17:35:58 =============================================================================== supervisor {local,httpc_sup} started [{pid,<0.278.0>}, {name,httpc_profile_sup}, {mfa,{httpc_profile_sup,start_link, [[{httpc,{default,only_session_cookies}}]]}}, {restart_type,permanent}, {shutdown,infinity}, {child_type,supervisor}] PROGRESS REPORT <0.277.0> 2011-05-07 17:35:58 =============================================================================== supervisor {local,httpc_sup} started [{pid,<0.280.0>}, {name,httpc_handler_sup}, {mfa,{httpc_handler_sup,start_link,[]}}, {restart_type,permanent}, {shutdown,infinity}, {child_type,supervisor}] PROGRESS REPORT <0.275.0> 2011-05-07 17:35:58 =============================================================================== supervisor {local,inets_sup} started [{pid,<0.277.0>}, {name,httpc_sup}, {mfa,{httpc_sup,start_link, [[{httpc,{default,only_session_cookies}}]]}}, {restart_type,permanent}, {shutdown,infinity}, {child_type,supervisor}] PROGRESS REPORT <0.275.0> 2011-05-07 17:35:58 =============================================================================== supervisor {local,inets_sup} started [{pid,<0.281.0>}, {name,httpd_sup}, {mfa,{httpd_sup,start_link,[[]]}}, {restart_type,permanent}, {shutdown,infinity}, {child_type,supervisor}] PROGRESS REPORT <0.275.0> 2011-05-07 17:35:58 =============================================================================== supervisor {local,inets_sup} started [{pid,<0.282.0>}, {name,tftp_sup}, {mfa,{tftp_sup,start_link,[[]]}}, {restart_type,permanent}, {shutdown,infinity}, {child_type,supervisor}] PROGRESS REPORT <0.7.0> 2011-05-07 17:35:58 =============================================================================== application inets started_at 'ns_1@10.1.4.243' INFO REPORT <0.64.0> 2011-05-07 17:35:58 =============================================================================== ns_1@10.1.4.243:ns_cluster:306: Reply from engage_cluster on {"10.1.4.244", 8091}: {ok,{struct,[{<<"availableStorage">>, {struct,[{<<"hdd">>, [{struct,[{<<"path">>,<<"/">>}, {<<"sizeKBytes">>,7583436}, {<<"usagePercent">>,16}]}, {struct,[{<<"path">>,<<"/dev">>}, {<<"sizeKBytes">>,508460}, {<<"usagePercent">>,1}]}, {struct,[{<<"path">>,<<"/dev/shm">>}, {<<"sizeKBytes">>,513048}, {<<"usagePercent">>,0}]}, {struct,[{<<"path">>,<<"/var/run">>}, {<<"sizeKBytes">>,513048}, {<<"usagePercent">>,1}]}, {struct,[{<<"path">>,<<"/var/lock">>}, {<<"sizeKBytes">>,513048}, {<<"usagePercent">>,0}]}, {struct,[{<<"path">>,<<"/lib/init/rw">>}, {<<"sizeKBytes">>,513048}, {<<"usagePercent">>,0}]}, {struct,[{<<"path">>,<<"/boot">>}, {<<"sizeKBytes">>,233191}, {<<"usagePercent">>,8}]}]}]}}, {<<"memoryQuota">>,801}, {<<"storageTotals">>, {struct,[{<<"ram">>, {struct,[{<<"usedByData">>,0}, {<<"total">>,1050726400}, {<<"quotaTotal">>,839909376}, {<<"used">>,371363840}]}}, {<<"hdd">>, {struct,[{<<"usedByData">>,0}, {<<"total">>,7765438464.0}, {<<"quotaTotal">>,7765438464.0}, {<<"used">>,1242470154}, {<<"free">>,6522968310.0}]}}]}}, {<<"storage">>, {struct,[{<<"ssd">>,[]}, {<<"hdd">>, [{struct,[{<<"path">>, <<"/var/opt/membase/1.6.5.4r/data/ns_1">>}, {<<"quotaMb">>,<<"none">>}, {<<"state">>,<<"ok">>}]}]}]}}, {<<"uptime">>,<<"64">>}, {<<"memoryTotal">>,1050726400}, {<<"memoryFree">>,679362560}, {<<"mcdMemoryReserved">>,801}, {<<"mcdMemoryAllocated">>,801}, {<<"otpNode">>,<<"ns_1@10.1.4.244">>}, {<<"otpCookie">>,<<"bszkoopiimchraxv">>}, {<<"clusterMembership">>,<<"active">>}, {<<"status">>,<<"healthy">>}, {<<"hostname">>,<<"10.1.4.244:8091">>}, {<<"clusterCompatibility">>,1}, {<<"version">>,<<"1.6.5.4r">>}, {<<"os">>,<<"i686-pc-linux-gnu">>}, {<<"ports">>, {struct,[{<<"proxy">>,11211},{<<"direct">>,11210}]}}]}} INFO REPORT <0.64.0> 2011-05-07 17:35:58 =============================================================================== ns_1@10.1.4.243:ns_cluster:371: port_please("ns_1", "10.1.4.244") = 21100 INFO REPORT <0.64.0> 2011-05-07 17:35:58 =============================================================================== ns_1@10.1.4.243:ns_cluster:461: Started node add transaction by adding node 'ns_1@10.1.4.244' to nodes_wanted INFO REPORT <0.66.0> 2011-05-07 17:35:58 =============================================================================== ns_node_disco_conf_events config on nodes_wanted INFO REPORT <0.66.0> 2011-05-07 17:35:58 =============================================================================== config change: nodes_wanted -> ['ns_1@10.1.4.243','ns_1@10.1.4.244'] INFO REPORT <0.285.0> 2011-05-07 17:35:58 =============================================================================== ns_node_disco cookie_sync INFO REPORT <0.285.0> 2011-05-07 17:35:58 =============================================================================== ns_node_disco: nodes_wanted updated: ['ns_1@10.1.4.243','ns_1@10.1.4.244'], with cookie: elyrdmyctqykhepq INFO REPORT <0.66.0> 2011-05-07 17:35:58 =============================================================================== ns_node_disco_conf_events config all INFO REPORT <0.86.0> 2011-05-07 17:35:58 =============================================================================== Pushing config INFO REPORT <0.86.0> 2011-05-07 17:35:58 =============================================================================== Pushing config done INFO REPORT <0.66.0> 2011-05-07 17:35:58 =============================================================================== config change: {node,'ns_1@10.1.4.244',membership} -> inactiveAdded INFO REPORT <0.66.0> 2011-05-07 17:35:58 =============================================================================== ns_node_disco_conf_events config all INFO REPORT <0.86.0> 2011-05-07 17:35:58 =============================================================================== Pushing config INFO REPORT <0.86.0> 2011-05-07 17:35:58 =============================================================================== Pushing config done INFO REPORT <0.64.0> 2011-05-07 17:35:58 =============================================================================== ns_1@10.1.4.243:ns_cluster:431: Posting the following to complete_join on "10.1.4.244:8091": {struct, [{<<"targetNode">>,'ns_1@10.1.4.244'}, {availableStorage, {struct, [{hdd, [{struct, [{path,<<"/">>}, {sizeKBytes,7583436}, {usagePercent,18}]}, {struct, [{path,<<"/dev">>}, {sizeKBytes,508460}, {usagePercent,1}]}, {struct, [{path,<<"/dev/shm">>}, {sizeKBytes,513048}, {usagePercent,0}]}, {struct, [{path,<<"/var/run">>}, {sizeKBytes,513048}, {usagePercent,1}]}, {struct, [{path,<<"/var/lock">>}, {sizeKBytes,513048}, {usagePercent,0}]}, {struct, [{path,<<"/lib/init/rw">>}, {sizeKBytes,513048}, {usagePercent,0}]}, {struct, [{path,<<"/boot">>}, {sizeKBytes,233191}, {usagePercent,8}]}]}]}}, {memoryQuota,801}, {storageTotals, {struct, [{ram, {struct, [{usedByData,48800032}, {total,1050726400}, {quotaTotal,839909376}, {used,674267136}]}}, {hdd, {struct, [{usedByData,28912640}, {total,7765438464}, {quotaTotal,7765438464}, {used,1397778923}, {free,6367659541}]}}]}}, {storage, {struct, [{ssd,[]}, {hdd, [{struct, [{path,<<"/var/opt/membase/1.6.5.4r/data/ns_1">>}, {quotaMb,none}, {state,ok}]}]}]}}, {uptime,<<"12">>}, {memoryTotal,1050726400}, {memoryFree,376459264}, {mcdMemoryReserved,801}, {mcdMemoryAllocated,801}, {otpNode,<<"ns_1@10.1.4.243">>}, {otpCookie,<<"elyrdmyctqykhepq">>}, {clusterMembership,<<"active">>}, {status,<<"healthy">>}, {hostname,<<"10.1.4.243:8091">>}, {clusterCompatibility,1}, {version,<<"1.6.5.4r">>}, {os,<<"i686-pc-linux-gnu">>}, {ports,{struct,[{proxy,11211},{direct,11210}]}}]} INFO REPORT <0.285.0> 2011-05-07 17:35:58 =============================================================================== ns_node_disco: nodes_wanted pong: ['ns_1@10.1.4.243'], with cookie: elyrdmyctqykhepq INFO REPORT <0.80.0> 2011-05-07 17:35:58 =============================================================================== ns_log: logging ns_node_disco:4:Node 'ns_1@10.1.4.243' saw that node 'ns_1@10.1.4.244' came up. INFO REPORT <0.64.0> 2011-05-07 17:35:59 =============================================================================== ns_1@10.1.4.243:ns_cluster:437: Reply from complete_join on "10.1.4.244:8091": {ok,[]} INFO REPORT <0.64.0> 2011-05-07 17:35:59 =============================================================================== ns_1@10.1.4.243:ns_cluster:92: add_node("10.1.4.244", 8091, ..) -> {ok, 'ns_1@10.1.4.244'} INFO REPORT <0.79.0> 2011-05-07 17:36:03 =============================================================================== Detected a new node (from node 'ns_1@10.1.4.243'). Moving config around. INFO REPORT <0.79.0> 2011-05-07 17:36:03 =============================================================================== ns_node_disco_log: nodes changed: ['ns_1@10.1.4.243','ns_1@10.1.4.244'] INFO REPORT <0.86.0> 2011-05-07 17:36:03 =============================================================================== Pulling config INFO REPORT <0.86.0> 2011-05-07 17:36:03 =============================================================================== Pulling config from: 'ns_1@10.1.4.244' INFO REPORT <0.86.0> 2011-05-07 17:36:03 =============================================================================== Pulling config done INFO REPORT <0.113.0> 2011-05-07 17:36:12 =============================================================================== ns_log: logging ns_orchestrator:4:Starting rebalance, KeepNodes = ['ns_1@10.1.4.243','ns_1@10.1.4.244'], EjectNodes = [] INFO REPORT <0.66.0> 2011-05-07 17:36:12 =============================================================================== config change: {node,'ns_1@10.1.4.244',membership} -> active INFO REPORT <0.66.0> 2011-05-07 17:36:12 =============================================================================== config change: {node,'ns_1@10.1.4.243',membership} -> active INFO REPORT <0.66.0> 2011-05-07 17:36:12 =============================================================================== ns_node_disco_conf_events config all INFO REPORT <0.86.0> 2011-05-07 17:36:12 =============================================================================== Pushing config INFO REPORT <0.86.0> 2011-05-07 17:36:12 =============================================================================== Pushing config done INFO REPORT <0.66.0> 2011-05-07 17:36:12 =============================================================================== config change: rebalance_status -> running INFO REPORT <0.66.0> 2011-05-07 17:36:12 =============================================================================== ns_node_disco_conf_events config all INFO REPORT <0.86.0> 2011-05-07 17:36:12 =============================================================================== Pushing config INFO REPORT <0.86.0> 2011-05-07 17:36:12 =============================================================================== Pushing config done INFO REPORT <0.66.0> 2011-05-07 17:36:12 =============================================================================== config change: buckets -> [{configs,[{"default", [{num_replicas,1}, {ram_quota,839909376}, {auth_type,sasl}, {sasl_password,[]}, {type,membase}, {num_vbuckets,1024}, {db_shards,4}, {ht_size,3079}, {tap_keepalive,0}, {tap_noop_interval,20}, {max_txn_size,1000}, {ht_locks,5}, {servers,['ns_1@10.1.4.243','ns_1@10.1.4.244']}, {map,[['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243',undefined], ['ns_1@10.1.4.243'|...], [...]|...]}]}]}] INFO REPORT <0.66.0> 2011-05-07 17:36:12 =============================================================================== ns_node_disco_conf_events config all INFO REPORT <0.86.0> 2011-05-07 17:36:12 =============================================================================== Pushing config INFO REPORT <0.86.0> 2011-05-07 17:36:12 =============================================================================== Pushing config done ERROR REPORT <0.227.0> 2011-05-07 17:36:12 =============================================================================== ns_1@10.1.4.243:stats_reader:176: Some nodes didn't respond: ['ns_1@10.1.4.244'] ERROR REPORT <0.218.0> 2011-05-07 17:36:12 =============================================================================== ns_1@10.1.4.243:stats_reader:176: Some nodes didn't respond: ['ns_1@10.1.4.244'] ERROR REPORT <0.217.0> 2011-05-07 17:36:13 =============================================================================== ns_1@10.1.4.243:stats_reader:176: Some nodes didn't respond: ['ns_1@10.1.4.244'] ERROR REPORT <0.220.0> 2011-05-07 17:36:13 =============================================================================== ns_1@10.1.4.243:stats_reader:176: Some nodes didn't respond: ['ns_1@10.1.4.244'] ERROR REPORT <0.204.0> 2011-05-07 17:36:13 =============================================================================== ns_1@10.1.4.243:stats_reader:176: Some nodes didn't respond: ['ns_1@10.1.4.244'] INFO REPORT <0.375.0> 2011-05-07 17:36:13 =============================================================================== ns_1@10.1.4.243:ns_rebalancer:420: Waiting for ['ns_1@10.1.4.244'] INFO REPORT <0.375.0> 2011-05-07 17:36:14 =============================================================================== ns_1@10.1.4.243:ns_rebalancer:420: Waiting for ['ns_1@10.1.4.244'] INFO REPORT <0.375.0> 2011-05-07 17:36:15 =============================================================================== ns_1@10.1.4.243:ns_rebalancer:420: Waiting for ['ns_1@10.1.4.244'] INFO REPORT <0.113.0> 2011-05-07 17:36:15 =============================================================================== ns_1@10.1.4.243:ns_orchestrator:164: Skipping janitor in state rebalancing: {rebalancing_state, <0.375.0>, {dict, 2, 16, 16, 8, 80, 48, {[], [], [], [], [], [], [], [], [], [], [], [], [], [], [], []}, {{[], [], [], [['ns_1@10.1.4.243'| 0.0]], [['ns_1@10.1.4.244'| 0.0]], [], [], [], [], [], [], [], [], [], [], []}}}} INFO REPORT <0.411.0> 2011-05-07 17:36:16 =============================================================================== vbucketmigrator<0.411.0>: Connecting to {Sock 10.1.4.244:11210} vbucketmigrator<0.411.0>: Authenticating towards: {Sock 10.1.4.244:11210} vbucketmigrator<0.411.0>: Authenticated towards: {Sock 10.1.4.244:11210} vbucketmigrator<0.411.0>: Connecting to {Sock 10.1.4.243:11210} vbucketmigrator<0.411.0>: Authenticating towards: {Sock 10.1.4.243:11210} vbucketmigrator<0.411.0>: Authenticated towards: {Sock 10.1.4.243:11210} vbucketmigrator<0.411.0>: Starting to move bucket 0 INFO REPORT <0.106.0> 2011-05-07 17:36:16 =============================================================================== memcached<0.106.0>: Vbucket is going dead. INFO REPORT <0.411.0> 2011-05-07 17:36:22 =============================================================================== vbucketmigrator<0.411.0>: Bucket 0 moved to the next server INFO REPORT <0.86.0> 2011-05-07 17:36:25 =============================================================================== Pulling config from: 'ns_1@10.1.4.244' INFO REPORT <0.113.0> 2011-05-07 17:36:25 =============================================================================== ns_1@10.1.4.243:ns_orchestrator:164: Skipping janitor in state rebalancing: {rebalancing_state, <0.375.0>, {dict, 2, 16, 16, 8, 80, 48, {[], [], [], [], [], [], [], [], [], [], [], [], [], [], [], []}, {{[], [], [], [['ns_1@10.1.4.243'| 0.0]], [['ns_1@10.1.4.244'| 0.0]], [], [], [], [], [], [], [], [], [], [], []}}}} INFO REPORT <0.113.0> 2011-05-07 17:36:35 =============================================================================== ns_1@10.1.4.243:ns_orchestrator:164: Skipping janitor in state rebalancing: {rebalancing_state, <0.375.0>, {dict, 2, 16, 16, 8, 80, 48, {[], [], [], [], [], [], [], [], [], [], [], [], [], [], [], []}, {{[], [], [], [['ns_1@10.1.4.243'| 0.0]], [['ns_1@10.1.4.244'| 0.0]], [], [], [], [], [], [], [], [], [], [], []}}}} INFO REPORT <0.94.0> 2011-05-07 17:36:45 =============================================================================== ns_1@10.1.4.243:ns_doctor:82: Current node statuses: [{'ns_1@10.1.4.243', [{last_heard,{1304,815005,869406}}, {active_buckets,["default"]}, {memory, [{total,22846832}, {processes,11885244}, {processes_used,11872668}, {system,10961588}, {atom,530957}, {atom_used,514432}, {binary,228352}, {code,4206123}, {ets,1066628}]}, {cluster_compatibility_version,1}, {version, [{kernel,"2.13.4"}, {mnesia,"4.4.12"}, {stdlib,"1.16.4"}, {os_mon,"2.2.4"}, {inets,"5.2"}, {ns_server,"1.6.5.4r"}, {menelaus,"1.6.5.4r"}, {sasl,"2.1.8"}]}, {system_arch,"i686-pc-linux-gnu"}, {wall_clock,60}, {memory_data,{1050726400,786362368,{<0.171.0>,3328604}}}, {disk_data, [{"/",7583436,18}, {"/dev",508460,1}, {"/dev/shm",513048,0}, {"/var/run",513048,1}, {"/var/lock",513048,0}, {"/lib/init/rw",513048,0}, {"/boot",233191,8}]}, {meminfo, <<"MemTotal: 1026100 kB\nMemFree: 258068 kB\nBuffers: 152780 kB\nCached: 452128 kB\nSwapCached: 0 kB\nActive: 476188 kB\nInactive: 253288 kB\nActive(anon): 120396 kB\nInactive(anon): 4412 kB\nActive(file): 355792 kB\nInactive(file): 248876 kB\nUnevictable: 0 kB\nMlocked: 0 kB\nHighTotal: 137160 kB\nHighFree: 252 kB\nLowTotal: 888940 kB\nLowFree: 257816 kB\nSwapTotal: 397304 kB\nSwapFree: 397304 kB\nDirty: 172 kB\nWriteback: 0 kB\nAnonPages: 124568 kB\nMapped: 10484 kB\nShmem: 240 kB\nSlab: 28216 kB\nSReclaimable: 23296 kB\nSUnreclaim: 4920 kB\nKernelStack: 984 kB\nPageTables: 1104 kB\nNFS_Unstable: 0 kB\nBounce: 0 kB\nWritebackTmp: 0 kB\nCommitLimit: 910352 kB\nCommitted_AS: 346888 kB\nVmallocTotal: 122880 kB\nVmallocUsed: 3540 kB\nVmallocChunk: 114424 kB\nHardwareCorrupted: 0 kB\nHugePages_Total: 0\nHugePages_Free: 0\nHugePages_Rsvd: 0\nHugePages_Surp: 0\nHugepagesize: 2048 kB\nDirectMap4k: 10232 kB\nDirectMap2M: 901120 kB\n">>}, {replication,[{"default",1.0}]}, {system_memory_data, [{system_total_memory,1050726400}, {free_swap,406839296}, {total_swap,406839296}, {cached_memory,462979072}, {buffered_memory,156446720}, {free_memory,264261632}, {total_memory,1050726400}]}, {statistics, [{wall_clock,{50313,1}}, {context_switches,{34797,0}}, {garbage_collection,{7445,43338680,0}}, {io,{{input,7332098},{output,3099032}}}, {reductions,{11530939,786422}}, {run_queue,0}, {runtime,{1700,150}}]}]}, {'ns_1@10.1.4.244', [{last_heard,{1304,815004,930228}}, {active_buckets,["default"]}, {memory, [{total,13953920}, {processes,3999828}, {processes_used,3995588}, {system,9954092}, {atom,514761}, {atom_used,490124}, {binary,220592}, {code,3973911}, {ets,343780}]}, {cluster_compatibility_version,1}, {version, [{kernel,"2.13.4"}, {mnesia,"4.4.12"}, {stdlib,"1.16.4"}, {os_mon,"2.2.4"}, {ns_server,"1.6.5.4r"}, {menelaus,"1.6.5.4r"}, {sasl,"2.1.8"}]}, {system_arch,"i686-pc-linux-gnu"}, {wall_clock,111}, {memory_data,{1050726400,371363840,{<6517.70.0>,486376}}}, {disk_data, [{"/",7583436,16}, {"/dev",508460,1}, {"/dev/shm",513048,0}, {"/var/run",513048,1}, {"/var/lock",513048,0}, {"/lib/init/rw",513048,0}, {"/boot",233191,8}]}, {meminfo, <<"MemTotal: 1026100 kB\nMemFree: 627268 kB\nBuffers: 51072 kB\nCached: 253484 kB\nSwapCached: 0 kB\nActive: 164612 kB\nInactive: 194280 kB\nActive(anon): 48416 kB\nInactive(anon): 6152 kB\nActive(file): 116196 kB\nInactive(file): 188128 kB\nUnevictable: 0 kB\nMlocked: 0 kB\nHighTotal: 137160 kB\nHighFree: 616 kB\nLowTotal: 888940 kB\nLowFree: 626652 kB\nSwapTotal: 397304 kB\nSwapFree: 397304 kB\nDirty: 316 kB\nWriteback: 0 kB\nAnonPages: 54336 kB\nMapped: 10036 kB\nShmem: 232 kB\nSlab: 30512 kB\nSReclaimable: 25684 kB\nSUnreclaim: 4828 kB\nKernelStack: 944 kB\nPageTables: 856 kB\nNFS_Unstable: 0 kB\nBounce: 0 kB\nWritebackTmp: 0 kB\nCommitLimit: 910352 kB\nCommitted_AS: 252496 kB\nVmallocTotal: 122880 kB\nVmallocUsed: 3540 kB\nVmallocChunk: 114844 kB\nHardwareCorrupted: 0 kB\nHugePages_Total: 0\nHugePages_Free: 0\nHugePages_Rsvd: 0\nHugePages_Surp: 0\nHugepagesize: 2048 kB\nDirectMap4k: 10232 kB\nDirectMap2M: 901120 kB\n">>}, {replication,[{"default",1.0}]}, {system_memory_data, [{system_total_memory,1050726400}, {free_swap,406839296}, {total_swap,406839296}, {cached_memory,259567616}, {buffered_memory,52297728}, {free_memory,642322432}, {total_memory,1050726400}]}, {statistics, [{wall_clock,{105008,1}}, {context_switches,{20149,0}}, {garbage_collection,{4415,9846519,0}}, {io,{{input,4511228},{output,714870}}}, {reductions,{3481992,231708}}, {run_queue,0}, {runtime,{820,40}}]}]}] INFO REPORT <0.113.0> 2011-05-07 17:36:45 =============================================================================== ns_1@10.1.4.243:ns_orchestrator:164: Skipping janitor in state rebalancing: {rebalancing_state, <0.375.0>, {dict, 2, 16, 16, 8, 80, 48, {[], [], [], [], [], [], [], [], [], [], [], [], [], [], [], []}, {{[], [], [], [['ns_1@10.1.4.243'| 0.0]], [['ns_1@10.1.4.244'| 0.0]], [], [], [], [], [], [], [], [], [], [], []}}}} INFO REPORT <0.113.0> 2011-05-07 17:36:55 =============================================================================== ns_1@10.1.4.243:ns_orchestrator:164: Skipping janitor in state rebalancing: {rebalancing_state, <0.375.0>, {dict, 2, 16, 16, 8, 80, 48, {[], [], [], [], [], [], [], [], [], [], [], [], [], [], [], []}, {{[], [], [], [['ns_1@10.1.4.243'| 0.0]], [['ns_1@10.1.4.244'| 0.0]], [], [], [], [], [], [], [], [], [], [], []}}}} PROGRESS REPORT <0.34.0> 2011-05-07 17:37:01 =============================================================================== supervisor {local,sasl_sup} started [{pid,<0.599.0>}, {name,rb_server}, {mfa,{rb,start_link, [[{report_dir,"/var/opt/membase/1.6.5.4r/logs"}, {start_log,"/opt/membase/1.6.5.4r/tmp/nslogs837591_8336.log"}, {type,all}, {max,all}, {report_dir}]]}}, {restart_type,temporary}, {shutdown,brutal_kill}, {child_type,worker}] INFO REPORT <0.113.0> 2011-05-07 17:37:05 =============================================================================== ns_1@10.1.4.243:ns_orchestrator:164: Skipping janitor in state rebalancing: {rebalancing_state, <0.375.0>, {dict, 2, 16, 16, 8, 80, 48, {[], [], [], [], [], [], [], [], [], [], [], [], [], [], [], []}, {{[], [], [], [['ns_1@10.1.4.243'| 0.0]], [['ns_1@10.1.4.244'| 0.0]], [], [], [], [], [], [], [], [], [], [], []}}}} INFO REPORT <0.86.0> 2011-05-07 17:37:10 =============================================================================== Pulling config from: 'ns_1@10.1.4.244' INFO REPORT <0.113.0> 2011-05-07 17:37:15 =============================================================================== ns_1@10.1.4.243:ns_orchestrator:164: Skipping janitor in state rebalancing: {rebalancing_state, <0.375.0>, {dict, 2, 16, 16, 8, 80, 48, {[], [], [], [], [], [], [], [], [], [], [], [], [], [], [], []}, {{[], [], [], [['ns_1@10.1.4.243'| 0.0]], [['ns_1@10.1.4.244'| 0.0]], [], [], [], [], [], [], [], [], [], [], []}}}} INFO REPORT <0.113.0> 2011-05-07 17:37:25 =============================================================================== ns_1@10.1.4.243:ns_orchestrator:164: Skipping janitor in state rebalancing: {rebalancing_state, <0.375.0>, {dict, 2, 16, 16, 8, 80, 48, {[], [], [], [], [], [], [], [], [], [], [], [], [], [], [], []}, {{[], [], [], [['ns_1@10.1.4.243'| 0.0]], [['ns_1@10.1.4.244'| 0.0]], [], [], [], [], [], [], [], [], [], [], []}}}} INFO REPORT <0.167.0> 2011-05-07 17:37:29 =============================================================================== ns_1@10.1.4.243:stats_collector:71: Stats for bucket "default": auth_cmds 5 auth_errors 0 bucket_conns 6 bytes_read 33928 bytes_written 919058 cas_badval 0 cas_hits 0 cas_misses 0 cmd_flush 0 cmd_get 0 cmd_set 0 conn_yields 0 connection_structures 16 curr_connections 16 curr_items 204600 curr_items_tot 204800 daemon_connections 10 decr_hits 0 decr_misses 0 delete_hits 0 delete_misses 0 ep_bg_fetched 0 ep_commit_num 0 ep_commit_time 0 ep_commit_time_total 0 ep_data_age 0 ep_data_age_highwat 0 ep_db_cleaner_status complete ep_db_strategy multiMTVBDB ep_dbinit 1 ep_dbname /var/opt/membase/1.6.5.4r/data/ns_1/default ep_dbshards 4 ep_expired 0 ep_flush_duration 0 ep_flush_duration_highwat 0 ep_flush_duration_total 0 ep_flush_preempts 0 ep_flusher_state running ep_flusher_todo 0 ep_io_num_read 205000 ep_io_num_write 0 ep_io_read_bytes 23370000 ep_io_write_bytes 0 ep_item_begin_failed 0 ep_item_commit_failed 0 ep_item_flush_expired 0 ep_item_flush_failed 0 ep_kv_size 35840000 ep_max_data_size 839909376 ep_max_txn_size 1000 ep_mem_high_wat 629932032 ep_mem_low_wat 503945625 ep_min_data_age 0 ep_num_active_non_resident 0 ep_num_eject_failures 0 ep_num_eject_replicas 0 ep_num_expiry_pager_runs 0 ep_num_non_resident 0 ep_num_not_my_vbuckets 0 ep_num_pager_runs 0 ep_num_value_ejects 0 ep_oom_errors 0 ep_overhead 12960032 ep_pending_ops 0 ep_pending_ops_max 0 ep_pending_ops_max_duration 0 ep_pending_ops_total 0 ep_queue_age_cap 900 ep_queue_size 0 ep_storage_age 0 ep_storage_age_highwat 0 ep_storage_type featured ep_store_max_concurrency 10 ep_store_max_readers 9 ep_store_max_readwrite 1 ep_tap_bg_fetch_requeued 0 ep_tap_bg_fetched 200 ep_tap_keepalive 300 ep_tmp_oom_errors 0 ep_too_old 0 ep_too_young 0 ep_total_cache_size 35840000 ep_total_del_items 0 ep_total_enqueued 0 ep_total_new_items 0 ep_total_persisted 0 ep_vbucket_del 0 ep_vbucket_del_fail 0 ep_version 1.6.5.3.1_8_gce31baa ep_warmed_up 204800 ep_warmup true ep_warmup_dups 0 ep_warmup_oom 0 ep_warmup_thread complete ep_warmup_time 809268 get_hits 0 get_misses 0 incr_hits 0 incr_misses 0 libevent 2.0.7-rc limit_maxbytes 67108864 mem_used 48800032 pid 8362 pointer_size 32 rejected_conns 0 rusage_system 0.284017 rusage_user 1.140071 tap_connect_received 1 tap_mutation_sent 400 tap_opaque_sent 1 tap_vbucket_set_sent 2 threads 4 time 1304815049 total_connections 16 uptime 106 version 1.4.4_382_g9df3289 INFO REPORT <0.113.0> 2011-05-07 17:37:35 =============================================================================== ns_1@10.1.4.243:ns_orchestrator:164: Skipping janitor in state rebalancing: {rebalancing_state, <0.375.0>, {dict, 2, 16, 16, 8, 80, 48, {[], [], [], [], [], [], [], [], [], [], [], [], [], [], [], []}, {{[], [], [], [['ns_1@10.1.4.243'| 0.0]], [['ns_1@10.1.4.244'| 0.0]], [], [], [], [], [], [], [], [], [], [], []}}}} INFO REPORT <0.86.0> 2011-05-07 17:37:42 =============================================================================== Pulling config from: 'ns_1@10.1.4.244' INFO REPORT <0.94.0> 2011-05-07 17:37:45 =============================================================================== ns_1@10.1.4.243:ns_doctor:82: Current node statuses: [{'ns_1@10.1.4.243', [{last_heard,{1304,815065,869391}}, {active_buckets,["default"]}, {memory, [{total,29049152}, {processes,17850884}, {processes_used,17838308}, {system,11198268}, {atom,536281}, {atom_used,522118}, {binary,200304}, {code,4299333}, {ets,1104364}]}, {cluster_compatibility_version,1}, {version, [{kernel,"2.13.4"}, {mnesia,"4.4.12"}, {stdlib,"1.16.4"}, {os_mon,"2.2.4"}, {inets,"5.2"}, {ns_server,"1.6.5.4r"}, {menelaus,"1.6.5.4r"}, {sasl,"2.1.8"}]}, {system_arch,"i686-pc-linux-gnu"}, {wall_clock,120}, {memory_data,{1050726400,794836992,{<0.539.0>,6171264}}}, {disk_data, [{"/",7583436,18}, {"/dev",508460,1}, {"/dev/shm",513048,0}, {"/var/run",513048,1}, {"/var/lock",513048,0}, {"/lib/init/rw",513048,0}, {"/boot",233191,8}]}, {meminfo, <<"MemTotal: 1026100 kB\nMemFree: 249644 kB\nBuffers: 152784 kB\nCached: 452388 kB\nSwapCached: 0 kB\nActive: 485784 kB\nInactive: 252724 kB\nActive(anon): 129176 kB\nInactive(anon): 4412 kB\nActive(file): 356608 kB\nInactive(file): 248312 kB\nUnevictable: 0 kB\nMlocked: 0 kB\nHighTotal: 137160 kB\nHighFree: 252 kB\nLowTotal: 888940 kB\nLowFree: 249392 kB\nSwapTotal: 397304 kB\nSwapFree: 397304 kB\nDirty: 88 kB\nWriteback: 0 kB\nAnonPages: 133300 kB\nMapped: 10488 kB\nShmem: 240 kB\nSlab: 28224 kB\nSReclaimable: 23304 kB\nSUnreclaim: 4920 kB\nKernelStack: 976 kB\nPageTables: 1116 kB\nNFS_Unstable: 0 kB\nBounce: 0 kB\nWritebackTmp: 0 kB\nCommitLimit: 910352 kB\nCommitted_AS: 355724 kB\nVmallocTotal: 122880 kB\nVmallocUsed: 3540 kB\nVmallocChunk: 114424 kB\nHardwareCorrupted: 0 kB\nHugePages_Total: 0\nHugePages_Free: 0\nHugePages_Rsvd: 0\nHugePages_Surp: 0\nHugepagesize: 2048 kB\nDirectMap4k: 10232 kB\nDirectMap2M: 901120 kB\n">>}, {replication,[{"default",1.0}]}, {system_memory_data, [{system_total_memory,1050726400}, {free_swap,406839296}, {total_swap,406839296}, {cached_memory,463245312}, {buffered_memory,156450816}, {free_memory,255635456}, {total_memory,1050726400}]}, {statistics, [{wall_clock,{110313,1}}, {context_switches,{200565,0}}, {garbage_collection,{46233,132047833,0}}, {io,{{input,72830142},{output,7735598}}}, {reductions,{40167280,776515}}, {run_queue,0}, {runtime,{4860,160}}]}]}, {'ns_1@10.1.4.244', [{last_heard,{1304,815064,929368}}, {active_buckets,["default"]}, {memory, [{total,14891864}, {processes,4888916}, {processes_used,4884676}, {system,10002948}, {atom,515245}, {atom_used,490161}, {binary,205416}, {code,3973911}, {ets,407060}]}, {cluster_compatibility_version,1}, {version, [{kernel,"2.13.4"}, {mnesia,"4.4.12"}, {stdlib,"1.16.4"}, {os_mon,"2.2.4"}, {ns_server,"1.6.5.4r"}, {menelaus,"1.6.5.4r"}, {sasl,"2.1.8"}]}, {system_arch,"i686-pc-linux-gnu"}, {wall_clock,171}, {memory_data,{1050726400,408743936,{<6517.255.0>,901104}}}, {disk_data, [{"/",7583436,16}, {"/dev",508460,1}, {"/dev/shm",513048,0}, {"/var/run",513048,1}, {"/var/lock",513048,0}, {"/lib/init/rw",513048,0}, {"/boot",233191,8}]}, {meminfo, <<"MemTotal: 1026100 kB\nMemFree: 625796 kB\nBuffers: 51168 kB\nCached: 253736 kB\nSwapCached: 0 kB\nActive: 165800 kB\nInactive: 194532 kB\nActive(anon): 49532 kB\nInactive(anon): 6152 kB\nActive(file): 116268 kB\nInactive(file): 188380 kB\nUnevictable: 0 kB\nMlocked: 0 kB\nHighTotal: 137160 kB\nHighFree: 252 kB\nLowTotal: 888940 kB\nLowFree: 625544 kB\nSwapTotal: 397304 kB\nSwapFree: 397304 kB\nDirty: 64 kB\nWriteback: 0 kB\nAnonPages: 55488 kB\nMapped: 10040 kB\nShmem: 232 kB\nSlab: 30504 kB\nSReclaimable: 25676 kB\nSUnreclaim: 4828 kB\nKernelStack: 936 kB\nPageTables: 860 kB\nNFS_Unstable: 0 kB\nBounce: 0 kB\nWritebackTmp: 0 kB\nCommitLimit: 910352 kB\nCommitted_AS: 256132 kB\nVmallocTotal: 122880 kB\nVmallocUsed: 3540 kB\nVmallocChunk: 114844 kB\nHardwareCorrupted: 0 kB\nHugePages_Total: 0\nHugePages_Free: 0\nHugePages_Rsvd: 0\nHugePages_Surp: 0\nHugepagesize: 2048 kB\nDirectMap4k: 10232 kB\nDirectMap2M: 901120 kB\n">>}, {replication,[{"default",1.0}]}, {system_memory_data, [{system_total_memory,1050726400}, {free_swap,406839296}, {total_swap,406839296}, {cached_memory,259825664}, {buffered_memory,52396032}, {free_memory,640815104}, {total_memory,1050726400}]}, {statistics, [{wall_clock,{165008,1}}, {context_switches,{26614,0}}, {garbage_collection,{5884,15620046,0}}, {io,{{input,4712800},{output,975757}}}, {reductions,{5977379,249500}}, {run_queue,0}, {runtime,{1140,40}}]}]}] INFO REPORT <0.113.0> 2011-05-07 17:37:45 =============================================================================== ns_1@10.1.4.243:ns_orchestrator:164: Skipping janitor in state rebalancing: {rebalancing_state, <0.375.0>, {dict, 2, 16, 16, 8, 80, 48, {[], [], [], [], [], [], [], [], [], [], [], [], [], [], [], []}, {{[], [], [], [['ns_1@10.1.4.243'| 0.0]], [['ns_1@10.1.4.244'| 0.0]], [], [], [], [], [], [], [], [], [], [], []}}}} INFO REPORT <0.113.0> 2011-05-07 17:37:55 =============================================================================== ns_1@10.1.4.243:ns_orchestrator:164: Skipping janitor in state rebalancing: {rebalancing_state, <0.375.0>, {dict, 2, 16, 16, 8, 80, 48, {[], [], [], [], [], [], [], [], [], [], [], [], [], [], [], []}, {{[], [], [], [['ns_1@10.1.4.243'| 0.0]], [['ns_1@10.1.4.244'| 0.0]], [], [], [], [], [], [], [], [], [], [], []}}}} INFO REPORT <0.113.0> 2011-05-07 17:38:05 =============================================================================== ns_1@10.1.4.243:ns_orchestrator:164: Skipping janitor in state rebalancing: {rebalancing_state, <0.375.0>, {dict, 2, 16, 16, 8, 80, 48, {[], [], [], [], [], [], [], [], [], [], [], [], [], [], [], []}, {{[], [], [], [['ns_1@10.1.4.243'| 0.0]], [['ns_1@10.1.4.244'| 0.0]], [], [], [], [], [], [], [], [], [], [], []}}}} INFO REPORT <0.113.0> 2011-05-07 17:38:15 =============================================================================== ns_1@10.1.4.243:ns_orchestrator:164: Skipping janitor in state rebalancing: {rebalancing_state, <0.375.0>, {dict, 2, 16, 16, 8, 80, 48, {[], [], [], [], [], [], [], [], [], [], [], [], [], [], [], []}, {{[], [], [], [['ns_1@10.1.4.243'| 0.0]], [['ns_1@10.1.4.244'| 0.0]], [], [], [], [], [], [], [], [], [], [], []}}}} INFO REPORT <0.86.0> 2011-05-07 17:38:20 =============================================================================== Pulling config from: 'ns_1@10.1.4.244' INFO REPORT <0.113.0> 2011-05-07 17:38:25 =============================================================================== ns_1@10.1.4.243:ns_orchestrator:164: Skipping janitor in state rebalancing: {rebalancing_state, <0.375.0>, {dict, 2, 16, 16, 8, 80, 48, {[], [], [], [], [], [], [], [], [], [], [], [], [], [], [], []}, {{[], [], [], [['ns_1@10.1.4.243'| 0.0]], [['ns_1@10.1.4.244'| 0.0]], [], [], [], [], [], [], [], [], [], [], []}}}} INFO REPORT <0.113.0> 2011-05-07 17:38:35 =============================================================================== ns_1@10.1.4.243:ns_orchestrator:164: Skipping janitor in state rebalancing: {rebalancing_state, <0.375.0>, {dict, 2, 16, 16, 8, 80, 48, {[], [], [], [], [], [], [], [], [], [], [], [], [], [], [], []}, {{[], [], [], [['ns_1@10.1.4.243'| 0.0]], [['ns_1@10.1.4.244'| 0.0]], [], [], [], [], [], [], [], [], [], [], []}}}} INFO REPORT <0.94.0> 2011-05-07 17:38:45 =============================================================================== ns_1@10.1.4.243:ns_doctor:82: Current node statuses: [{'ns_1@10.1.4.243', [{last_heard,{1304,815125,869463}}, {active_buckets,["default"]}, {memory, [{total,29402952}, {processes,18168460}, {processes_used,18155932}, {system,11234492}, {atom,536281}, {atom_used,522118}, {binary,203096}, {code,4299333}, {ets,1137540}]}, {cluster_compatibility_version,1}, {version, [{kernel,"2.13.4"}, {mnesia,"4.4.12"}, {stdlib,"1.16.4"}, {os_mon,"2.2.4"}, {inets,"5.2"}, {ns_server,"1.6.5.4r"}, {menelaus,"1.6.5.4r"}, {sasl,"2.1.8"}]}, {system_arch,"i686-pc-linux-gnu"}, {wall_clock,180}, {memory_data,{1050726400,792424448,{<0.539.0>,6171264}}}, {disk_data, [{"/",7583436,18}, {"/dev",508460,1}, {"/dev/shm",513048,0}, {"/var/run",513048,1}, {"/var/lock",513048,0}, {"/lib/init/rw",513048,0}, {"/boot",233191,8}]}, {meminfo, <<"MemTotal: 1026100 kB\nMemFree: 250264 kB\nBuffers: 152784 kB\nCached: 452628 kB\nSwapCached: 0 kB\nActive: 484416 kB\nInactive: 252864 kB\nActive(anon): 127696 kB\nInactive(anon): 4412 kB\nActive(file): 356720 kB\nInactive(file): 248452 kB\nUnevictable: 0 kB\nMlocked: 0 kB\nHighTotal: 137160 kB\nHighFree: 252 kB\nLowTotal: 888940 kB\nLowFree: 250012 kB\nSwapTotal: 397304 kB\nSwapFree: 397304 kB\nDirty: 84 kB\nWriteback: 0 kB\nAnonPages: 131868 kB\nMapped: 10488 kB\nShmem: 240 kB\nSlab: 28224 kB\nSReclaimable: 23304 kB\nSUnreclaim: 4920 kB\nKernelStack: 976 kB\nPageTables: 1120 kB\nNFS_Unstable: 0 kB\nBounce: 0 kB\nWritebackTmp: 0 kB\nCommitLimit: 910352 kB\nCommitted_AS: 356968 kB\nVmallocTotal: 122880 kB\nVmallocUsed: 3540 kB\nVmallocChunk: 114424 kB\nHardwareCorrupted: 0 kB\nHugePages_Total: 0\nHugePages_Free: 0\nHugePages_Rsvd: 0\nHugePages_Surp: 0\nHugepagesize: 2048 kB\nDirectMap4k: 10232 kB\nDirectMap2M: 901120 kB\n">>}, {replication,[{"default",1.0}]}, {system_memory_data, [{system_total_memory,1050726400}, {free_swap,406839296}, {total_swap,406839296}, {cached_memory,463491072}, {buffered_memory,156450816}, {free_memory,256270336}, {total_memory,1050726400}]}, {statistics, [{wall_clock,{170313,1}}, {context_switches,{216079,0}}, {garbage_collection,{49850,166444459,0}}, {io,{{input,73116759},{output,8269670}}}, {reductions,{44834264,742166}}, {run_queue,0}, {runtime,{5800,160}}]}]}, {'ns_1@10.1.4.244', [{last_heard,{1304,815124,930152}}, {active_buckets,["default"]}, {memory, [{total,14917488}, {processes,4893716}, {processes_used,4889476}, {system,10023772}, {atom,515245}, {atom_used,490161}, {binary,191552}, {code,3973911}, {ets,441556}]}, {cluster_compatibility_version,1}, {version, [{kernel,"2.13.4"}, {mnesia,"4.4.12"}, {stdlib,"1.16.4"}, {os_mon,"2.2.4"}, {ns_server,"1.6.5.4r"}, {menelaus,"1.6.5.4r"}, {sasl,"2.1.8"}]}, {system_arch,"i686-pc-linux-gnu"}, {wall_clock,231}, {memory_data,{1050726400,409911296,{<6517.255.0>,901104}}}, {disk_data, [{"/",7583436,16}, {"/dev",508460,1}, {"/dev/shm",513048,0}, {"/var/run",513048,1}, {"/var/lock",513048,0}, {"/lib/init/rw",513048,0}, {"/boot",233191,8}]}, {meminfo, <<"MemTotal: 1026100 kB\nMemFree: 625300 kB\nBuffers: 51260 kB\nCached: 253984 kB\nSwapCached: 0 kB\nActive: 166012 kB\nInactive: 194812 kB\nActive(anon): 49676 kB\nInactive(anon): 6152 kB\nActive(file): 116336 kB\nInactive(file): 188660 kB\nUnevictable: 0 kB\nMlocked: 0 kB\nHighTotal: 137160 kB\nHighFree: 252 kB\nLowTotal: 888940 kB\nLowFree: 625048 kB\nSwapTotal: 397304 kB\nSwapFree: 397304 kB\nDirty: 36 kB\nWriteback: 0 kB\nAnonPages: 55636 kB\nMapped: 10040 kB\nShmem: 232 kB\nSlab: 30516 kB\nSReclaimable: 25684 kB\nSUnreclaim: 4832 kB\nKernelStack: 936 kB\nPageTables: 860 kB\nNFS_Unstable: 0 kB\nBounce: 0 kB\nWritebackTmp: 0 kB\nCommitLimit: 910352 kB\nCommitted_AS: 256132 kB\nVmallocTotal: 122880 kB\nVmallocUsed: 3540 kB\nVmallocChunk: 114844 kB\nHardwareCorrupted: 0 kB\nHugePages_Total: 0\nHugePages_Free: 0\nHugePages_Rsvd: 0\nHugePages_Surp: 0\nHugepagesize: 2048 kB\nDirectMap4k: 10232 kB\nDirectMap2M: 901120 kB\n">>}, {replication,[{"default",1.0}]}, {system_memory_data, [{system_total_memory,1050726400}, {free_swap,406839296}, {total_swap,406839296}, {cached_memory,260079616}, {buffered_memory,52490240}, {free_memory,640307200}, {total_memory,1050726400}]}, {statistics, [{wall_clock,{225008,1}}, {context_switches,{32560,0}}, {garbage_collection,{7366,20221908,0}}, {io,{{input,4924122},{output,1239942}}}, {reductions,{7471927,248592}}, {run_queue,0}, {runtime,{1460,50}}]}]}] INFO REPORT <0.113.0> 2011-05-07 17:38:45 =============================================================================== ns_1@10.1.4.243:ns_orchestrator:164: Skipping janitor in state rebalancing: {rebalancing_state, <0.375.0>, {dict, 2, 16, 16, 8, 80, 48, {[], [], [], [], [], [], [], [], [], [], [], [], [], [], [], []}, {{[], [], [], [['ns_1@10.1.4.243'| 0.0]], [['ns_1@10.1.4.244'| 0.0]], [], [], [], [], [], [], [], [], [], [], []}}}} INFO REPORT <0.113.0> 2011-05-07 17:38:55 =============================================================================== ns_1@10.1.4.243:ns_orchestrator:164: Skipping janitor in state rebalancing: {rebalancing_state, <0.375.0>, {dict, 2, 16, 16, 8, 80, 48, {[], [], [], [], [], [], [], [], [], [], [], [], [], [], [], []}, {{[], [], [], [['ns_1@10.1.4.243'| 0.0]], [['ns_1@10.1.4.244'| 0.0]], [], [], [], [], [], [], [], [], [], [], []}}}} INFO REPORT <0.86.0> 2011-05-07 17:39:02 =============================================================================== Pulling config from: 'ns_1@10.1.4.244' INFO REPORT <0.113.0> 2011-05-07 17:39:05 =============================================================================== ns_1@10.1.4.243:ns_orchestrator:164: Skipping janitor in state rebalancing: {rebalancing_state, <0.375.0>, {dict, 2, 16, 16, 8, 80, 48, {[], [], [], [], [], [], [], [], [], [], [], [], [], [], [], []}, {{[], [], [], [['ns_1@10.1.4.243'| 0.0]], [['ns_1@10.1.4.244'| 0.0]], [], [], [], [], [], [], [], [], [], [], []}}}} INFO REPORT <0.167.0> 2011-05-07 17:39:09 =============================================================================== ns_1@10.1.4.243:stats_collector:71: Stats for bucket "default": auth_cmds 5 auth_errors 0 bucket_conns 6 bytes_read 38674 bytes_written 1698702 cas_badval 0 cas_hits 0 cas_misses 0 cmd_flush 0 cmd_get 0 cmd_set 0 conn_yields 0 connection_structures 16 curr_connections 16 curr_items 204600 curr_items_tot 204800 daemon_connections 10 decr_hits 0 decr_misses 0 delete_hits 0 delete_misses 0 ep_bg_fetched 0 ep_commit_num 0 ep_commit_time 0 ep_commit_time_total 0 ep_data_age 0 ep_data_age_highwat 0 ep_db_cleaner_status complete ep_db_strategy multiMTVBDB ep_dbinit 1 ep_dbname /var/opt/membase/1.6.5.4r/data/ns_1/default ep_dbshards 4 ep_expired 0 ep_flush_duration 0 ep_flush_duration_highwat 0 ep_flush_duration_total 0 ep_flush_preempts 0 ep_flusher_state running ep_flusher_todo 0 ep_io_num_read 205000 ep_io_num_write 0 ep_io_read_bytes 23370000 ep_io_write_bytes 0 ep_item_begin_failed 0 ep_item_commit_failed 0 ep_item_flush_expired 0 ep_item_flush_failed 0 ep_kv_size 35840000 ep_max_data_size 839909376 ep_max_txn_size 1000 ep_mem_high_wat 629932032 ep_mem_low_wat 503945625 ep_min_data_age 0 ep_num_active_non_resident 0 ep_num_eject_failures 0 ep_num_eject_replicas 0 ep_num_expiry_pager_runs 0 ep_num_non_resident 0 ep_num_not_my_vbuckets 0 ep_num_pager_runs 0 ep_num_value_ejects 0 ep_oom_errors 0 ep_overhead 12960032 ep_pending_ops 0 ep_pending_ops_max 0 ep_pending_ops_max_duration 0 ep_pending_ops_total 0 ep_queue_age_cap 900 ep_queue_size 0 ep_storage_age 0 ep_storage_age_highwat 0 ep_storage_type featured ep_store_max_concurrency 10 ep_store_max_readers 9 ep_store_max_readwrite 1 ep_tap_bg_fetch_requeued 0 ep_tap_bg_fetched 200 ep_tap_keepalive 300 ep_tmp_oom_errors 0 ep_too_old 0 ep_too_young 0 ep_total_cache_size 35840000 ep_total_del_items 0 ep_total_enqueued 0 ep_total_new_items 0 ep_total_persisted 0 ep_vbucket_del 0 ep_vbucket_del_fail 0 ep_version 1.6.5.3.1_8_gce31baa ep_warmed_up 204800 ep_warmup true ep_warmup_dups 0 ep_warmup_oom 0 ep_warmup_thread complete ep_warmup_time 809268 get_hits 0 get_misses 0 incr_hits 0 incr_misses 0 libevent 2.0.7-rc limit_maxbytes 67108864 mem_used 48800032 pid 8362 pointer_size 32 rejected_conns 0 rusage_system 0.296018 rusage_user 1.168073 tap_connect_received 1 tap_mutation_sent 400 tap_opaque_sent 1 tap_vbucket_set_sent 2 threads 4 time 1304815148 total_connections 16 uptime 205 version 1.4.4_382_g9df3289 INFO REPORT <0.113.0> 2011-05-07 17:39:15 =============================================================================== ns_1@10.1.4.243:ns_orchestrator:164: Skipping janitor in state rebalancing: {rebalancing_state, <0.375.0>, {dict, 2, 16, 16, 8, 80, 48, {[], [], [], [], [], [], [], [], [], [], [], [], [], [], [], []}, {{[], [], [], [['ns_1@10.1.4.243'| 0.0]], [['ns_1@10.1.4.244'| 0.0]], [], [], [], [], [], [], [], [], [], [], []}}}} INFO REPORT <0.113.0> 2011-05-07 17:39:25 =============================================================================== ns_1@10.1.4.243:ns_orchestrator:164: Skipping janitor in state rebalancing: {rebalancing_state, <0.375.0>, {dict, 2, 16, 16, 8, 80, 48, {[], [], [], [], [], [], [], [], [], [], [], [], [], [], [], []}, {{[], [], [], [['ns_1@10.1.4.243'| 0.0]], [['ns_1@10.1.4.244'| 0.0]], [], [], [], [], [], [], [], [], [], [], []}}}} INFO REPORT <0.113.0> 2011-05-07 17:39:35 =============================================================================== ns_1@10.1.4.243:ns_orchestrator:164: Skipping janitor in state rebalancing: {rebalancing_state, <0.375.0>, {dict, 2, 16, 16, 8, 80, 48, {[], [], [], [], [], [], [], [], [], [], [], [], [], [], [], []}, {{[], [], [], [['ns_1@10.1.4.243'| 0.0]], [['ns_1@10.1.4.244'| 0.0]], [], [], [], [], [], [], [], [], [], [], []}}}} INFO REPORT <0.86.0> 2011-05-07 17:39:40 =============================================================================== Pulling config from: 'ns_1@10.1.4.244' INFO REPORT <0.94.0> 2011-05-07 17:39:45 =============================================================================== ns_1@10.1.4.243:ns_doctor:82: Current node statuses: [{'ns_1@10.1.4.243', [{last_heard,{1304,815185,869497}}, {active_buckets,["default"]}, {memory, [{total,25581864}, {processes,14291364}, {processes_used,14278340}, {system,11290500}, {atom,536281}, {atom_used,522118}, {binary,223288}, {code,4299333}, {ets,1172108}]}, {cluster_compatibility_version,1}, {version, [{kernel,"2.13.4"}, {mnesia,"4.4.12"}, {stdlib,"1.16.4"}, {os_mon,"2.2.4"}, {inets,"5.2"}, {ns_server,"1.6.5.4r"}, {menelaus,"1.6.5.4r"}, {sasl,"2.1.8"}]}, {system_arch,"i686-pc-linux-gnu"}, {wall_clock,240}, {memory_data,{1050726400,788742144,{<0.171.0>,4114276}}}, {disk_data, [{"/",7583436,18}, {"/dev",508460,1}, {"/dev/shm",513048,0}, {"/var/run",513048,1}, {"/var/lock",513048,0}, {"/lib/init/rw",513048,0}, {"/boot",233191,8}]}, {meminfo, <<"MemTotal: 1026100 kB\nMemFree: 255944 kB\nBuffers: 152784 kB\nCached: 452868 kB\nSwapCached: 0 kB\nActive: 478608 kB\nInactive: 253236 kB\nActive(anon): 122032 kB\nInactive(anon): 4412 kB\nActive(file): 356576 kB\nInactive(file): 248824 kB\nUnevictable: 0 kB\nMlocked: 0 kB\nHighTotal: 137160 kB\nHighFree: 252 kB\nLowTotal: 888940 kB\nLowFree: 255692 kB\nSwapTotal: 397304 kB\nSwapFree: 397304 kB\nDirty: 96 kB\nWriteback: 0 kB\nAnonPages: 126256 kB\nMapped: 10488 kB\nShmem: 240 kB\nSlab: 28232 kB\nSReclaimable: 23308 kB\nSUnreclaim: 4924 kB\nKernelStack: 976 kB\nPageTables: 1108 kB\nNFS_Unstable: 0 kB\nBounce: 0 kB\nWritebackTmp: 0 kB\nCommitLimit: 910352 kB\nCommitted_AS: 351416 kB\nVmallocTotal: 122880 kB\nVmallocUsed: 3540 kB\nVmallocChunk: 114424 kB\nHardwareCorrupted: 0 kB\nHugePages_Total: 0\nHugePages_Free: 0\nHugePages_Rsvd: 0\nHugePages_Surp: 0\nHugepagesize: 2048 kB\nDirectMap4k: 10232 kB\nDirectMap2M: 901120 kB\n">>}, {replication,[{"default",1.0}]}, {system_memory_data, [{system_total_memory,1050726400}, {free_swap,406839296}, {total_swap,406839296}, {cached_memory,463736832}, {buffered_memory,156450816}, {free_memory,262086656}, {total_memory,1050726400}]}, {statistics, [{wall_clock,{230313,1}}, {context_switches,{232183,0}}, {garbage_collection,{53760,202180847,0}}, {io,{{input,74163788},{output,9520051}}}, {reductions,{49695088,856503}}, {run_queue,0}, {runtime,{6780,120}}]}]}, {'ns_1@10.1.4.244', [{last_heard,{1304,815184,930170}}, {active_buckets,["default"]}, {memory, [{total,14969536}, {processes,4878984}, {processes_used,4866424}, {system,10090552}, {atom,515245}, {atom_used,490161}, {binary,222936}, {code,3973911}, {ets,476132}]}, {cluster_compatibility_version,1}, {version, [{kernel,"2.13.4"}, {mnesia,"4.4.12"}, {stdlib,"1.16.4"}, {os_mon,"2.2.4"}, {ns_server,"1.6.5.4r"}, {menelaus,"1.6.5.4r"}, {sasl,"2.1.8"}]}, {system_arch,"i686-pc-linux-gnu"}, {wall_clock,291}, {memory_data,{1050726400,410419200,{<6517.255.0>,901104}}}, {disk_data, [{"/",7583436,16}, {"/dev",508460,1}, {"/dev/shm",513048,0}, {"/var/run",513048,1}, {"/var/lock",513048,0}, {"/lib/init/rw",513048,0}, {"/boot",233191,8}]}, {meminfo, <<"MemTotal: 1026100 kB\nMemFree: 624468 kB\nBuffers: 51392 kB\nCached: 254220 kB\nSwapCached: 0 kB\nActive: 166068 kB\nInactive: 195252 kB\nActive(anon): 49780 kB\nInactive(anon): 6152 kB\nActive(file): 116288 kB\nInactive(file): 189100 kB\nUnevictable: 0 kB\nMlocked: 0 kB\nHighTotal: 137160 kB\nHighFree: 280 kB\nLowTotal: 888940 kB\nLowFree: 624188 kB\nSwapTotal: 397304 kB\nSwapFree: 397304 kB\nDirty: 100 kB\nWriteback: 0 kB\nAnonPages: 55712 kB\nMapped: 10040 kB\nShmem: 232 kB\nSlab: 30528 kB\nSReclaimable: 25700 kB\nSUnreclaim: 4828 kB\nKernelStack: 928 kB\nPageTables: 860 kB\nNFS_Unstable: 0 kB\nBounce: 0 kB\nWritebackTmp: 0 kB\nCommitLimit: 910352 kB\nCommitted_AS: 256132 kB\nVmallocTotal: 122880 kB\nVmallocUsed: 3540 kB\nVmallocChunk: 114844 kB\nHardwareCorrupted: 0 kB\nHugePages_Total: 0\nHugePages_Free: 0\nHugePages_Rsvd: 0\nHugePages_Surp: 0\nHugepagesize: 2048 kB\nDirectMap4k: 10232 kB\nDirectMap2M: 901120 kB\n">>}, {replication,[{"default",1.0}]}, {system_memory_data, [{system_total_memory,1050726400}, {free_swap,406839296}, {total_swap,406839296}, {cached_memory,260321280}, {buffered_memory,52625408}, {free_memory,639455232}, {total_memory,1050726400}]}, {statistics, [{wall_clock,{285008,1}}, {context_switches,{38832,0}}, {garbage_collection,{8971,24991237,0}}, {io,{{input,5789194},{output,2144943}}}, {reductions,{8992598,251574}}, {run_queue,0}, {runtime,{1800,40}}]}]}] INFO REPORT <0.113.0> 2011-05-07 17:39:45 =============================================================================== ns_1@10.1.4.243:ns_orchestrator:164: Skipping janitor in state rebalancing: {rebalancing_state, <0.375.0>, {dict, 2, 16, 16, 8, 80, 48, {[], [], [], [], [], [], [], [], [], [], [], [], [], [], [], []}, {{[], [], [], [['ns_1@10.1.4.243'| 0.0]], [['ns_1@10.1.4.244'| 0.0]], [], [], [], [], [], [], [], [], [], [], []}}}} INFO REPORT <0.113.0> 2011-05-07 17:39:55 =============================================================================== ns_1@10.1.4.243:ns_orchestrator:164: Skipping janitor in state rebalancing: {rebalancing_state, <0.375.0>, {dict, 2, 16, 16, 8, 80, 48, {[], [], [], [], [], [], [], [], [], [], [], [], [], [], [], []}, {{[], [], [], [['ns_1@10.1.4.243'| 0.0]], [['ns_1@10.1.4.244'| 0.0]], [], [], [], [], [], [], [], [], [], [], []}}}} INFO REPORT <0.86.0> 2011-05-07 17:39:56 =============================================================================== Pulling config from: 'ns_1@10.1.4.244' INFO REPORT <0.113.0> 2011-05-07 17:40:05 =============================================================================== ns_1@10.1.4.243:ns_orchestrator:164: Skipping janitor in state rebalancing: {rebalancing_state, <0.375.0>, {dict, 2, 16, 16, 8, 80, 48, {[], [], [], [], [], [], [], [], [], [], [], [], [], [], [], []}, {{[], [], [], [['ns_1@10.1.4.243'| 0.0]], [['ns_1@10.1.4.244'| 0.0]], [], [], [], [], [], [], [], [], [], [], []}}}} INFO REPORT <0.86.0> 2011-05-07 17:40:10 =============================================================================== Pulling config from: 'ns_1@10.1.4.244' INFO REPORT <0.113.0> 2011-05-07 17:40:15 =============================================================================== ns_1@10.1.4.243:ns_orchestrator:164: Skipping janitor in state rebalancing: {rebalancing_state, <0.375.0>, {dict, 2, 16, 16, 8, 80, 48, {[], [], [], [], [], [], [], [], [], [], [], [], [], [], [], []}, {{[], [], [], [['ns_1@10.1.4.243'| 0.0]], [['ns_1@10.1.4.244'| 0.0]], [], [], [], [], [], [], [], [], [], [], []}}}} INFO REPORT <0.113.0> 2011-05-07 17:40:25 =============================================================================== ns_1@10.1.4.243:ns_orchestrator:164: Skipping janitor in state rebalancing: {rebalancing_state, <0.375.0>, {dict, 2, 16, 16, 8, 80, 48, {[], [], [], [], [], [], [], [], [], [], [], [], [], [], [], []}, {{[], [], [], [['ns_1@10.1.4.243'| 0.0]], [['ns_1@10.1.4.244'| 0.0]], [], [], [], [], [], [], [], [], [], [], []}}}} INFO REPORT <0.86.0> 2011-05-07 17:40:27 =============================================================================== Pulling config from: 'ns_1@10.1.4.244' INFO REPORT <0.113.0> 2011-05-07 17:40:35 =============================================================================== ns_1@10.1.4.243:ns_orchestrator:164: Skipping janitor in state rebalancing: {rebalancing_state, <0.375.0>, {dict, 2, 16, 16, 8, 80, 48, {[], [], [], [], [], [], [], [], [], [], [], [], [], [], [], []}, {{[], [], [], [['ns_1@10.1.4.243'| 0.0]], [['ns_1@10.1.4.244'| 0.0]], [], [], [], [], [], [], [], [], [], [], []}}}} INFO REPORT <0.94.0> 2011-05-07 17:40:45 =============================================================================== ns_1@10.1.4.243:ns_doctor:82: Current node statuses: [{'ns_1@10.1.4.243', [{last_heard,{1304,815245,869458}}, {active_buckets,["default"]}, {memory, [{total,25722808}, {processes,14350916}, {processes_used,14337428}, {system,11371892}, {atom,536281}, {atom_used,522118}, {binary,271704}, {code,4299333}, {ets,1205196}]}, {cluster_compatibility_version,1}, {version, [{kernel,"2.13.4"}, {mnesia,"4.4.12"}, {stdlib,"1.16.4"}, {os_mon,"2.2.4"}, {inets,"5.2"}, {ns_server,"1.6.5.4r"}, {menelaus,"1.6.5.4r"}, {sasl,"2.1.8"}]}, {system_arch,"i686-pc-linux-gnu"}, {wall_clock,300}, {memory_data,{1050726400,789147648,{<0.171.0>,4114276}}}, {disk_data, [{"/",7583436,18}, {"/dev",508460,1}, {"/dev/shm",513048,0}, {"/var/run",513048,1}, {"/var/lock",513048,0}, {"/lib/init/rw",513048,0}, {"/boot",233191,8}]}, {meminfo, <<"MemTotal: 1026100 kB\nMemFree: 255324 kB\nBuffers: 152784 kB\nCached: 453120 kB\nSwapCached: 0 kB\nActive: 478824 kB\nInactive: 253368 kB\nActive(anon): 122132 kB\nInactive(anon): 4412 kB\nActive(file): 356692 kB\nInactive(file): 248956 kB\nUnevictable: 0 kB\nMlocked: 0 kB\nHighTotal: 137160 kB\nHighFree: 252 kB\nLowTotal: 888940 kB\nLowFree: 255072 kB\nSwapTotal: 397304 kB\nSwapFree: 397304 kB\nDirty: 92 kB\nWriteback: 0 kB\nAnonPages: 126256 kB\nMapped: 10488 kB\nShmem: 240 kB\nSlab: 28228 kB\nSReclaimable: 23308 kB\nSUnreclaim: 4920 kB\nKernelStack: 976 kB\nPageTables: 1120 kB\nNFS_Unstable: 0 kB\nBounce: 0 kB\nWritebackTmp: 0 kB\nCommitLimit: 910352 kB\nCommitted_AS: 352656 kB\nVmallocTotal: 122880 kB\nVmallocUsed: 3540 kB\nVmallocChunk: 114424 kB\nHardwareCorrupted: 0 kB\nHugePages_Total: 0\nHugePages_Free: 0\nHugePages_Rsvd: 0\nHugePages_Surp: 0\nHugepagesize: 2048 kB\nDirectMap4k: 10232 kB\nDirectMap2M: 901120 kB\n">>}, {replication,[{"default",1.0}]}, {system_memory_data, [{system_total_memory,1050726400}, {free_swap,406839296}, {total_swap,406839296}, {cached_memory,463994880}, {buffered_memory,156450816}, {free_memory,261451776}, {total_memory,1050726400}]}, {statistics, [{wall_clock,{290313,1}}, {context_switches,{248150,0}}, {garbage_collection,{57493,238196141,0}}, {io,{{input,74550411},{output,10059223}}}, {reductions,{54631253,846933}}, {run_queue,0}, {runtime,{7690,140}}]}]}, {'ns_1@10.1.4.244', [{last_heard,{1304,815244,930256}}, {active_buckets,["default"]}, {memory, [{total,15110496}, {processes,4996080}, {processes_used,4983520}, {system,10114416}, {atom,515245}, {atom_used,490161}, {binary,212176}, {code,3973911}, {ets,510564}]}, {cluster_compatibility_version,1}, {version, [{kernel,"2.13.4"}, {mnesia,"4.4.12"}, {stdlib,"1.16.4"}, {os_mon,"2.2.4"}, {ns_server,"1.6.5.4r"}, {menelaus,"1.6.5.4r"}, {sasl,"2.1.8"}]}, {system_arch,"i686-pc-linux-gnu"}, {wall_clock,351}, {memory_data,{1050726400,411299840,{<6517.255.0>,901104}}}, {disk_data, [{"/",7583436,16}, {"/dev",508460,1}, {"/dev/shm",513048,0}, {"/var/run",513048,1}, {"/var/lock",513048,0}, {"/lib/init/rw",513048,0}, {"/boot",233191,8}]}, {meminfo, <<"MemTotal: 1026100 kB\nMemFree: 623440 kB\nBuffers: 51484 kB\nCached: 254620 kB\nSwapCached: 0 kB\nActive: 166816 kB\nInactive: 195404 kB\nActive(anon): 50216 kB\nInactive(anon): 6152 kB\nActive(file): 116600 kB\nInactive(file): 189252 kB\nUnevictable: 0 kB\nMlocked: 0 kB\nHighTotal: 137160 kB\nHighFree: 252 kB\nLowTotal: 888940 kB\nLowFree: 623188 kB\nSwapTotal: 397304 kB\nSwapFree: 397304 kB\nDirty: 84 kB\nWriteback: 0 kB\nAnonPages: 56180 kB\nMapped: 10108 kB\nShmem: 232 kB\nSlab: 30540 kB\nSReclaimable: 25708 kB\nSUnreclaim: 4832 kB\nKernelStack: 928 kB\nPageTables: 860 kB\nNFS_Unstable: 0 kB\nBounce: 0 kB\nWritebackTmp: 0 kB\nCommitLimit: 910352 kB\nCommitted_AS: 256132 kB\nVmallocTotal: 122880 kB\nVmallocUsed: 3540 kB\nVmallocChunk: 114844 kB\nHardwareCorrupted: 0 kB\nHugePages_Total: 0\nHugePages_Free: 0\nHugePages_Rsvd: 0\nHugePages_Surp: 0\nHugepagesize: 2048 kB\nDirectMap4k: 10232 kB\nDirectMap2M: 901120 kB\n">>}, {replication,[{"default",1.0}]}, {system_memory_data, [{system_total_memory,1050726400}, {free_swap,406839296}, {total_swap,406839296}, {cached_memory,260730880}, {buffered_memory,52719616}, {free_memory,638402560}, {total_memory,1050726400}]}, {statistics, [{wall_clock,{345007,0}}, {context_switches,{44766,0}}, {garbage_collection,{10361,29683202,0}}, {io,{{input,6029484},{output,2409944}}}, {reductions,{10484708,249757}}, {run_queue,0}, {runtime,{2080,50}}]}]}] INFO REPORT <0.113.0> 2011-05-07 17:40:45 =============================================================================== ns_1@10.1.4.243:ns_orchestrator:164: Skipping janitor in state rebalancing: {rebalancing_state, <0.375.0>, {dict, 2, 16, 16, 8, 80, 48, {[], [], [], [], [], [], [], [], [], [], [], [], [], [], [], []}, {{[], [], [], [['ns_1@10.1.4.243'| 0.0]], [['ns_1@10.1.4.244'| 0.0]], [], [], [], [], [], [], [], [], [], [], []}}}} INFO REPORT <0.167.0> 2011-05-07 17:40:49 =============================================================================== ns_1@10.1.4.243:stats_collector:71: Stats for bucket "default": auth_cmds 6 auth_errors 0 bucket_conns 7 bytes_read 43458 bytes_written 2478400 cas_badval 0 cas_hits 0 cas_misses 0 cmd_flush 0 cmd_get 0 cmd_set 0 conn_yields 0 connection_structures 17 curr_connections 17 curr_items 204600 curr_items_tot 204800 daemon_connections 10 decr_hits 0 decr_misses 0 delete_hits 0 delete_misses 0 ep_bg_fetched 0 ep_commit_num 0 ep_commit_time 0 ep_commit_time_total 0 ep_data_age 0 ep_data_age_highwat 0 ep_db_cleaner_status complete ep_db_strategy multiMTVBDB ep_dbinit 1 ep_dbname /var/opt/membase/1.6.5.4r/data/ns_1/default ep_dbshards 4 ep_expired 0 ep_flush_duration 0 ep_flush_duration_highwat 0 ep_flush_duration_total 0 ep_flush_preempts 0 ep_flusher_state running ep_flusher_todo 0 ep_io_num_read 205000 ep_io_num_write 0 ep_io_read_bytes 23370000 ep_io_write_bytes 0 ep_item_begin_failed 0 ep_item_commit_failed 0 ep_item_flush_expired 0 ep_item_flush_failed 0 ep_kv_size 35840000 ep_max_data_size 839909376 ep_max_txn_size 1000 ep_mem_high_wat 629932032 ep_mem_low_wat 503945625 ep_min_data_age 0 ep_num_active_non_resident 0 ep_num_eject_failures 0 ep_num_eject_replicas 0 ep_num_expiry_pager_runs 0 ep_num_non_resident 0 ep_num_not_my_vbuckets 0 ep_num_pager_runs 0 ep_num_value_ejects 0 ep_oom_errors 0 ep_overhead 12960032 ep_pending_ops 0 ep_pending_ops_max 0 ep_pending_ops_max_duration 0 ep_pending_ops_total 0 ep_queue_age_cap 900 ep_queue_size 0 ep_storage_age 0 ep_storage_age_highwat 0 ep_storage_type featured ep_store_max_concurrency 10 ep_store_max_readers 9 ep_store_max_readwrite 1 ep_tap_bg_fetch_requeued 0 ep_tap_bg_fetched 200 ep_tap_keepalive 300 ep_tmp_oom_errors 0 ep_too_old 0 ep_too_young 0 ep_total_cache_size 35840000 ep_total_del_items 0 ep_total_enqueued 0 ep_total_new_items 0 ep_total_persisted 0 ep_vbucket_del 0 ep_vbucket_del_fail 0 ep_version 1.6.5.3.1_8_gce31baa ep_warmed_up 204800 ep_warmup true ep_warmup_dups 0 ep_warmup_oom 0 ep_warmup_thread complete ep_warmup_time 809268 get_hits 0 get_misses 0 incr_hits 0 incr_misses 0 libevent 2.0.7-rc limit_maxbytes 67108864 mem_used 48800032 pid 8362 pointer_size 32 rejected_conns 0 rusage_system 0.300018 rusage_user 1.208075 tap_connect_received 1 tap_mutation_sent 400 tap_opaque_sent 1 tap_vbucket_set_sent 2 threads 4 time 1304815248 total_connections 17 uptime 305 version 1.4.4_382_g9df3289 INFO REPORT <0.86.0> 2011-05-07 17:40:55 =============================================================================== Pulling config from: 'ns_1@10.1.4.244' INFO REPORT <0.113.0> 2011-05-07 17:40:55 =============================================================================== ns_1@10.1.4.243:ns_orchestrator:164: Skipping janitor in state rebalancing: {rebalancing_state, <0.375.0>, {dict, 2, 16, 16, 8, 80, 48, {[], [], [], [], [], [], [], [], [], [], [], [], [], [], [], []}, {{[], [], [], [['ns_1@10.1.4.243'| 0.0]], [['ns_1@10.1.4.244'| 0.0]], [], [], [], [], [], [], [], [], [], [], []}}}} INFO REPORT <0.113.0> 2011-05-07 17:41:05 =============================================================================== ns_1@10.1.4.243:ns_orchestrator:164: Skipping janitor in state rebalancing: {rebalancing_state, <0.375.0>, {dict, 2, 16, 16, 8, 80, 48, {[], [], [], [], [], [], [], [], [], [], [], [], [], [], [], []}, {{[], [], [], [['ns_1@10.1.4.243'| 0.0]], [['ns_1@10.1.4.244'| 0.0]], [], [], [], [], [], [], [], [], [], [], []}}}}