带zookeeper群集的mesos主崩溃

gmxoilav  于 2021-06-21  发布在  Mesos
关注(0)|答案(4)|浏览(389)

我正在部署一个zookeeper集群,它有3个节点。我用它来保持我的mesos主机的高可用性。我下载zookeeper-3.4.6.tar.gz tarball并将其解压缩到/opt,将其重命名为/opt/zookeeper,进入目录,编辑conf/zoo.cfg(粘贴在下面),在datadir中创建myid文件(设置为/var/lib/zookeeper in zoo.cfg),然后使用./bin/zkserver.sh start启动zookeeper,一切正常。我一个接一个地启动所有3个节点,它们看起来都很好。我用./bin/zkcli.sh连接服务器,没问题。
但是当我启动mesos(3个主节点和3个从节点,每个节点运行一个主节点和一个从节点)时,主节点很快就崩溃了,一个接一个,出现在网页上http://mesos_master:5050,从属选项卡,不显示从属。但是当我只管理一个Zookeeper的时候,这些都很好。所以我认为这是Zookeeper的问题。
我的ubuntu服务器上有3台pv主机。他们都在运行ubuntu14.04 lts:node-01,node-02,node-03,我有 /etc/hosts 在这三个节点中:

172.16.2.70     node-01
172.16.2.81     node-02
172.16.2.80     node-03

我在所有三个节点上安装了zookeeper和mesos。zookeeper配置文件如下(所有三个节点):

tickTime=2000
dataDir=/var/lib/zookeeper
clientPort=2181
initLimit=5
syncLimit=2
server.1=node-01:2888:3888
server.2=node-02:2888:3888
server.3=node-03:2888:3888

启动正常,运行良好。然后我使用命令行启动mesos主服务 ./bin/mesos-master.sh --zk=zk://172.16.2.70:2181,172.16.2.81:2181,172.16.2.80:2181/mesos --work_dir=/var/lib/mesos --quorum=2 ,几秒钟后,它会给我这样的错误:

F0817 15:09:19.995256  2250 master.cpp:1253] Recovery failed: Failed to recover registrar: Failed to perform fetch within 1mins

***Check failure stack trace:***

    @     0x7fa2b8be71a2  google::LogMessage::Fail()
    @     0x7fa2b8be70ee  google::LogMessage::SendToLog()
    @     0x7fa2b8be6af0  google::LogMessage::Flush()
    @     0x7fa2b8be9a04  google::LogMessageFatal::~LogMessageFatal()

▽
    @     0x7fa2b81a899a  mesos::internal::master::fail()

▽
    @     0x7fa2b8262f8f  _ZNSt5_BindIFPFvRKSsS1_EPKcSt12_PlaceholderILi1EEEE6__callIvJS1_EJLm0ELm1EEEET_OSt5tupleIJDpT0_EESt12_Index_tupleIJXspT1_EEE

▽
    @     0x7fa2b823fba7  _ZNSt5_BindIFPFvRKSsS1_EPKcSt12_PlaceholderILi1EEEEclIJS1_EvEET0_DpOT_
    @     0x7fa2b820f9f3  _ZZNK7process6FutureI7NothingE8onFailedISt5_BindIFPFvRKSsS6_EPKcSt12_PlaceholderILi1EEEEvEERKS2_OT_NS2_6PreferEENUlS6_E_clES6_
    @     0x7fa2b826305c  _ZNSt17_Function_handlerIFvRKSsEZNK7process6FutureI7NothingE8onFailedISt5_BindIFPFvS1_S1_EPKcSt12_PlaceholderILi1EEEEvEERKS6_OT_NS6_6PreferEEUlS1_E_E9_M_invokeERKSt9_Any_dataS1_
    @           0x4a44e7  std::function<>::operator()()
    @           0x49f3a7  _ZN7process8internal3runISt8functionIFvRKSsEEJS4_EEEvRKSt6vectorIT_SaIS8_EEDpOT0_
    @           0x499480  process::Future<>::fail()
    @     0x7fa2b806b4b4  process::Promise<>::fail()
    @     0x7fa2b826011b  process::internal::thenf<>()
    @     0x7fa2b82a0757  _ZNSt5_BindIFPFvRKSt8functionIFN7process6FutureI7NothingEERKN5mesos8internal8RegistryEEERKSt10shared_ptrINS1_7PromiseIS3_EEERKNS2_IS7_EEESB_SH_St12_PlaceholderILi1EEEE6__callIvISM_EILm0ELm1ELm2EEEET_OSt5tupleIIDpT0_EESt12_Index_tupleIIXspT1_EEE
    @     0x7fa2b82962d9  std::_Bind<>::operator()<>()
    @     0x7fa2b827ee89  std::_Function_handler<>::_M_invoke()
I0817 15:09:20.098639  2248 http.cpp:283] HTTP GET for /master/state.json from 172.16.2.84:54542 with User-Agent='Mozilla/5.0 (Macintosh; Intel Mac OS X 10_10_4) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/44.0.2403.155 Safari/537.36'
    @     0x7fa2b8296507  std::function<>::operator()()
    @     0x7fa2b827efaf  _ZZNK7process6FutureIN5mesos8internal8RegistryEE5onAnyIRSt8functionIFvRKS4_EEvEES8_OT_NS4_6PreferEENUlS8_E_clES8_
    @     0x7fa2b82a07fe  _ZNSt17_Function_handlerIFvRKN7process6FutureIN5mesos8internal8RegistryEEEEZNKS5_5onAnyIRSt8functionIS8_EvEES7_OT_NS5_6PreferEEUlS7_E_E9_M_invokeERKSt9_Any_dataS7_
    @     0x7fa2b8296507  std::function<>::operator()()
    @     0x7fa2b82e4419  process::internal::run<>()
    @     0x7fa2b82da22a  process::Future<>::fail()
    @     0x7fa2b83136b5  std::_Mem_fn<>::operator()<>()
    @     0x7fa2b830efdf  _ZNSt5_BindIFSt7_Mem_fnIMN7process6FutureIN5mesos8internal8RegistryEEEFbRKSsEES6_St12_PlaceholderILi1EEEE6__callIbIS8_EILm0ELm1EEEET_OSt5tupleIIDpT0_EESt12_Index_tupleIIXspT1_EEE
    @     0x7fa2b8307d7f  _ZNSt5_BindIFSt7_Mem_fnIMN7process6FutureIN5mesos8internal8RegistryEEEFbRKSsEES6_St12_PlaceholderILi1EEEEclIJS8_EbEET0_DpOT_
    @     0x7fa2b82fe431  _ZZNK7process6FutureIN5mesos8internal8RegistryEE8onFailedISt5_BindIFSt7_Mem_fnIMS4_FbRKSsEES4_St12_PlaceholderILi1EEEEbEERKS4_OT_NS4_6PreferEENUlS9_E_clES9_
    @     0x7fa2b830f065  _ZNSt17_Function_handlerIFvRKSsEZNK7process6FutureIN5mesos8internal8RegistryEE8onFailedISt5_BindIFSt7_Mem_fnIMS8_FbS1_EES8_St12_PlaceholderILi1EEEEbEERKS8_OT_NS8_6PreferEEUlS1_E_E9_M_invokeERKSt9_Any_dataS1_
    @           0x4a44e7  std::function<>::operator()()
    @           0x49f3a7  _ZN7process8internal3runISt8functionIFvRKSsEEJS4_EEEvRKSt6vectorIT_SaIS8_EEDpOT0_
    @     0x7fa2b82da202  process::Future<>::fail()
    @     0x7fa2b82d2d82  process::Promise<>::fail()
Aborted

有时警告是这样的,然后以上面相同的输出崩溃:

0817 15:09:49.745750  2104 recover.cpp:111] Unable to finish the recover protocol in 10secs, retrying

我想知道zookeeper在我的案例中是否部署和运行良好,以及如何定位问题所在。欢迎您的回答和建议。谢谢。

nqwrtyyt

nqwrtyyt1#

将你的mesos大师的作品目录分成不同的目录,不要对所有的大师使用共享的作品目录,因为zk

pdkcd3nj

pdkcd3nj2#

我陷入了同样的问题,我尝试了不同的方法和不同的选择,最后 --ip 这个选择对我有效。最初我用 --hostname 选项

mesos-master --ip=192.168.0.13 --quorum=2 --zk=zk://m1:2181,m2:2181,m3:2181/mesos --work_dir=/opt/mm1 --log_dir=/opt/mm1/logs
5f0d552i

5f0d552i3#

实际上,在我的例子中,这是因为我没有打开防火墙端口5050,以允许三台服务器相互通信。更新防火墙规则后,它将按预期开始工作。

tez616oj

tez616oj4#

您需要检查所有mesos/zookeeper主节点是否可以正确通信。为此,您需要:
zookeeper端口打开:tcp 2181、2888、3888
mesos端口打开:tcp 5050
ping可用(icmp消息0和8)
如果在配置中使用fqdn而不是ip,请检查dns解析是否也正常工作。

相关问题