-
Notifications
You must be signed in to change notification settings - Fork 20
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
zookeeper needs to be restarted after upgrading instance type #80
Comments
This looks like a rolling restart steps for zk. Would be an issue of zk itself. The full logs would help to further analyse the issue. |
The cluster has been running for several months w/o issues and I probably don't have the logs of point of start anymore |
Not sure the root cause. The node looks not able to connect to zk on 172.31.4.202. 2018-10-11 10:10:29,214 [myid:2] - WARN [WorkerSender[myid=2]:QuorumCnxManager@584] - Cannot open channel to 3 at election address zoo-prod-2.firecamp-prod-firecamp.com/172.31.4.202:3888 |
Yeah, I'm not sure as well. The main point is that it's got fixed after restarting zookeeper. Probably we need to wait for all zookeeper nodes to start up before starting up kafka? |
Are you able to get logs from node 172.31.4.202? Kafka does rely on ZK. If zk cluster is not working, kafka will not work. We could consider to introduce the dependency between services. While, how to detecting one service is healthy might not be easy. It has to look into the service internal status. For kafka, it is not necessary to do so. Kafka itself will wait till zk is running. |
What kind of logs do you need? |
zk logs, to see if it showed some information about why connecting fails. |
@JuniusLuo, please check #80 (comment) for zookeeper logs |
there is only one log file. We need the log file for zk on 172.31.4.202 |
It looks like 4.202 is zoo2. zoo2 was not started at 10:10:29. The first log in zoo2.log.gz was at 10:11:14. Probably when system was not stable, zk instance kept restarting itself. Restarting the zk service helps to bring all instances up around the same time. This looks like a zk issue itself. Probably zk instance should just wait and retry. |
Could you please make firecamp-manager aware of this issue, so Kafka will be restarted only after all ZK instances are up and running? |
The manager service aims to be a common service. Monitoring the specific service healthy status is too specific to the service. Looks not the best fit to have the manager service to do this. |
@JuniusLuo , in general, I agree with you. But this particular service - Kafka - does not work at all without Zookeeper. So, they're tied by design. Thinking that would be a good feature to handle this case in the manager. |
This is not an easy task. This is like requiring the full monitoring ability to ZooKeeper. For example, if ZooKeeper fails because of some bug/issue, Kafka will not work as well. |
We just need to start Kafka containers after Zookeeper is got up. Do you really think it's hard to implement? |
Currently we don't have the control for which service starts first. Firecamp is simple. The manage service is simply responsible for initializing the service and updating the service configs, such as creating volumes, etc. Then it is ECS's responsible to schedule the containers, and the manage service does not involve on the scheduling. Filecamp plugin will talk with DynamoDB to grab the volume and update network. |
Understand. What do you think of adding this issue somewhere in the wiki? So people might be aware of such things |
How to reproduce:
Zookeeper logs before step 4 contain a lot of similar messages:
The text was updated successfully, but these errors were encountered: