-
Notifications
You must be signed in to change notification settings - Fork 9
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
rolling restart #124
rolling restart #124
Conversation
# | ||
_LOG.info('Overriding ephemeral volumes to be able to set up AWS auto recovery alarm ') | ||
block_devices = [] | ||
for bd in ami.block_device_mappings: |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
@adyach could you please explain this to me later. I don't understand it.
UserData=taupage_user_data, | ||
InstanceType=self.cluster_config.get_instance_type(), | ||
SubnetId=subnet['SubnetId'], | ||
PrivateIpAddress=ip, |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
@adyach why not let AWS choose an available IP address for you?
logging.basicConfig(level=getattr(logging, 'INFO', None)) | ||
|
||
|
||
@click.group() | ||
def cli(): | ||
logo = """ |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
.__
____ ____ ____ | |
_/ ___\/ _ \ / _ \| |
\ \__( <_> | <_> ) |__
\___ >____/ \____/|____/
\/
This is really cool 👍 Some corner cases that could happen but I'm not sure we need to address them right now:
|
Thank you for review!
1. It is intentionally blocking restart if there is prefered replica
imbalance, we observed in the past version of Kafka, it could imfluence
behavior. I agree, we need to observe it for sometime.
2. The restart can stop if bubuku performing restart was stopped, in that
case the restart should be started again from scratch. This is
intentionally done like that, because it is quite bad situation, when
broker is down while it is restarting other one. The state is not saved,
this is not perfect, we can improve it if required. If one of the steps
fails with exception, it will be triggered again.
On Thu 20. Dec 2018 at 12:50, Ricardo de Cillo ***@***.***> wrote:
This is really cool 👍
Some corner cases that could happen but I'm not sure we need to address
them right now:
1. Kafka leader imbalance could make this pause permanently between
restarts. We've seen that happening. Let's see how it goes.
2. The action is consumed and posted back on each step. The problem
with that is that some problems during the execution of a step could cause
an unrecoverable failure. I mean, if at step 3 out of 5 something bad
happens and an unrecoverable exception is launched, wouldn't that cause the
complete process to stop and we would have to somehow identify where it
stopped? Please, correct me if I understood it wrong.
—
You are receiving this because you were mentioned.
Reply to this email directly, view it on GitHub
<#124 (comment)>,
or mute the thread
<https://github.com/notifications/unsubscribe-auth/AEo3QmS5BsvbPbJDZ8S-VBU1fcGkOjGQks5u63mVgaJpZM4ZbtFn>
.
--
With great enthusiasm,
Andrey
|
Ok, now I remember that bubuku has a "catch all" exception and it would retry the action, right? So the only chance of having a partial rolling restart would be to kill bubuku process with -9, for example. If that's the case, then I agree that it's quite dramatic event and maybe we would like to start all over. |
👍 |
I checked the lastest change. Was it necessary because it would retry too often? |
👍 |
1 similar comment
👍 |
Description
Rolling restart of the cluster:
bubuku-cli rolling-restart --image-tag 123 --instnce-type t2.nano