Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

azure-events-az: Node Health Attribute (-1000000) Not Automatically Reset After Critical Failure in Resource-Agent Update #2025

Open
rnirek opened this issue Feb 20, 2025 · 3 comments

Comments

@rnirek
Copy link

rnirek commented Feb 20, 2025

We have observed that with the recent update to the resource-agent configuration, as outlined in Cluster Lab resource-agents, once a node is marked with a value of -1000000—indicating a critical failure or an unhealthy state—the attribute remains unchanged until it is manually modified. Consequently, cluster services will not restart until this manual intervention occurs.

Could you clarify the rationale behind why the value is not automatically reset to 0?

@oalbrigt oalbrigt changed the title Node Health Attribute (-1000000) Not Automatically Reset After Critical Failure in Resource-Agent Update azure-events-az: Node Health Attribute (-1000000) Not Automatically Reset After Critical Failure in Resource-Agent Update Feb 24, 2025
@oalbrigt
Copy link
Contributor

@MSSedusch Can you explain why the value was changed from what it was in the old azure-events agent?

@MSSedusch
Copy link
Contributor

@HappyTobi @msftrobiro @rdeltcheva

@oalbrigt @rnirek the node should be put back online here:

def putNodeOnline(self, node=None):

@HappyTobi can you debug why/if the node is not put back online? Do we have to also put it in the else branch here?

@rnirek
Copy link
Author

rnirek commented Feb 25, 2025

@MSSedusch While that was our assumption, we have observed that whenever a node is shut down, it is automatically marked as inactive with a value of -100000, preventing it from being brought back online until the value is manually updated. This behavior has been consistently observed across several clusters.

clusterHelper._exec("crm_attribute", "--node", node, "--name", attr_healthstate, "--update", "-1000000", "--lifetime=forever")

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants