-
Notifications
You must be signed in to change notification settings - Fork 75
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
stage.0: Module function cephprocesses.wait threw an exception. Exception: 'openattic' #657
Comments
Did your cluster go into HEALTH_ERR? The steps in Stage 0 for a minion are serialized for an already running cluster. It's in /srv/salt/ceph/stage/0/minion/default.sls. The ceph.wait state is simply paranoia on our part that the previous update on some minion caused an issue and the cluster did not recover. We bail out. Unfortunately, the HEALTH_ERR status isn't terribly granular so we do not have a systematic guarantee of correlating cause and effect. |
No - the cluster and all services are up and running well and I am testing to go through stages 0..5 where I do not expect failures in case I did not change anything. Any idea how to get more debug information, here? |
There is a module called cephprocesses.py which checks for services to be up. You can try it either on the respective node with:
or target the node directly
or condensed in a runner that checks all services for all roles on all nodes.
Appending a |
result:
on the node itself (this is on the admin node)
Output of salt-call cephprocesses.check attached: |
this confirms that #661 will fix your issue |
Manually applied the change in #661 and executed salt "*" saltutil.sync_all. --> After this the error is gone! THANKS! |
I'm having same issue with prometheus: salt cl5.opn.shft cephprocesses.checkcl5.opn.shft: |
Which deepsea version are you using @khodayard ? |
Versions Report |
@khodayard There is no |
@jschmid1 thank you for your response. this is my policy.cfg now: :~ # cat /srv/pillar/ceph/proposals/policy.cfg but I'm getting the same result: Failures summary: ceph.metapackage (/srv/salt/ceph/metapackage): I've even tried to upgrade deepsea to the latest version from github but it failed and I had to revert a snapshot: :~ # deepsea stage run ceph.stage.0 Thanks again. |
Make sure to run stage.2 after changing the policy.cfg |
Running stage.2 fixed that problem, thank you. |
On an already deployed cluster I get this error when executing stage.0:
The openattic.service us up an running well and I can access openATTIC with a web-browser without any problems.
Any idea what might be wrong, here?
The text was updated successfully, but these errors were encountered: