-
Notifications
You must be signed in to change notification settings - Fork 23
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Self-hosted workers fail immediately, get marked "offline" in the runners list. #388
Comments
👋 @mikolajpabiszczak the reason is because the runner has been marked to do just one job with the parameter |
@DavidGOrtega: I do know that. So let me emphasise this again:
(I added some clarifications in the opening message) |
@mikolajpabiszczak
There must be something that you do not have permissions to do with your token? |
Just to be sure and move one step forward can you please your |
These were not changed since the working runs, but I checked it again. We are using a company application, so checking up wrt. this list Repository level:
Organisation level:
Additionally, in the repository settings:
|
It looks like that app needs an additional scope it might not have? https://docs.github.com/en/rest/actions/workflow-runs#list-workflow-runs-for-a-repository @mikolajpabiszczak to confirm is an issue with app generated token can you try and curl the endpoint with one of the generated tokens?
|
Did some tests, indeed the culprit was the lack of sufficient permissions: after adding Thx for your time and help! And yes, the guide needs an update in this case. ;D |
@mikolajpabiszczak thanks for the report and help, we'll keep this open until we update the docs |
I am frankly not sure if this is the issue on CML side, but let me describe it.
CML versions tested: 0.11.0 and 0.17.0
Cloud provider: AWS
Remark: the very same workflow worked when I last used it (3 months ago)
Deploy self-hosted runner:
this deployment job finishes successfully, but when it finishes the instance (as checked in AWS console) has not yet performed status checks (this was not the case when the workflow worked 3 last time) / is still in the
Initialisation
stage.The next job (which runs on self-hosted runner) gets closed basically immediately (in 4s):
The runner has received a shutdown signal
although the instance itself is not getting cancelled: it goes through AWS status checks and remains running (to clarify: instance deployed as
single
),One more thing: if I deploy the worker as
reusable
it will be marked as offline in the list of workers after the job fails and will not be accessible…I deployed the reusable instance and got logs after failure:
The text was updated successfully, but these errors were encountered: