Updating existing machines in '****' with rolling strategy
[1/4] Waiting for 683dde2a77e618 [worker] to have state: started
Error: timeout reached waiting for machine to started failed to wait for VM 683dde2a77e618 in started state: Get "https://api.machines.dev/v1/apps/****/machines/683dde2a77e618/wait?instance_id=01H7BBZKNQKS8NY24X0S177FVT&state=started&timeout=60": net/http: request canceled
note: you can change this timeout with the --wait-timeout flag
London:
Error: failed to fetch an image or build from source: error rendering push status stream: received unexpected HTTP status: 500 Internal Server Error
I’m brand new. I think I’m experiencing the same issue. I was able to redeploy by destroying all machines. But the issue returns each time I try to deploy. It’s always the sidekiq one. With the message “Waiting for XXXX [sidekiq] to have state: started”
I’ll see if I can figure out how to restart that one machine as someone suggested above.
Still happening for me this morning. Starting or restarting the “stuck” app doesn’t help (I have 8 machines across 4 processes, I’ve tried doing that with all of them.)
I’m going to sign up to the Starter plan and try their email support and see if the support is better there - this is my first month running something live with Fly, it’ll be a real shame to have to move elsewhere, but not being able to deploy for 24 hours is kinda a blocker …
Hi all, we identified a bug in fly deploy introduced a couple days ago (starting in v0.1.72) causing deploys to incorrectly wait on an app’s standby machines to enter the started state (which will never happen, resulting in a timeout error), when they should properly be skipped.
We are working on publishing a new release with a fix (which will be v0.1.74). A new version of flyctl (v0.1.74) has been released with a fix for this issue.
In the meantime, you can manually downgrade to v0.1.71 (after disabling autoupdate first) to work around this issue: