I am getting the same error with our NextJS application for FlyIO Apps V2.
Our FLyIO Application has been deploying just fine for some time now, and only broke once we introduced the release_command
into our fly.toml
:
app = 'arch1ve-web'
primary_region = 'iad'
kill_signal = "SIGINT"
kill_timeout = "5s"
[build]
dockerfile = "Dockerfile"
[deploy]
release_command = "echo 'hello world'"
With this simple release_command
, our application will fail to deploy, offering the following error similar to @RoboZoom:
✖ Failed: error getting release_command logs: 422 Unprocessable Entity
Error: release command failed - aborting deployment. error getting release_command logs: 422 Unprocessable Entity
Checking the Grafana Logs dashboard adds a little more color, showing this might be related to some networking or the Firecracker runtime, but still not very helpful:
INFO [fly api proxy] listening at /.fly/api
INFO Preparing to run: `docker-entrypoint.sh cd ../../packages/db && echo 'hello world'` as root
INFO Starting init (commit: 04656915)...
Similar to @RoboZoom, we are not able to get any more context on our logs via the fly logs
command for this app.
To help the FlyIO team or other community members gain more context to our application, we can share that the Dockerfile was always working up until this recent addition, and has a final lines to start the app as:
RUN pnpm install --frozen-lockfile
ENV NODE_ENV production
EXPOSE 3000
WORKDIR /app/apps/web
CMD ["pnpm", "start"]
Furthermore, we have tried and verified the exemplary failing release command on a locally built docker container, as well as the intended release command that we will soon replace echo 'hello world'
with (a migration script), both of which worked just fine on our local Docker container.
- Is this a FlyIO platform error? If so then why is this error not reported on the FlyIO Status page?
- Does this have to do with the instance being in the
iad
region specifically? One or two other forum posts seemed to have this as a potential root cause w/DNS in iad
- How can one monitor-for and prevent this critical failure in the future if something like production database migrations depends on it?
Thank you for any help here!