Constantly getting this error when running now scale xyz:
> Error! An unexpected error occurred in scale: Error: Timeout while verifying instance count (10m)
at waitForScale (/snapshot/repo/dist/now.js:8878:13)
at <anonymous>
+1
Getting when deploying to bru1
+1 to sf1
> Your deployment's code and logs will be publicly accessible because you are subscribed to the OSS plan.
> NOTE: You can use `now --public` or upgrade your plan (https://zeit.co/account/plan) to skip this prompt
> Using Node.js 8.11.1 (default)
> https://experiment-qcjvbbvnfs.now.sh [in clipboard] (bru1) [912ms]
> Building…
> â–² npm install
> â–² Snapshotting deployment
> â–² Saving deployment image (463)
> Build completed
â § Verifying instances in bru1> Error! An unexpected error occurred!
Error: Timeout while verifying instance count (10m)
at waitForScale (/snapshot/repo/dist/now.js:6281:13)
at <anonymous> Error: Timeout while verifying instance count (10m)
at waitForScale (/snapshot/repo/dist/now.js:6281:13)
at <anonymous>
same here, deploying to bru1
We are looking into this. To confirm, do you see anything in the app logs that would point to startup issues / not opening a HTTP port?
The app I uploaded doesn't record logs, the deployment was successful after a couple of tries, but the url provided gives 503 error, here's the given link https://brantu-task-ctmovfwubg.now.sh
what was happening is that the upload process goes fine but it gets stuck at verifying and then times out
Same here. Just sent you guys an email before heading here to check the conversation.

I get the same on sfo1 trying to deploy anything with now
Same!
Verifying instances in sfo1> Error! An unexpected error occurred!
Error: Timeout while verifying instance count (10m)
at waitForScale (/snapshot/repo/dist/now.js:6281:13)
at <anonymous> Error: Timeout while verifying instance count (10m)
at waitForScale (/snapshot/repo/dist/now.js:6281:13)
at <anonymous>
Ah! There was an error in my code, which I only discovered by deploying via localhost instead of now. Once fixed, I no longer received a timeout. So it might only occur if there’s an unrelated bug.
/Users/__USER__/Apps/__MY_APP__/node_modules/graphql-yoga/node_modules/graphql-tools/dist/schemaGenerator.js:341
throw new SchemaError(typeName + "." + fieldName + " defined in resolvers, but not in schema");
^
@heymartinadams So were you able to successfully deploy after this?
Yes indeed, @nnennajohn! 😎
Bummer for me. Successfully got it on heroku though, so I know its not some rouge function. :) Will wait to hear from the support team after the weekend. Hopeful. :)
+1, both bru and sfo, timeout
Sorry, getting the same with bru1 and sfo1 since last night.
update: I also had broken code (third party dep) and it works fine after fixing that, same as @heymartinadams. This is a non issue for me!
Or rather... this is a messaging issue.
Any updates on this? this is holding me back.
@bookofdom see if you can deploy to localhost using the node or nodemon command in your CLI — this should give you an indication to any errors in your code (that are unrelated to now).
Just wanted to add, I'm seeing this too. It's only affecting one of my projects. I'm able to build and run it locally:

i, too, am getting this error:
â ¼ Verifying instances in sfo1> Error! An unexpected error occurred!
Error: Timeout while verifying instance count (10m)
at waitForScale (/snapshot/repo/dist/now.js:8446:13)
at <anonymous> Error: Timeout while verifying instance count (10m)
at waitForScale (/snapshot/repo/dist/now.js:8446:13)
at <anonymous>
there are no errors in my code, i ran a standard JS task to clean and report any JS error, which came back with none.
Same here! Waiting for a way to deploy my app. :O
@rauchg It seems that the deploy process doesn't start the Node.js application. Earlier you saw the application bootstrapping before verifying the Now instances.
Yeah, same here, I deployed my app built on Probot on Now a month back. At that time, it was working fine.
After making a few changes, I redeployed it after updating CLI but faced similar error.
Subscribing this thread to receive new updates on the same. 🙂
This should never happen again in the latest both stable and canary releases!
Please reopen if anybody else experiences this issue.
Yup keep seeing this again more and more last few days in our deployments (always bru1). This is completely breaking down our continuous deployments for 2 projects:

Note sure what causes the log rate limiting, but we are on a paid account and far away from any limits nor do we deploy excessively.
Update:
This might be only happening on our CI agents where its pulling the latest version of the now CLI during every build. I'll check if this goes away to setting this to same as local version, but if that makes it go away then an error was introduced in a later version.
I'm pretty sure this is happening to you because you ran into a logs rate limit so we are unable to get the logging events to ensure that the instance is running. Temporarily you can just run the deploy with --no-verify but maybe we should consider this case on our side @rauchg
It's happened again to me. my app, and I also tested with now-example/hello-go. When I used --no-verify, deploying was completed but the apps didn't work well.
now
> Deploying ~/Workspace/now-examples/go-hello under
> https://go-hello-zfxrcdeozk.now.sh [in clipboard] (sfo1) [4s]
> Building…
> Sending build context to Docker daemon 4.096kB
> Step 1/7 : FROM golang:1.9-alpine as base
> ---> b0260be938c6
> Step 2/7 : WORKDIR /usr/src
> ---> Using cache
> ---> 3c896b471e39
> Step 3/7 : COPY . .
> ---> 606d53d6f6f1
> Step 4/7 : RUN CGO_ENABLED=0 go build -ldflags "-s -w" -o main
> ---> Running in 1d409d44bdc9
> Removing intermediate container 1d409d44bdc9
> ---> 323e9d832b79
> Step 5/7 : FROM scratch
> --->
> Step 6/7 : COPY --from=base /usr/src/main /go-http-microservice
> ---> cfe6c5685dea
> Step 7/7 : CMD ["/go-http-microservice"]
> ---> Running in da80dd77058f
> Removing intermediate container da80dd77058f
> ---> 15f3d2c0f800
> Successfully built 15f3d2c0f800
> Successfully tagged registry.now.systems/now/685e5f16b3f11c86f95fb93bcbe4b83178a48758:latest
> â–² Storing image
> Build completed
â ¹ Verifying instances in sfo1> Error! An unexpected error occurred!
Error: Timeout while verifying instance count (10m)
at waitForScale (/snapshot/repo/dist/now.js:6281:13)
at <anonymous> Error: Timeout while verifying instance count (10m)
at waitForScale (/snapshot/repo/dist/now.js:6281:13)
at <anonymous>
I'm having same here :raised_hand_with_fingers_splayed: (I'm using now-cli version 12.1.2)
UPDATE: I changed the region from bru1 to sfo1 and it seems working back.. issues with bru1 region? I have another deployment on bru1 though.
All the process went fine, here the logs from the deployment:
11/14 12:11 PM (6m) yarn
11/14 12:11 PM (6m) yarn install v1.9.4
11/14 12:11 PM (6m) [1/4] Resolving packages...
11/14 12:11 PM (6m) [2/4] Fetching packages...
11/14 12:12 PM (6m) info [email protected]: The platform "linux" is incompatible with this module.
11/14 12:12 PM (6m) info "[email protected]" is an optional dependency and failed compatibility check.Excluding it from installation.
11/14 12:12 PM (6m) info [email protected]: The platform "linux" is incompatible with this module.
11/14 12:12 PM (6m) info "[email protected]" is an optional dependency and failed compatibility check. Excluding it from installation.
11/14 12:12 PM (6m) [3/4] Linking dependencies...
11/14 12:12 PM (6m) [4/4] Building fresh packages...
11/14 12:12 PM (6m) Done in 23.48s.
11/14 12:12 PM (6m) npm run build
11/14 12:12 PM (6m)
11/14 12:12 PM (6m) > [email protected] build /home/nowuser/src
11/14 12:12 PM (6m) > react-scripts build
11/14 12:12 PM (6m)
11/14 12:12 PM (6m) Creating an optimized production build...
11/14 12:12 PM (6m) Compiled successfully.
11/14 12:12 PM (6m)
11/14 12:12 PM (6m) File sizes after gzip:
11/14 12:12 PM (6m)
11/14 12:12 PM (6m) 144.03 KB build/static/js/main.88c0f8b5.js
11/14 12:12 PM (6m) 18.72 KB build/static/css/main.b014df2d.css
11/14 12:12 PM (6m)
11/14 12:12 PM (6m) The project was built assuming it is hosted at the server root.
11/14 12:12 PM (6m) To override this, specify the homepage in your package.json.
11/14 12:12 PM (6m) For example, add this to build it for GitHub Pages:
11/14 12:12 PM (6m)
11/14 12:12 PM (6m) "homepage": "http://myname.github.io/myapp",
11/14 12:12 PM (6m)
11/14 12:12 PM (6m) The build folder is ready to be deployed.
11/14 12:12 PM (6m) You may serve it with a static server:
11/14 12:12 PM (6m)
11/14 12:12 PM (6m) yarn global add serve
11/14 12:12 PM (6m) serve -s build
11/14 12:12 PM (6m)
11/14 12:12 PM (5m) Snapshotting deployment
11/14 12:12 PM (5m) Saving deployment image (25.8M)
$ now
....
> â–² Snapshotting deployment
> â–² Saving deployment image (25.8M)
> Build completed
> Verifying instantiation in bru1
> Error! Instance verification timed out (3m)
> Read more: https://err.sh/now-cli/verification-timeout
I am having the same problem with bru over the last few days.
Every now and then my deployment refuses to scale in bru with verification timeout (yes, it really is not running). There is no error in logs.
@hleumas We have fixed a problem with scaling affecting BRU1 a few hours ago. Can you retry?
It's working seamlessly. Thanks @paulogdm .
+1
I have the same issue in BRU1 region today. Is there a way to increase timeout?
We have not been able to scale up BRU1 on the first try for about a month.
Between 40% and 80% of my deploys fail (it varies from day to day( on BRU1 due to this problem. I'm running out of options.
Today I've had 2 successful deploys out of 12 and still going.
This should never happen again in the latest both stable and canary releases!
Please reopen if anybody else experiences this issue.
@javivelasco This should be reopened. I have instance verification timeouts for latest stable release versions in bru1. Tested versions 14.2.0 on up all the way to lts 15.3.0.
It feels like there just aren't enough resources allocated for bru1 anymore:

Is this true?
Most helpful comment
This should never happen again in the latest both stable and canary releases!
Please reopen if anybody else experiences this issue.