Simon Willison

Results 1833 comments of Simon Willison

Actually I disagree that Datasette isn't setup to make use of container scaling: part of the idea behind the [Baked Data](https://simonwillison.net/2021/Jul/28/baked-data/) pattern is that you can scale to handle effectively...

Here's the relevant part of the `datasette publish` command from that failed Actions workflow: ``` datasette publish cloudrun f7.db nlrb.db opdr.db old_nlrb.db voluntary_recognitions.db work_stoppages.db lm20.db chips.db \ --memory 8Gi \...

Tried duplicating this error locally but the following command succeeded when I expected it to fail: datasette publish cloudrun fixtures.db --memory 8Gi --cpu 2 --service issue-1779

Maybe I need to upgrade: ``` % gcloud --version Google Cloud SDK 378.0.0 alpha 2022.03.18 bq 2.0.74 core 2022.03.18 gsutil 5.8 Updates are available for some Google Cloud CLI components....

`datasette publish cloudrun fixtures.db --memory 8Gi --cpu 2 --service issue-1779` still works.

Just spotted this in the failing Actions workflow: gcloud config set run/region us-central1 I tried that locally too but the deploy still succeeds.

Just tried this instead, and it still worked and deployed OK: datasette publish cloudrun fixtures.db --memory 16Gi --cpu 4 --service issue-1779 @fgregg I'm not able to replicate your deployment failure...

(I deleted my `issue-1779` project using the UI at https://console.cloud.google.com/run?project=datasette-222320)

Here's the start of the man page for `gcloud run deploy`: ``` NAME gcloud run deploy - deploy a container to Cloud Run SYNOPSIS gcloud run deploy [[SERVICE] --namespace=NAMESPACE] [--args=[ARG,...]]...

Tested that with: datasette publish cloudrun fixtures.db --service issue-1779 --min-instances 2 --max-instances 4