spark-operator icon indicating copy to clipboard operation
spark-operator copied to clipboard

Missing docker pull gcr.io/spark-operator/spark:v3.1.1 image

Open oschrenk opened this issue 2 years ago • 12 comments

To move forward with a legacy application, I need to pull gcr.io/spark-operator/spark:v3.1.1 image

It seems that image does not exist anymore

❯ docker pull gcr.io/spark-operator/spark:v3.1.1

Maybe somebody removed it, because of:

https://github.com/GoogleCloudPlatform/spark-on-k8s-operator/issues/1800

what would be valid alternatives?

oschrenk avatar Nov 02 '23 14:11 oschrenk

Or is there some way to build it myself?

The Dockerfile in root already requires that image

❯ docker build .
[+] Building 0.5s (4/4) FINISHED                                                  docker:desktop-linux
 => [internal] load build definition from Dockerfile                                              0.0s
 => => transferring dockerfile: 1.49kB                                                            0.0s
 => [internal] load .dockerignore                                                                 0.0s
 => => transferring context: 47B                                                                  0.0s
 => ERROR [internal] load metadata for gcr.io/spark-operator/spark:v3.1.1                         0.5s
 => CANCELED [internal] load metadata for docker.io/library/golang:1.19.2-alpine                  0.5s
------
 > [internal] load metadata for gcr.io/spark-operator/spark:v3.1.1:
------
Dockerfile:37
--------------------
  35 |     RUN CGO_ENABLED=0 GOOS=linux GOARCH=amd64 GO111MODULE=on go build -a -o /usr/bin/spark-operator main.go
  36 |
  37 | >>> FROM ${SPARK_IMAGE}
  38 |     USER root
  39 |     COPY --from=builder /usr/bin/spark-operator /usr/bin/
--------------------
ERROR: failed to solve: gcr.io/spark-operator/spark:v3.1.1: gcr.io/spark-operator/spark:v3.1.1: not found

oschrenk avatar Nov 02 '23 14:11 oschrenk

Also encountering the same issue on both my k8s cluster:

Warning  Failed     22m (x4 over 23m)     kubelet            Failed to pull image "gcr.io/spark-operator/spark:v3.1.1": rpc error: code = Unknown desc = Error response from daemon: manifest for gcr.io/spar
k-operator/spark:v3.1.1 not found: manifest unknown: Failed to fetch "v3.1.1" from request "/v2/spark-operator/spark/manifests/v3.1.1".

and by using docker pull gcr.io/spark-operator/spark:v3.1.1:

Error response from daemon: manifest for gcr.io/spark-operator/spark:v3.1.1 not found: manifest unknown: Failed to fetch "v3.1.1" from request "/v2/spark-operator/spark/manifests/v3.1.1".

Seems like the image is no longer available?

Checking https://console.cloud.google.com/gcr/images/spark-operator/GLOBAL/spark and no images exist.

Souheil-Yazji avatar Nov 06 '23 14:11 Souheil-Yazji

Also encountering the same issue on both my k8s cluster: and by using docker pull gcr.io/spark-operator/spark:v3.1.1:

If is is possible, use docker pull apache/spark:v3.1.3 instead

plfred avatar Nov 06 '23 14:11 plfred

Once we found out about the missing docker image, we've been using images generated from the https://github.com/apache/spark-docker repository without any major problems.

jalvarez avatar Nov 07 '23 16:11 jalvarez

Once we found out about the missing docker image, we've been using images generated from the https://github.com/apache/spark-docker repository without any major problems.

There is no 3.1.1? Which is the version recommended

AlejandroUPC avatar Nov 08 '23 06:11 AlejandroUPC

Is there any spark-py image which works with python2.7 for time being i migrate my code to python3 ?

gcr.io/spark-operator/spark:v3.0.0 was working fine for me.

Ayushri avatar Nov 29 '23 12:11 Ayushri

Hi @jalvarez I am trying to setup the Spark on K8s using spark-on-k8s-operator, if we build the Apache spark docker image from https://github.com/apache/spark-docker repository then should we change mainApplicationFile parameter in one of the examples code?

https://github.com/GoogleCloudPlatform/spark-on-k8s-operator/blob/master/examples/spark-pi.yaml#L27

vinayak-shanawad avatar Jan 23 '24 06:01 vinayak-shanawad

Hi @jalvarez I am trying to setup the Spark on K8s using spark-on-k8s-operator, if we build the Apache spark docker image from https://github.com/apache/spark-docker repository then should we change mainApplicationFile parameter in one of the examples code?

https://github.com/GoogleCloudPlatform/spark-on-k8s-operator/blob/master/examples/spark-pi.yaml#L27

If you build spark image from version 3.1.1, then you needn't change mainApplicationFile parameter.

ChenYi015 avatar Jan 25 '24 07:01 ChenYi015

Currently all the examples in https://github.com/GoogleCloudPlatform/spark-on-k8s-operator/tree/master/examples appear to be referencing this non existing image. This implies the quick start guide is also broken.

Is there any properly working repository anywhere?

dbackeus avatar Feb 13 '24 14:02 dbackeus

Hopefully someone will merge the PR or correct all examples + Dockerfile.

brito-rafa avatar Feb 22 '24 19:02 brito-rafa

Hopefully someone will merge the PR or correct all examples + Dockerfile.

Probably not, given that this repo may soon be transferred to a new owner (the Kubeflow Org)

Souheil-Yazji avatar Feb 23 '24 20:02 Souheil-Yazji

I am referencing the tutorial and can confirm that it complains about the missing image.

image

https://github.com/kubeflow/spark-operator/blob/master/examples/spark-pi.yaml

kenho811 avatar Apr 23 '24 14:04 kenho811

This issue has been automatically marked as stale because it has not had recent activity. It will be closed if no further activity occurs. Thank you for your contributions.

github-actions[bot] avatar Aug 14 '24 04:08 github-actions[bot]