-
Notifications
You must be signed in to change notification settings - Fork 1.7k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
v2 compatible on full Kubeflow #5680
Comments
/assign @Bobgy |
I'm trying to run v2 compatible pipeline on full Kubeflow, as expected, it fails when uploading to minio storage:
|
Interesting, when I switch pipeline_root to a gcs path like gs://gongyuan-dev/v2/. I checked how that's configured -- the mlmd grpc service host and port come from k8s env vars:
EDIT: I was wrong, the env vars come from a configmap, see
|
Because configmaps needs to be deployed, I propose a simpler default mechanism:
Going forward, a proper configuration for artifact storage will need to be introduced -- that's when we can make this configurable. |
UPDATE: after trying this out, I found that this doesn't work on Kubeflow on GCP, because we do not use the bucket Therefore, I think the current approach is probably fine for full Kubeflow on other platforms with MinIO, but not a good fit for Kubeflow on GCP. |
So actually, configurability of credentials is not a must have for making v2compat run on full Kubeflow. It's enough to make default pipeline root configurable. For KF on GCP, we can default to gcs path. For others, using minio://mlpipeline is enough. |
/reopen |
@Bobgy: Reopened this issue. In response to this:
Instructions for interacting with me using PR comments are available here. If you have questions or suggestions related to my behavior, please file an issue against the kubernetes/test-infra repository. |
I think one issue that I previously missed is that, we need to make the new task API available from pipeline pods. That requires some istio authorization policy change. |
TODOs:
|
UPDATE:
PRs have been written:
===
This issue tracks efforts to make v2 compatible pipelines run on full Kubeflow.
Depends on #4649
The major differences from KFP standalone are:
mlpipeline
.For MLMD grpc service, we are already reading a configmap called
metadata-grpc-configmap
. The configmap is already configured in full Kubeflow for each user namespace:pipelines/manifests/kustomize/base/installs/multi-user/pipelines-profile-controller/sync.py
Lines 52 to 63 in 7608fcd
Therefore, it already works out of the box.
Because in Kubeflow on GCP, the default bucket is not
mlpipeline
, we need configurations that can set default pipeline root and credentials for each namespace, this is essentially #4649 for v2 compatible pipelines.The following comments are investigation logs towards the conclusion I summarized here.
The text was updated successfully, but these errors were encountered: