Skip to content

Commit

Permalink
Updated version to 1.0.0-rc.2
Browse files Browse the repository at this point in the history
  • Loading branch information
Bobgy committed Jun 16, 2020
1 parent e27f474 commit 0796340
Show file tree
Hide file tree
Showing 65 changed files with 83 additions and 83 deletions.
2 changes: 1 addition & 1 deletion VERSION
Original file line number Diff line number Diff line change
@@ -1 +1 @@
1.0.0-rc.1
1.0.0-rc.2
4 changes: 2 additions & 2 deletions backend/api/python_http_client/README.md
Original file line number Diff line number Diff line change
Expand Up @@ -3,8 +3,8 @@ This file contains REST API specification for Kubeflow Pipelines. The file is au

This Python package is automatically generated by the [OpenAPI Generator](https://openapi-generator.tech) project:

- API version: 1.0.0-rc.1
- Package version: 1.0.0-rc.1
- API version: 1.0.0-rc.2
- Package version: 1.0.0-rc.2
- Build package: org.openapitools.codegen.languages.PythonClientCodegen
For more information, please visit [https://www.google.com](https://www.google.com)

Expand Down
2 changes: 1 addition & 1 deletion backend/api/python_http_client/kfp_server_api/__init__.py
Original file line number Diff line number Diff line change
Expand Up @@ -28,7 +28,7 @@

from __future__ import absolute_import

__version__ = "1.0.0-rc.1"
__version__ = "1.0.0-rc.2"

# import apis into sdk package
from kfp_server_api.api.experiment_service_api import ExperimentServiceApi
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -92,7 +92,7 @@ def __init__(self, configuration=None, header_name=None, header_value=None,
self.default_headers[header_name] = header_value
self.cookie = cookie
# Set default User-Agent.
self.user_agent = 'OpenAPI-Generator/1.0.0-rc.1/python'
self.user_agent = 'OpenAPI-Generator/1.0.0-rc.2/python'
self.client_side_validation = configuration.client_side_validation

def __enter__(self):
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -365,8 +365,8 @@ def to_debug_report(self):
return "Python SDK Debug Report:\n"\
"OS: {env}\n"\
"Python Version: {pyversion}\n"\
"Version of the API: 1.0.0-rc.1\n"\
"SDK Package Version: 1.0.0-rc.1".\
"Version of the API: 1.0.0-rc.2\n"\
"SDK Package Version: 1.0.0-rc.2".\
format(env=sys.platform, pyversion=sys.version)

def get_host_settings(self):
Expand Down
2 changes: 1 addition & 1 deletion backend/api/python_http_client/setup.py
Original file line number Diff line number Diff line change
Expand Up @@ -27,7 +27,7 @@
from setuptools import setup, find_packages # noqa: H301

NAME = "kfp-server-api"
VERSION = "1.0.0-rc.1"
VERSION = "1.0.0-rc.2"
# To install the library, run the following
#
# python setup.py install
Expand Down
2 changes: 1 addition & 1 deletion backend/api/swagger/kfp_api_single_file.swagger.json

Some generated files are not rendered by default. Learn more about how customized files appear on GitHub.

2 changes: 1 addition & 1 deletion components/gcp/bigquery/query/README.md
Original file line number Diff line number Diff line change
Expand Up @@ -85,7 +85,7 @@ KFP_PACKAGE = 'https://storage.googleapis.com/ml-pipeline/release/0.1.14/kfp.tar
import kfp.components as comp

bigquery_query_op = comp.load_component_from_url(
'https://raw.githubusercontent.com/kubeflow/pipelines/1.0.0-rc.1/components/gcp/bigquery/query/component.yaml')
'https://raw.githubusercontent.com/kubeflow/pipelines/1.0.0-rc.2/components/gcp/bigquery/query/component.yaml')
help(bigquery_query_op)
```

Expand Down
2 changes: 1 addition & 1 deletion components/gcp/bigquery/query/component.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -59,7 +59,7 @@ outputs:
type: UI metadata
implementation:
container:
image: gcr.io/ml-pipeline/ml-pipeline-gcp:1.0.0-rc.1
image: gcr.io/ml-pipeline/ml-pipeline-gcp:1.0.0-rc.2
args: [
--ui_metadata_path, {outputPath: MLPipeline UI metadata},
kfp_component.google.bigquery, query,
Expand Down
2 changes: 1 addition & 1 deletion components/gcp/bigquery/query/sample.ipynb
Original file line number Diff line number Diff line change
Expand Up @@ -104,7 +104,7 @@
"import kfp.components as comp\n",
"\n",
"bigquery_query_op = comp.load_component_from_url(\n",
" 'https://raw.githubusercontent.com/kubeflow/pipelines/1.0.0-rc.1/components/gcp/bigquery/query/component.yaml')\n",
" 'https://raw.githubusercontent.com/kubeflow/pipelines/1.0.0-rc.2/components/gcp/bigquery/query/component.yaml')\n",
"help(bigquery_query_op)"
]
},
Expand Down
2 changes: 1 addition & 1 deletion components/gcp/container/component_sdk/python/setup.py
Original file line number Diff line number Diff line change
Expand Up @@ -15,7 +15,7 @@
from setuptools import setup

PACKAGE_NAME = 'kfp-component'
VERSION = '1.0.0-rc.1'
VERSION = '1.0.0-rc.2'

setup(
name=PACKAGE_NAME,
Expand Down
2 changes: 1 addition & 1 deletion components/gcp/dataflow/launch_python/README.md
Original file line number Diff line number Diff line change
Expand Up @@ -91,7 +91,7 @@ The steps to use the component in a pipeline are:
```python
import kfp.components as comp

dataflow_python_op = comp.load_component_from_url('https://raw.githubusercontent.com/kubeflow/pipelines/1.0.0-rc.1/components/gcp/dataflow/launch_python/component.yaml')
dataflow_python_op = comp.load_component_from_url('https://raw.githubusercontent.com/kubeflow/pipelines/1.0.0-rc.2/components/gcp/dataflow/launch_python/component.yaml')
help(dataflow_python_op)
```

Expand Down
2 changes: 1 addition & 1 deletion components/gcp/dataflow/launch_python/component.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -53,7 +53,7 @@ outputs:
type: UI metadata
implementation:
container:
image: gcr.io/ml-pipeline/ml-pipeline-gcp:1.0.0-rc.1
image: gcr.io/ml-pipeline/ml-pipeline-gcp:1.0.0-rc.2
args: [
--ui_metadata_path, {outputPath: MLPipeline UI metadata},
kfp_component.google.dataflow, launch_python,
Expand Down
2 changes: 1 addition & 1 deletion components/gcp/dataflow/launch_python/sample.ipynb
Original file line number Diff line number Diff line change
Expand Up @@ -92,7 +92,7 @@
"import kfp.components as comp\n",
"\n",
"dataflow_python_op = comp.load_component_from_url(\n",
" 'https://raw.githubusercontent.com/kubeflow/pipelines/1.0.0-rc.1/components/gcp/dataflow/launch_python/component.yaml')\n",
" 'https://raw.githubusercontent.com/kubeflow/pipelines/1.0.0-rc.2/components/gcp/dataflow/launch_python/component.yaml')\n",
"help(dataflow_python_op)"
]
},
Expand Down
2 changes: 1 addition & 1 deletion components/gcp/dataflow/launch_template/README.md
Original file line number Diff line number Diff line change
Expand Up @@ -64,7 +64,7 @@ KFP_PACKAGE = 'https://storage.googleapis.com/ml-pipeline/release/0.1.14/kfp.tar
import kfp.components as comp

dataflow_template_op = comp.load_component_from_url(
'https://raw.githubusercontent.com/kubeflow/pipelines/1.0.0-rc.1/components/gcp/dataflow/launch_template/component.yaml')
'https://raw.githubusercontent.com/kubeflow/pipelines/1.0.0-rc.2/components/gcp/dataflow/launch_template/component.yaml')
help(dataflow_template_op)
```

Expand Down
2 changes: 1 addition & 1 deletion components/gcp/dataflow/launch_template/component.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -63,7 +63,7 @@ outputs:
type: UI metadata
implementation:
container:
image: gcr.io/ml-pipeline/ml-pipeline-gcp:1.0.0-rc.1
image: gcr.io/ml-pipeline/ml-pipeline-gcp:1.0.0-rc.2
args: [
--ui_metadata_path, {outputPath: MLPipeline UI metadata},
kfp_component.google.dataflow, launch_template,
Expand Down
2 changes: 1 addition & 1 deletion components/gcp/dataflow/launch_template/sample.ipynb
Original file line number Diff line number Diff line change
Expand Up @@ -82,7 +82,7 @@
"import kfp.components as comp\n",
"\n",
"dataflow_template_op = comp.load_component_from_url(\n",
" 'https://raw.githubusercontent.com/kubeflow/pipelines/1.0.0-rc.1/components/gcp/dataflow/launch_template/component.yaml')\n",
" 'https://raw.githubusercontent.com/kubeflow/pipelines/1.0.0-rc.2/components/gcp/dataflow/launch_template/component.yaml')\n",
"help(dataflow_template_op)"
]
},
Expand Down
2 changes: 1 addition & 1 deletion components/gcp/dataproc/create_cluster/README.md
Original file line number Diff line number Diff line change
Expand Up @@ -88,7 +88,7 @@ Follow these steps to use the component in a pipeline:
```python
import kfp.components as comp

dataproc_create_cluster_op = comp.load_component_from_url('https://raw.githubusercontent.com/kubeflow/pipelines/1.0.0-rc.1/components/gcp/dataproc/create_cluster/component.yaml')
dataproc_create_cluster_op = comp.load_component_from_url('https://raw.githubusercontent.com/kubeflow/pipelines/1.0.0-rc.2/components/gcp/dataproc/create_cluster/component.yaml')
help(dataproc_create_cluster_op)
```

Expand Down
2 changes: 1 addition & 1 deletion components/gcp/dataproc/create_cluster/component.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -70,7 +70,7 @@ outputs:
type: UI metadata
implementation:
container:
image: gcr.io/ml-pipeline/ml-pipeline-gcp:1.0.0-rc.1
image: gcr.io/ml-pipeline/ml-pipeline-gcp:1.0.0-rc.2
args: [
--ui_metadata_path, {outputPath: MLPipeline UI metadata},
kfp_component.google.dataproc, create_cluster,
Expand Down
2 changes: 1 addition & 1 deletion components/gcp/dataproc/create_cluster/sample.ipynb
Original file line number Diff line number Diff line change
Expand Up @@ -88,7 +88,7 @@
"import kfp.components as comp\n",
"\n",
"dataproc_create_cluster_op = comp.load_component_from_url(\n",
" 'https://raw.githubusercontent.com/kubeflow/pipelines/1.0.0-rc.1/components/gcp/dataproc/create_cluster/component.yaml')\n",
" 'https://raw.githubusercontent.com/kubeflow/pipelines/1.0.0-rc.2/components/gcp/dataproc/create_cluster/component.yaml')\n",
"help(dataproc_create_cluster_op)"
]
},
Expand Down
2 changes: 1 addition & 1 deletion components/gcp/dataproc/delete_cluster/README.md
Original file line number Diff line number Diff line change
Expand Up @@ -66,7 +66,7 @@ Follow these steps to use the component in a pipeline:
```python
import kfp.components as comp

dataproc_delete_cluster_op = comp.load_component_from_url('https://raw.githubusercontent.com/kubeflow/pipelines/1.0.0-rc.1/components/gcp/dataproc/delete_cluster/component.yaml')
dataproc_delete_cluster_op = comp.load_component_from_url('https://raw.githubusercontent.com/kubeflow/pipelines/1.0.0-rc.2/components/gcp/dataproc/delete_cluster/component.yaml')
help(dataproc_delete_cluster_op)
```

Expand Down
2 changes: 1 addition & 1 deletion components/gcp/dataproc/delete_cluster/component.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -36,7 +36,7 @@ inputs:
type: Integer
implementation:
container:
image: gcr.io/ml-pipeline/ml-pipeline-gcp:1.0.0-rc.1
image: gcr.io/ml-pipeline/ml-pipeline-gcp:1.0.0-rc.2
args: [
kfp_component.google.dataproc, delete_cluster,
--project_id, {inputValue: project_id},
Expand Down
2 changes: 1 addition & 1 deletion components/gcp/dataproc/delete_cluster/sample.ipynb
Original file line number Diff line number Diff line change
Expand Up @@ -71,7 +71,7 @@
"import kfp.components as comp\n",
"\n",
"dataproc_delete_cluster_op = comp.load_component_from_url(\n",
" 'https://raw.githubusercontent.com/kubeflow/pipelines/1.0.0-rc.1/components/gcp/dataproc/delete_cluster/component.yaml')\n",
" 'https://raw.githubusercontent.com/kubeflow/pipelines/1.0.0-rc.2/components/gcp/dataproc/delete_cluster/component.yaml')\n",
"help(dataproc_delete_cluster_op)"
]
},
Expand Down
2 changes: 1 addition & 1 deletion components/gcp/dataproc/submit_hadoop_job/README.md
Original file line number Diff line number Diff line change
Expand Up @@ -83,7 +83,7 @@ Follow these steps to use the component in a pipeline:
```python
import kfp.components as comp

dataproc_submit_hadoop_job_op = comp.load_component_from_url('https://raw.githubusercontent.com/kubeflow/pipelines/1.0.0-rc.1/components/gcp/dataproc/submit_hadoop_job/component.yaml')
dataproc_submit_hadoop_job_op = comp.load_component_from_url('https://raw.githubusercontent.com/kubeflow/pipelines/1.0.0-rc.2/components/gcp/dataproc/submit_hadoop_job/component.yaml')
help(dataproc_submit_hadoop_job_op)
```

Expand Down
2 changes: 1 addition & 1 deletion components/gcp/dataproc/submit_hadoop_job/component.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -80,7 +80,7 @@ outputs:
type: UI metadata
implementation:
container:
image: gcr.io/ml-pipeline/ml-pipeline-gcp:1.0.0-rc.1
image: gcr.io/ml-pipeline/ml-pipeline-gcp:1.0.0-rc.2
args: [
--ui_metadata_path, {outputPath: MLPipeline UI metadata},
kfp_component.google.dataproc, submit_hadoop_job,
Expand Down
2 changes: 1 addition & 1 deletion components/gcp/dataproc/submit_hadoop_job/sample.ipynb
Original file line number Diff line number Diff line change
Expand Up @@ -86,7 +86,7 @@
"import kfp.components as comp\n",
"\n",
"dataproc_submit_hadoop_job_op = comp.load_component_from_url(\n",
" 'https://raw.githubusercontent.com/kubeflow/pipelines/1.0.0-rc.1/components/gcp/dataproc/submit_hadoop_job/component.yaml')\n",
" 'https://raw.githubusercontent.com/kubeflow/pipelines/1.0.0-rc.2/components/gcp/dataproc/submit_hadoop_job/component.yaml')\n",
"help(dataproc_submit_hadoop_job_op)"
]
},
Expand Down
2 changes: 1 addition & 1 deletion components/gcp/dataproc/submit_hive_job/README.md
Original file line number Diff line number Diff line change
Expand Up @@ -73,7 +73,7 @@ Follow these steps to use the component in a pipeline:
```python
import kfp.components as comp

dataproc_submit_hive_job_op = comp.load_component_from_url('https://raw.githubusercontent.com/kubeflow/pipelines/1.0.0-rc.1/components/gcp/dataproc/submit_hive_job/component.yaml')
dataproc_submit_hive_job_op = comp.load_component_from_url('https://raw.githubusercontent.com/kubeflow/pipelines/1.0.0-rc.2/components/gcp/dataproc/submit_hive_job/component.yaml')
help(dataproc_submit_hive_job_op)
```

Expand Down
2 changes: 1 addition & 1 deletion components/gcp/dataproc/submit_hive_job/component.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -75,7 +75,7 @@ outputs:
type: UI metadata
implementation:
container:
image: gcr.io/ml-pipeline/ml-pipeline-gcp:1.0.0-rc.1
image: gcr.io/ml-pipeline/ml-pipeline-gcp:1.0.0-rc.2
args: [
--ui_metadata_path, {outputPath: MLPipeline UI metadata},
kfp_component.google.dataproc, submit_hive_job,
Expand Down
2 changes: 1 addition & 1 deletion components/gcp/dataproc/submit_hive_job/sample.ipynb
Original file line number Diff line number Diff line change
Expand Up @@ -77,7 +77,7 @@
"import kfp.components as comp\n",
"\n",
"dataproc_submit_hive_job_op = comp.load_component_from_url(\n",
" 'https://raw.githubusercontent.com/kubeflow/pipelines/1.0.0-rc.1/components/gcp/dataproc/submit_hive_job/component.yaml')\n",
" 'https://raw.githubusercontent.com/kubeflow/pipelines/1.0.0-rc.2/components/gcp/dataproc/submit_hive_job/component.yaml')\n",
"help(dataproc_submit_hive_job_op)"
]
},
Expand Down
2 changes: 1 addition & 1 deletion components/gcp/dataproc/submit_pig_job/README.md
Original file line number Diff line number Diff line change
Expand Up @@ -82,7 +82,7 @@ Follow these steps to use the component in a pipeline:
```python
import kfp.components as comp

dataproc_submit_pig_job_op = comp.load_component_from_url('https://raw.githubusercontent.com/kubeflow/pipelines/1.0.0-rc.1/components/gcp/dataproc/submit_pig_job/component.yaml')
dataproc_submit_pig_job_op = comp.load_component_from_url('https://raw.githubusercontent.com/kubeflow/pipelines/1.0.0-rc.2/components/gcp/dataproc/submit_pig_job/component.yaml')
help(dataproc_submit_pig_job_op)
```

Expand Down
2 changes: 1 addition & 1 deletion components/gcp/dataproc/submit_pig_job/component.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -75,7 +75,7 @@ outputs:
type: UI metadata
implementation:
container:
image: gcr.io/ml-pipeline/ml-pipeline-gcp:1.0.0-rc.1
image: gcr.io/ml-pipeline/ml-pipeline-gcp:1.0.0-rc.2
args: [
--ui_metadata_path, {outputPath: MLPipeline UI metadata},
kfp_component.google.dataproc, submit_pig_job,
Expand Down
2 changes: 1 addition & 1 deletion components/gcp/dataproc/submit_pig_job/sample.ipynb
Original file line number Diff line number Diff line change
Expand Up @@ -80,7 +80,7 @@
"import kfp.components as comp\n",
"\n",
"dataproc_submit_pig_job_op = comp.load_component_from_url(\n",
" 'https://raw.githubusercontent.com/kubeflow/pipelines/1.0.0-rc.1/components/gcp/dataproc/submit_pig_job/component.yaml')\n",
" 'https://raw.githubusercontent.com/kubeflow/pipelines/1.0.0-rc.2/components/gcp/dataproc/submit_pig_job/component.yaml')\n",
"help(dataproc_submit_pig_job_op)"
]
},
Expand Down
2 changes: 1 addition & 1 deletion components/gcp/dataproc/submit_pyspark_job/README.md
Original file line number Diff line number Diff line change
Expand Up @@ -79,7 +79,7 @@ Follow these steps to use the component in a pipeline:
```python
import kfp.components as comp

dataproc_submit_pyspark_job_op = comp.load_component_from_url('https://raw.githubusercontent.com/kubeflow/pipelines/1.0.0-rc.1/components/gcp/dataproc/submit_pyspark_job/component.yaml')
dataproc_submit_pyspark_job_op = comp.load_component_from_url('https://raw.githubusercontent.com/kubeflow/pipelines/1.0.0-rc.2/components/gcp/dataproc/submit_pyspark_job/component.yaml')
help(dataproc_submit_pyspark_job_op)
```

Expand Down
2 changes: 1 addition & 1 deletion components/gcp/dataproc/submit_pyspark_job/component.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -69,7 +69,7 @@ outputs:
type: UI metadata
implementation:
container:
image: gcr.io/ml-pipeline/ml-pipeline-gcp:1.0.0-rc.1
image: gcr.io/ml-pipeline/ml-pipeline-gcp:1.0.0-rc.2
args: [
--ui_metadata_path, {outputPath: MLPipeline UI metadata},
kfp_component.google.dataproc, submit_pyspark_job,
Expand Down
2 changes: 1 addition & 1 deletion components/gcp/dataproc/submit_pyspark_job/sample.ipynb
Original file line number Diff line number Diff line change
Expand Up @@ -82,7 +82,7 @@
"import kfp.components as comp\n",
"\n",
"dataproc_submit_pyspark_job_op = comp.load_component_from_url(\n",
" 'https://raw.githubusercontent.com/kubeflow/pipelines/1.0.0-rc.1/components/gcp/dataproc/submit_pyspark_job/component.yaml')\n",
" 'https://raw.githubusercontent.com/kubeflow/pipelines/1.0.0-rc.2/components/gcp/dataproc/submit_pyspark_job/component.yaml')\n",
"help(dataproc_submit_pyspark_job_op)"
]
},
Expand Down
2 changes: 1 addition & 1 deletion components/gcp/dataproc/submit_spark_job/README.md
Original file line number Diff line number Diff line change
Expand Up @@ -95,7 +95,7 @@ Follow these steps to use the component in a pipeline:
import kfp.components as comp

dataproc_submit_spark_job_op = comp.load_component_from_url(
'https://raw.githubusercontent.com/kubeflow/pipelines/1.0.0-rc.1/components/gcp/dataproc/submit_spark_job/component.yaml')
'https://raw.githubusercontent.com/kubeflow/pipelines/1.0.0-rc.2/components/gcp/dataproc/submit_spark_job/component.yaml')
help(dataproc_submit_spark_job_op)
```

Expand Down
2 changes: 1 addition & 1 deletion components/gcp/dataproc/submit_spark_job/component.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -76,7 +76,7 @@ outputs:
type: UI metadata
implementation:
container:
image: gcr.io/ml-pipeline/ml-pipeline-gcp:1.0.0-rc.1
image: gcr.io/ml-pipeline/ml-pipeline-gcp:1.0.0-rc.2
args: [
--ui_metadata_path, {outputPath: MLPipeline UI metadata},
kfp_component.google.dataproc, submit_spark_job,
Expand Down
2 changes: 1 addition & 1 deletion components/gcp/dataproc/submit_spark_job/sample.ipynb
Original file line number Diff line number Diff line change
Expand Up @@ -93,7 +93,7 @@
"import kfp.components as comp\n",
"\n",
"dataproc_submit_spark_job_op = comp.load_component_from_url(\n",
" 'https://raw.githubusercontent.com/kubeflow/pipelines/1.0.0-rc.1/components/gcp/dataproc/submit_spark_job/component.yaml')\n",
" 'https://raw.githubusercontent.com/kubeflow/pipelines/1.0.0-rc.2/components/gcp/dataproc/submit_spark_job/component.yaml')\n",
"help(dataproc_submit_spark_job_op)"
]
},
Expand Down
2 changes: 1 addition & 1 deletion components/gcp/dataproc/submit_sparksql_job/README.md
Original file line number Diff line number Diff line change
Expand Up @@ -74,7 +74,7 @@ Follow these steps to use the component in a pipeline:
```python
import kfp.components as comp

dataproc_submit_sparksql_job_op = comp.load_component_from_url('https://raw.githubusercontent.com/kubeflow/pipelines/1.0.0-rc.1/components/gcp/dataproc/submit_sparksql_job/component.yaml')
dataproc_submit_sparksql_job_op = comp.load_component_from_url('https://raw.githubusercontent.com/kubeflow/pipelines/1.0.0-rc.2/components/gcp/dataproc/submit_sparksql_job/component.yaml')
help(dataproc_submit_sparksql_job_op)
```

Expand Down
2 changes: 1 addition & 1 deletion components/gcp/dataproc/submit_sparksql_job/component.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -75,7 +75,7 @@ outputs:
type: UI metadata
implementation:
container:
image: gcr.io/ml-pipeline/ml-pipeline-gcp:1.0.0-rc.1
image: gcr.io/ml-pipeline/ml-pipeline-gcp:1.0.0-rc.2
args: [
--ui_metadata_path, {outputPath: MLPipeline UI metadata},
kfp_component.google.dataproc, submit_sparksql_job,
Expand Down
2 changes: 1 addition & 1 deletion components/gcp/dataproc/submit_sparksql_job/sample.ipynb
Original file line number Diff line number Diff line change
Expand Up @@ -78,7 +78,7 @@
"import kfp.components as comp\n",
"\n",
"dataproc_submit_sparksql_job_op = comp.load_component_from_url(\n",
" 'https://raw.githubusercontent.com/kubeflow/pipelines/1.0.0-rc.1/components/gcp/dataproc/submit_sparksql_job/component.yaml')\n",
" 'https://raw.githubusercontent.com/kubeflow/pipelines/1.0.0-rc.2/components/gcp/dataproc/submit_sparksql_job/component.yaml')\n",
"help(dataproc_submit_sparksql_job_op)"
]
},
Expand Down
2 changes: 1 addition & 1 deletion components/gcp/ml_engine/batch_predict/README.md
Original file line number Diff line number Diff line change
Expand Up @@ -87,7 +87,7 @@ KFP_PACKAGE = 'https://storage.googleapis.com/ml-pipeline/release/0.1.14/kfp.tar
import kfp.components as comp

mlengine_batch_predict_op = comp.load_component_from_url(
'https://raw.githubusercontent.com/kubeflow/pipelines/1.0.0-rc.1/components/gcp/ml_engine/batch_predict/component.yaml')
'https://raw.githubusercontent.com/kubeflow/pipelines/1.0.0-rc.2/components/gcp/ml_engine/batch_predict/component.yaml')
help(mlengine_batch_predict_op)
```

Expand Down
2 changes: 1 addition & 1 deletion components/gcp/ml_engine/batch_predict/component.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -69,7 +69,7 @@ outputs:
type: UI metadata
implementation:
container:
image: gcr.io/ml-pipeline/ml-pipeline-gcp:1.0.0-rc.1
image: gcr.io/ml-pipeline/ml-pipeline-gcp:1.0.0-rc.2
args: [
--ui_metadata_path, {outputPath: MLPipeline UI metadata},
kfp_component.google.ml_engine, batch_predict,
Expand Down
2 changes: 1 addition & 1 deletion components/gcp/ml_engine/batch_predict/sample.ipynb
Original file line number Diff line number Diff line change
Expand Up @@ -106,7 +106,7 @@
"import kfp.components as comp\n",
"\n",
"mlengine_batch_predict_op = comp.load_component_from_url(\n",
" 'https://raw.githubusercontent.com/kubeflow/pipelines/1.0.0-rc.1/components/gcp/ml_engine/batch_predict/component.yaml')\n",
" 'https://raw.githubusercontent.com/kubeflow/pipelines/1.0.0-rc.2/components/gcp/ml_engine/batch_predict/component.yaml')\n",
"help(mlengine_batch_predict_op)"
]
},
Expand Down
Loading

0 comments on commit 0796340

Please sign in to comment.