Skip to content

Commit

Permalink
fix(doc): Use Managed NodeGroup for EKS Cluster (#4609)
Browse files Browse the repository at this point in the history
  • Loading branch information
Yao Xiao committed Oct 12, 2020
1 parent e8fb58a commit 1c2a4e8
Showing 1 changed file with 14 additions and 40 deletions.
54 changes: 14 additions & 40 deletions manifests/kustomize/env/aws/README.md
Original file line number Diff line number Diff line change
@@ -1,47 +1,19 @@
# Sample installation

1. Create an EKS cluster and setup kubectl context
1. Create an EKS cluster

Using configuration file to simplify EKS cluster creation process:
```
apiVersion: eksctl.io/v1alpha5
kind: ClusterConfig
metadata:
name: kfworkshop
region: us-west-2
version: '1.17'
# If your region has multiple availability zones, you can specify 3 of them.
availabilityZones: ["us-west-2b", "us-west-2c", "us-west-2d"]
# NodeGroup holds all configuration attributes that are specific to a nodegroup
# You can have several node group in your cluster.
nodeGroups:
- name: cpu-nodegroup
instanceType: m5.xlarge
desiredCapacity: 2
minSize: 0
maxSize: 4
volumeSize: 50
# ssh:
# allow: true
# publicKeyPath: '~/.ssh/id_rsa.pub'
# Example of GPU node group
- name: Tesla-V100
instanceType: p3.8xlarge
# Make sure the availability zone here is one of cluster availability zones.
availabilityZones: ["us-west-2b"]
desiredCapacity: 0
minSize: 0
maxSize: 4
volumeSize: 50
# ssh:
# allow: true
# publicKeyPath: '~/.ssh/id_rsa.pub'
```
Run this command to create EKS cluster
```
eksctl create cluster -f cluster.yaml
eksctl create cluster \
--name AWS-KFP \
--version 1.17 \
--region us-west-2 \
--nodegroup-name linux-nodes \
--node-type m5.xlarge \
--nodes 2 \
--nodes-min 1 \
--nodes-max 4 \
--managed
```

2. Prepare S3
Expand All @@ -67,14 +39,16 @@ Follow this [doc](https://www.kubeflow.org/docs/aws/rds/#deploy-amazon-rds-mysql

```
kubectl apply -k ../../cluster-scoped-resources
# If upper one action got failed, e.x. you used wrong value, try delete, fix and apply again
# kubectl delete -k ../../cluster-scoped-resources
kubectl wait crd/applications.app.k8s.io --for condition=established --timeout=60s
kubectl apply -k ./
# If upper one action got failed, e.x. you used wrong value, try delete, fix and apply again
# kubectl delete -k ./
kubectl wait applications/mypipeline -n kubeflow --for condition=Ready --timeout=1800s
kubectl wait applications/pipeline -n kubeflow --for condition=Ready --timeout=1800s
kubectl port-forward -n kubeflow svc/ml-pipeline-ui 8080:80
```
Expand Down

0 comments on commit 1c2a4e8

Please sign in to comment.