Anuj Bansal c8f77d2059 Update gitignore, filenames, CF userdata 3 anos atrás
..
.gitignore c8f77d2059 Update gitignore, filenames, CF userdata 3 anos atrás
Pulumi.yaml 9fd1bade66 devops: separate ansible and aws from kubernetes deployments 3 anos atrás
README.md 9fd1bade66 devops: separate ansible and aws from kubernetes deployments 3 anos atrás
index.ts 9fd1bade66 devops: separate ansible and aws from kubernetes deployments 3 anos atrás
package.json 9fd1bade66 devops: separate ansible and aws from kubernetes deployments 3 anos atrás
tsconfig.json 9fd1bade66 devops: separate ansible and aws from kubernetes deployments 3 anos atrás

README.md

Amazon EKS Cluster: Hello World!

This example deploys an EKS Kubernetes cluster with custom ipfs image

Deploying the App

To deploy your infrastructure, follow the below steps.

Prerequisites

  1. Install Pulumi
  2. Install Node.js
  3. Install a package manager for Node.js, such as npm or Yarn.
  4. Configure AWS Credentials
  5. Optional (for debugging): Install kubectl

Steps

After cloning this repo, from this working directory, run these commands:

  1. Install the required Node.js packages:

This installs the dependent packages needed for our Pulumi program.

   $ npm install
  1. Create a new stack, which is an isolated deployment target for this example:

This will initialize the Pulumi program in TypeScript.

   $ pulumi stack init
  1. Set the required configuration variables in Pulumi.<stack>.yaml

    $ pulumi config set-all --plaintext aws:region=us-east-1 --plaintext aws:profile=joystream-user \
    --plaintext wsProviderEndpointURI='wss://rome-rpc-endpoint.joystream.org:9944/' \
    --plaintext isMinikube=true --plaintext isAnonymous=true
    

If running for production use the below mentioned config

   $ pulumi config set-all --plaintext aws:region=us-east-1 --plaintext aws:profile=joystream-user \
    --plaintext wsProviderEndpointURI='wss://rome-rpc-endpoint.joystream.org:9944/' --plaintext isAnonymous=false --plaintext isMinikube=false \
    --plaintext providerId=<ID> --plaintext keyFile=<PATH> --plaintext publicURL=<DOMAIN> --secret passphrase=<PASSPHRASE>

You can also set the storage and the colossusPort config parameters if required

  1. Stand up the EKS cluster:

Running pulumi up -y will deploy the EKS cluster. Note, provisioning a new EKS cluster takes between 10-15 minutes.

  1. Once the stack if up and running, we will modify the Caddy config to get SSL certificate for the load balancer

Modify the config variable isLoadBalancerReady

   $ pulumi config set isLoadBalancerReady true

Run pulumi up -y to update the Caddy config

  1. Access the Kubernetes Cluster using kubectl

To access your new Kubernetes cluster using kubectl, we need to set up the kubeconfig file and download kubectl. We can leverage the Pulumi stack output in the CLI, as Pulumi facilitates exporting these objects for us.

   $ pulumi stack output kubeconfig --show-secrets > kubeconfig
   $ export KUBECONFIG=$PWD/kubeconfig
   $ kubectl get nodes

We can also use the stack output to query the cluster for our newly created Deployment:

   $ kubectl get deployment $(pulumi stack output deploymentName) --namespace=$(pulumi stack output namespaceName)
   $ kubectl get service $(pulumi stack output serviceName) --namespace=$(pulumi stack output namespaceName)

To get logs

   $ kubectl config set-context --current --namespace=$(pulumi stack output namespaceName)
   $ kubectl get pods
   $ kubectl logs <PODNAME> --all-containers

To run a command on a pod

   $ kubectl exec ${POD_NAME} -c ${CONTAINER_NAME} -- ${CMD} ${ARG1}

To see complete pulumi stack output

   $ pulumi stack output

To execute a command

   $ kubectl exec --stdin --tty <PODNAME> -c colossus -- /bin/bash
  1. Once you've finished experimenting, tear down your stack's resources by destroying and removing it:

    $ pulumi destroy --yes
    $ pulumi stack rm --yes