Skip to content
This repository has been archived by the owner on Feb 15, 2024. It is now read-only.

cse-labs/weaveworks

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

7 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Kubernetes in Codespaces

Setup a Kubernetes cluster using k3d running in GitHub Codespaces

License

Overview

This is a template that will setup a Kubernetes developer cluster using k3d in a GitHub Codespace or local Dev Container

We use this for inner-loop Kubernetes development. Note that it is not appropriate for production use but is a great Developer Experience. Feedback calls the approach game-changing - we hope you agree!

For ideas, feature requests, and discussions, please use GitHub discussions so we can collaborate and follow up.

This Codespace is tested with zsh and oh-my-zsh - it "should" work with bash but hasn't been fully tested. For the HoL, please use zsh to avoid any issues.

You can run the dev container locally and you can also connect to the Codespace with a local version of VS Code.

Please experiment and add any issues to the GitHub Discussion. We LOVE PRs!

The motivation for creating and using Codespaces is highlighted by this GitHub Blog Post. "It eliminated the fragility and single-track model of local development environments, but it also gave us a powerful new point of leverage for improving GitHub’s developer experience."

Cory Wilkerson, Senior Director of Engineering at GitHub, recorded a podcast where he shared the GitHub journey to Codespaces

Create your repo

You must have access to Codespaces as an individual or part of a GitHub Team or GitHub Enterprise Cloud

If you are a member of this GitHub organization, you can skip this step and open with Codespaces

Create your repo from this template and add your application code

  • Click the Use this template button
  • Enter your repo details

Open with Codespaces

Note this screen shot is a little out of date with the released version of Codespaces

We LOVE PRs ... :)

  • Click the Code button on your repo
  • Click Open with Codespaces
  • Click New Codespace
  • Choose the 4 core option
    • 2 core isn't enough to run everything well

Create Codespace

Open Workspace

Important!

Another late change - wait until the Codespace is ready before opening the workspace

We LOVE PRs ... :)

  • When prompted, choose Open Workspace

Build and Deploy a k3d Cluster

# build the cluster
make all

Running Codespace

Validate Deployment

Output from make all should resemble this


default      jumpbox                                   1/1   Running   0   25s
default      ngsa-memory                               1/1   Running   0   33s
default      webv                                      1/1   Running   0   31s
logging      fluentbit                                 1/1   Running   0   31s
monitoring   grafana-64f7dbcf96-cfmtd                  1/1   Running   0   32s
monitoring   prometheus-deployment-67cbf97f84-tjxm7    1/1   Running   0   32s

Service endpoints

  • All endpoints are usable in your browser via clicking on the Ports tab
    • Select the open in browser icon on the far right
  • Some popup blockers block the new browser tab
  • If you get a gateway error, just hit refresh - it will clear once the port-forward is ready
# check endpoints
make check

Validate deployment with k9s

  • From the Codespace terminal window, start k9s
    • Type k9s and press enter
    • Press 0 to select all namespaces
    • Wait for all pods to be in the Running state (look for the STATUS column)
    • Use the arrow key to select nsga-memory then press the l key to view logs from the pod
    • To go back, press the esc key
    • Use the arrow key to select jumpbox then press s key to open a shell in the container
      • Hit the ngsa-memory NodePort from within the cluster by executing http ngsa-memory:8080/version
      • Verify 200 status in the response
      • To exit - exit
    • To view other deployed resources - press shift + : followed by the deployment type (e.g. secret, services, deployment, etc).
    • To exit - :q <enter>

k9s

Other interesting endpoints

Open curl.http

curl.http is used in conjuction with the Visual Studio Code REST Client extension.

When you open curl.http, you should see a clickable Send Request text above each of the URLs

REST Client example

Clicking on Send Request should open a new panel in Visual Studio Code with the response from that request like so:

REST Client example response

Jump Box

A jump box pod is created so that you can execute commands in the cluster

  • use the kj alias

    • kubectl exec -it jumpbox -- bash -l
      • note: -l causes a login and processes .profile
      • note: sh -l will work, but the results will not be displayed in the terminal due to a bug
  • use the kje alias

    • kubectl exec -it jumpbox --
  • example

    • run http against the ClusterIP
      • kje http ngsa-memory:8080/version

View Prometheus Dashboard

  • Click on the ports tab of the terminal window

  • Click on the open in browser icon on the Prometheus port (30000)

  • This will open Prometheus in a new browser tab

  • From the Prometheus tab

    • Begin typing NgsaAppDuration_bucket in the Expression search
    • Click Execute
    • This will display the histogram that Grafana uses for the charts

How to expose a Service via a NodePort

Note: you have to make the changes before you run make all

Forward the NodePort on Codespaces

Goal: The steps needed to make the Grafana dashboard accessible via a Codespaces forwarded port

  • Open .devcontainer/devcontainer.json
  • Add port 32000 to forwardPorts
  • Add the port label "32000": { "label": "Grafana" } to portsAttributes
  • Verify Grafana (32000) in the Ports tab of the Terminal view

Set the NodePort on Grafana's deployed service

  • Open deploy/grafana/deployment.yaml
  • In the configs for kind: service, set the ports
    • This example forwards local port 3000 to NodePort 32000
  ports:
    - port: 3000
      targetPort: 3000
      nodePort: 32000

Expose the NodePort from k3d to localhost

  • Open deploy/k3d.yaml
  • Under ports map nodePort 32000 to local port 32000
  • Under - port: 32000:32000 add the nodeFilter, server[0]
    • Explanation
      • There is only one node, which has the Grafana pod
      • That node in the cluster is server node 0 (aka server[0])
      • The node filter indicates which node to send traffic to
      • For multi-node clusters, you have to update
ports:
  - port: 32000:32000
    nodeFilters:
      - server[0]

Launch Grafana Dashboard

  • Grafana login info

    • admin
    • akdc-512
  • Once make all completes successfully

    • Click on the ports tab of the terminal window
    • Click on the open in browser icon on the Grafana port (32000)
    • This will open Grafana in a new browser tab

Codespace Ports

View Grafana Dashboard

  • Click on Home at the top of the page
  • From the dashboards page, click on NGSA

Grafana

Run a load test

# from Codespaces terminal

# run a baseline test (will generate warnings in Grafana)
make test

# run a 60 second load test
make load-test
  • Switch to the Grafana brower tab
  • The test will generate 400 / 404 results
  • The requests metric will go from green to yellow to red as load increases
    • It may skip yellow
  • As the test completes
    • The metric will go back to green (1.0)
    • The request graph will return to normal

Load Test

View Fluent Bit Logs

  • Start k9s from the Codespace terminal
  • Press 0 to show all namespaces
  • Select fluentbit and press enter
  • Press enter again to see the logs
  • Press s to Toggle AutoScroll
  • Press w to Toggle Wrap
  • Review logs that will be sent to Log Analytics when configured
    • See deploy/loganalytics for directions

Build and deploy a local version of WebValidate

  • Switch back to your Codespaces tab
# from Codespaces terminal

# make and deploy a local version of WebV to k8s
make webv

Build and deploy a local version of ngsa-memory

  • Switch back to your Codespaces tab
# from Codespaces terminal

# make and deploy a local version of ngsa-memory to k8s
make app

Next Steps

Makefile is a good place to start exploring

dapr Lab

make sure you are in the root of the repo

Create and run a Web API app with dapr

Create a new dotnet webapi project

mkdir -p dapr-app
cd dapr-app
dotnet new webapi --no-https

Run the app with dapr

dapr run -a myapp -p 5000 -H 3500 -- dotnet run

Check the endpoints

  • open dapr.http
    • click on the dotnet app send request link
    • click on the dapr endpoint send request link

Open Zipkin

  • Click on the Ports tab
    • Open the Zipkin link
    • Click on Run Query
      • Explore the traces generated automatically with dapr

Stop the app by pressing ctl-c

Clean up

cd ..
rm -rf dapr-app

Add dapr SDK to the weather app

Changes to the app have already been made and are detailed below

  • Open .vscode/launch.json
    • Added .NET Core Launch (web) with Dapr configuration
  • Open .vscode/task.json
    • Added daprd-debug and daprd-down tasks
  • Open weather/weather.csproj
    • Added dapr.aspnetcore package reference
  • Open weather/Startup.cs
    • Injected dapr into the services
      • Line 29 services.AddControllers().AddDapr()
    • Added Cloud Events
      • Line 40 app.UseCloudEvents()
  • Open weather/Controllers/WeatherForecastController.cs
    • PostWeatherForecast is a new function for sending pub-sub events
      • Added the Dapr.Topic attribute
      • Got the daprClient via Dependency Injection
      • Published the model to the State Store
    • Get
      • Added the daprClient via Dependency Injection
      • Retrieved the model from the State Store
    • Set a breakpoint on lines 30 and 38

Run the dapr weather app

  • Click on one of the VS Code panels to make sure it has the focus, then Press F5 to run
  • Alternatively, you can use the hamburger menu, then Run and Start Debugging
  • Open dapr.http
    • Send a message via dapr
      • Click on Send Request under post to dapr
      • Click continue when you hit the breakpoint
      • 200 OK
    • Get the model from the State Store
      • Click on Send Request under dapr endpoint
      • Click continue when you hit the breakpoint
      • Verify the value from the POST request appears
    • Change the temperatureC value in POST request and repeat

FAQ

  • Why don't we use helm to deploy Kubernetes manifests?
    • The target audience for this repository is app developers so we chose simplicity for the Developer Experience.
    • In our daily work, we use Helm for deployments and it is installed in the Codespace should you want to use it.
  • Why k3d instead of Kind?
    • We love kind! Most of our code will run unchanged in kind (except the cluster commands)
    • We had to choose one or the other as we don't have the resources to validate both
    • We chose k3d for these main reasons
      • Smaller memory footprint
      • Faster startup time
      • Secure by default
      • Based on K3s which is a certified Kubernetes distro
        • Many customers run K3s on the edge as well as in CI-CD pipelines
      • Rancher provides support - including 24x7 (for a fee)
      • K3s has a vibrant community
      • K3s is a CNCF sandbox project

Engineering Docs

How to file issues and get help

This project uses GitHub Issues to track bugs and feature requests. Please search the existing issues before filing new issues to avoid duplicates. For new issues, file your bug or feature request as a new issue.

For help and questions about using this project, please open a GitHub issue.

Contributing

This project welcomes contributions and suggestions. Most contributions require you to agree to a Contributor License Agreement (CLA) declaring that you have the right to, and actually do, grant us the rights to use your contribution. For details, visit https://cla.opensource.microsoft.com

When you submit a pull request, a CLA bot will automatically determine whether you need to provide a CLA and decorate the PR appropriately (e.g., status check, comment). Simply follow the instructions provided by the bot. You will only need to do this once across all repos using our CLA.

This project has adopted the Microsoft Open Source Code of Conduct. For more information see the Code of Conduct FAQ or contact [email protected] with any additional questions or comments.

Trademarks

This project may contain trademarks or logos for projects, products, or services.

Authorized use of Microsoft trademarks or logos is subject to and must follow Microsoft's Trademark & Brand Guidelines.

Use of Microsoft trademarks or logos in modified versions of this project must not cause confusion or imply Microsoft sponsorship.

Any use of third-party trademarks or logos are subject to those third-party's policies.

About

No description, website, or topics provided.

Resources

License

Code of conduct

Security policy

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published