author
int64 658
755k
| date
stringlengths 19
19
| timezone
int64 -46,800
43.2k
| hash
stringlengths 40
40
| message
stringlengths 5
490
| mods
list | language
stringclasses 20
values | license
stringclasses 3
values | repo
stringlengths 5
68
| original_message
stringlengths 12
491
|
---|---|---|---|---|---|---|---|---|---|
596,229 |
27.11.2018 14:52:28
| -3,600 |
e76b816653c1281156a278ccb1b9196964895de6
|
nodepool delete from API
|
[
{
"change_type": "MODIFY",
"old_path": "pkg/api/handlers/terminate_cluster.go",
"new_path": "pkg/api/handlers/terminate_cluster.go",
"diff": "@@ -31,9 +31,11 @@ func (d *terminateCluster) Handle(params operations.TerminateClusterParams, prin\nreturn NewErrorResponse(&operations.TerminateClusterDefault{}, 500, err.Error())\n}\n- _, err = editCluster(kluster, principal, params.Name, func(kluster *v1.Kluster) {\n+ _, err = editCluster(kluster, principal, params.Name, func(kluster *v1.Kluster) error {\nkluster.Status.Phase = models.KlusterPhaseTerminating\nkluster.Status.Message = \"Cluster terminating\"\n+\n+ return nil\n})\nif err != nil {\nreturn NewErrorResponse(&operations.TerminateClusterDefault{}, 500, err.Error())\n"
},
{
"change_type": "MODIFY",
"old_path": "pkg/api/handlers/update_cluster.go",
"new_path": "pkg/api/handlers/update_cluster.go",
"diff": "@@ -20,7 +20,28 @@ type updateCluster struct {\nfunc (d *updateCluster) Handle(params operations.UpdateClusterParams, principal *models.Principal) middleware.Responder {\n- kluster, err := editCluster(d.Kubernikus.Kubernikus().Klusters(d.Namespace), principal, params.Name, func(kluster *v1.Kluster) {\n+ kluster, err := editCluster(d.Kubernikus.Kubernikus().Klusters(d.Namespace), principal, params.Name, func(kluster *v1.Kluster) error {\n+\n+ // find the deleted nodepools\n+ deletedNodePoolNames, err := detectNodePoolChanges(kluster.Spec.NodePools, params.Body.Spec.NodePools)\n+ if err != nil {\n+ return err\n+ }\n+\n+ // clear the status for the deleted nodepools\n+ if len(deletedNodePoolNames) > 0 {\n+ nodePoolInfo := kluster.Status.NodePools\n+ for _, name := range deletedNodePoolNames {\n+ for i, statusNodePool := range nodePoolInfo {\n+ if name == statusNodePool.Name {\n+ nodePoolInfo = append(nodePoolInfo[:i], nodePoolInfo[i+1:]...)\n+ }\n+\n+ }\n+ }\n+ kluster.Status.NodePools = nodePoolInfo\n+ }\n+\nnodePools := params.Body.Spec.NodePools\n//set default image\nfor i, pool := range nodePools {\n@@ -35,6 +56,8 @@ func (d *updateCluster) Handle(params operations.UpdateClusterParams, principal\nif params.Body.Spec.Openstack.SecurityGroupName != \"\" {\nkluster.Spec.Openstack.SecurityGroupName = params.Body.Spec.Openstack.SecurityGroupName\n}\n+\n+ return nil\n})\nif err != nil {\nif apierrors.IsNotFound(err) {\n"
},
{
"change_type": "MODIFY",
"old_path": "pkg/api/handlers/util.go",
"new_path": "pkg/api/handlers/util.go",
"diff": "@@ -2,6 +2,8 @@ package handlers\nimport (\n\"fmt\"\n+ \"github.com/pkg/errors\"\n+\n\"net/http\"\n\"strings\"\n@@ -31,13 +33,16 @@ func qualifiedName(name string, accountId string) string {\nreturn fmt.Sprintf(\"%s-%s\", name, accountId)\n}\n-func editCluster(client kubernikusv1.KlusterInterface, principal *models.Principal, name string, updateFunc func(k *v1.Kluster)) (*v1.Kluster, error) {\n+func editCluster(client kubernikusv1.KlusterInterface, principal *models.Principal, name string, updateFunc func(k *v1.Kluster) error) (*v1.Kluster, error) {\nkluster, err := client.Get(qualifiedName(name, principal.Account), metav1.GetOptions{})\nif err != nil {\nreturn nil, err\n}\n- updateFunc(kluster)\n+ err = updateFunc(kluster)\n+ if err != nil {\n+ return nil, err\n+ }\nupdatedCluster, err := client.Update(kluster)\nif err != nil {\n@@ -62,3 +67,46 @@ func getTracingLogger(request *http.Request) kitlog.Logger {\n}\nreturn logger\n}\n+\n+// detectNodePoolChanges checks for the changes between node pool lists\n+func detectNodePoolChanges(oldNodePools, newNodePools []models.NodePool) (nodePoolsToDelete []string, err error) {\n+\n+ nodePoolsToDelete = make([]string, 0)\n+\n+ // For each old node pool\n+ for _, old := range oldNodePools {\n+ foundInNew := false\n+ // For each new node pool\n+ for _, new := range newNodePools {\n+ // Found in both!\n+ if old.Name == new.Name {\n+ foundInNew = true\n+\n+ err := nodePoolEqualsWithScaling(old, new)\n+ if err != nil {\n+ return nodePoolsToDelete, err\n+ }\n+ }\n+ }\n+ if !foundInNew {\n+ if old.Size != 0 {\n+ return nodePoolsToDelete, errors.New(\"nodepool with size larger than 0 cannot be deleted: \" + old.Name)\n+ } else {\n+ nodePoolsToDelete = append(nodePoolsToDelete, old.Name)\n+ }\n+\n+ }\n+ }\n+\n+ return nodePoolsToDelete, nil\n+}\n+\n+// nodePoolEqualsWithScaling checks whether the node pool is only scaled without any changes\n+func nodePoolEqualsWithScaling(old, new models.NodePool) error {\n+\n+ if old.Flavor != new.Flavor || old.Image != new.Image || old.Name != new.Name {\n+ return errors.New(\"nodepool data cannot be changed except size: \" + old.Name)\n+ }\n+\n+ return nil\n+}\n"
},
{
"change_type": "ADD",
"old_path": null,
"new_path": "pkg/api/handlers/util_test.go",
"diff": "+package handlers\n+\n+import (\n+ \"testing\"\n+\n+ \"github.com/stretchr/testify/assert\"\n+\n+ \"github.com/sapcc/kubernikus/pkg/api/models\"\n+)\n+\n+func TestDetectNodePoolChanges(t *testing.T) {\n+\n+ np := models.NodePool{Name: \"pool\", Size: 0, Image: \"image\", Flavor: \"flavor\"}\n+ npScaled := models.NodePool{Name: \"pool\", Size: 5, Image: \"image\", Flavor: \"flavor\"}\n+ npChanged := models.NodePool{Name: \"pool\", Size: 0, Image: \"image:v2\", Flavor: \"flavor\"}\n+ npNew := models.NodePool{Name: \"pool_new\", Size: 0, Image: \"image\", Flavor: \"flavor\"}\n+\n+ nodePoolListOriginal := []models.NodePool{np, npNew}\n+ nodePoolListScaled := []models.NodePool{npScaled, npNew}\n+ nodePoolListRemoved := []models.NodePool{np}\n+ nodePoolListChanged := []models.NodePool{npChanged}\n+\n+ deleteList, err := detectNodePoolChanges(nodePoolListOriginal, nodePoolListScaled)\n+ assert.Len(t, deleteList, 0)\n+ assert.Nil(t, err)\n+\n+ deleteList, err = detectNodePoolChanges(nodePoolListOriginal, nodePoolListRemoved)\n+ assert.Len(t, deleteList, 1)\n+ assert.Nil(t, err)\n+\n+ deleteList, err = detectNodePoolChanges(nodePoolListOriginal, nodePoolListChanged)\n+ assert.Len(t, deleteList, 0)\n+ assert.NotNil(t, err)\n+\n+}\n+\n+func TestNodePoolEqualsWithScaling(t *testing.T) {\n+\n+ np := models.NodePool{Name: \"pool\", Size: 0, Image: \"image\", Flavor: \"flavor\"}\n+ npScaled := models.NodePool{Name: \"pool\", Size: 5, Image: \"image\", Flavor: \"flavor\"}\n+ npChanged := models.NodePool{Name: \"pool\", Size: 0, Image: \"image:v2\", Flavor: \"flavor\"}\n+\n+ assert.Nil(t, nodePoolEqualsWithScaling(np, npScaled))\n+ assert.NotNil(t, nodePoolEqualsWithScaling(np, npChanged))\n+\n+}\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
nodepool delete from API
|
596,229 |
27.11.2018 15:44:32
| -3,600 |
87e5cbb6706db9cffe56659d4e1595e5cd6a5eca
|
TestClusterUpdate changed: add new nodepool and change only size of previous nodepool
|
[
{
"change_type": "MODIFY",
"old_path": "pkg/api/rest/api_test.go",
"new_path": "pkg/api/rest/api_test.go",
"diff": "@@ -212,6 +212,12 @@ func TestClusterUpdate(t *testing.T) {\nRouterID: \"changed\",\n},\nNodePools: []models.NodePool{\n+ {\n+ Flavor: \"flavour\",\n+ Image: \"image\",\n+ Name: \"poolname\",\n+ Size: 5,\n+ },\n{\nFlavor: \"newflavour\",\nImage: \"newimage\",\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
TestClusterUpdate changed: add new nodepool and change only size of previous nodepool
|
596,226 |
04.12.2018 09:36:35
| -3,600 |
8c9d24db82d6e2660943b66ff7d1fd642cb01404
|
Turn off parallel e2e tests
|
[
{
"change_type": "MODIFY",
"old_path": "ci/task_e2e_tests.yaml",
"new_path": "ci/task_e2e_tests.yaml",
"diff": "@@ -22,6 +22,7 @@ run:\ncd gopath/src/github.com/sapcc/kubernikus\napk add --no-progress --no-cache make git curl\n+ export RUN_PARALLEL=false\nmake test-e2e | tee test.output\nrc=$?\nif [ \"$SENTRY_DSN\" != \"\" ] && [ $rc -ne 0 ] ; then\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
Turn off parallel e2e tests (#343)
|
596,240 |
04.12.2018 10:15:46
| -3,600 |
b764ee9a37c7e6dff4be7ffba9b2c3e94ff3255f
|
bump k8s version to v1.10.11
|
[
{
"change_type": "MODIFY",
"old_path": "charts/kube-master/values.yaml",
"new_path": "charts/kube-master/values.yaml",
"diff": "# Declare variables to be passed into your templates.\nimage:\nrepository: sapcc/hyperkube\n- tag: v1.10.8\n+ tag: v1.10.11\npullPolicy: IfNotPresent\n# Settings for the openstack cloudprovider\n"
},
{
"change_type": "MODIFY",
"old_path": "pkg/templates/node_1.10.go",
"new_path": "pkg/templates/node_1.10.go",
"diff": "@@ -87,7 +87,7 @@ systemd:\n--mount volume=var-log,target=/var/log \\\n--mount volume=etc-machine-id,target=/etc/machine-id \\\n--insecure-options=image\"\n- Environment=\"KUBELET_IMAGE_TAG=v1.10.8\"\n+ Environment=\"KUBELET_IMAGE_TAG=v1.10.11\"\nEnvironment=\"KUBELET_IMAGE_URL=docker://sapcc/hyperkube\"\nEnvironment=\"KUBELET_IMAGE_ARGS=--name=kubelet --exec=/kubelet\"\nExecStartPre=/bin/mkdir -p /etc/kubernetes/manifests\n@@ -168,7 +168,7 @@ systemd:\n--mount volume=lib-modules,target=/lib/modules \\\n--stage1-from-dir=stage1-fly.aci \\\n--insecure-options=image \\\n- docker://sapcc/hyperkube:v1.10.8 \\\n+ docker://sapcc/hyperkube:v1.10.11 \\\n--name kube-proxy \\\n--exec=/hyperkube \\\n-- \\\n"
},
{
"change_type": "MODIFY",
"old_path": "pkg/util/constants.go",
"new_path": "pkg/util/constants.go",
"diff": "@@ -5,5 +5,5 @@ const (\nCA_ISSUER_KUBERNIKUS_IDENTIFIER_1 = \"Kubernikus\"\n// This is the default Kubernetes version that clusters are created in\n- DEFAULT_KUBERNETES_VERSION = \"1.10.8\"\n+ DEFAULT_KUBERNETES_VERSION = \"1.10.11\"\n)\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
bump k8s version to v1.10.11
|
596,226 |
04.12.2018 13:52:41
| -3,600 |
ea7c8d767ccf3fe2474ed8ece97d2f39aa2ddfe5
|
Add oomkill-exporter chart to kubernikus-system
|
[
{
"change_type": "MODIFY",
"old_path": "charts/kubernikus-system/requirements.lock",
"new_path": "charts/kubernikus-system/requirements.lock",
"diff": "@@ -14,5 +14,8 @@ dependencies:\n- name: prometheus-node-exporter\nrepository: https://kubernetes-charts.storage.googleapis.com/\nversion: 0.3.0\n-digest: sha256:70c42cabcaf252a2342bca375a0607a83b514fb2e8f8d44a48a52004630cd6d1\n-generated: 2018-09-17T16:26:54.648440908+02:00\n+- name: oomkill-exporter\n+ repository: file://vendor/oomkill-exporter\n+ version: 0.1.0\n+digest: sha256:93d33808c11f78a7c2d82aa81ebfb86e576c65d186e5d35cc0ed62e34aba6289\n+generated: 2018-12-04T13:39:32.115337967+01:00\n"
},
{
"change_type": "MODIFY",
"old_path": "charts/kubernikus-system/requirements.yaml",
"new_path": "charts/kubernikus-system/requirements.yaml",
"diff": "@@ -14,3 +14,6 @@ dependencies:\n- name: prometheus-node-exporter\nrepository: https://kubernetes-charts.storage.googleapis.com/\nversion: 0.3.0\n+ - name: oomkill-exporter\n+ repository: file://vendor/oomkill-exporter\n+ version: 0.1.0\n"
},
{
"change_type": "ADD",
"old_path": null,
"new_path": "charts/kubernikus-system/vendor/oomkill-exporter/Chart.yaml",
"diff": "+apiVersion: v1\n+description: OOMKill exporter for Kubernetes\n+name: oomkill-exporter\n+version: 0.1.0\n"
},
{
"change_type": "ADD",
"old_path": null,
"new_path": "charts/kubernikus-system/vendor/oomkill-exporter/templates/_helpers.tpl",
"diff": "+{{/* vim: set filetype=mustache: */}}\n+{{/*\n+Expand the name of the chart.\n+*/}}\n+{{- define \"name\" -}}\n+{{- default .Chart.Name .Values.nameOverride | trunc 63 | trimSuffix \"-\" -}}\n+{{- end -}}\n"
},
{
"change_type": "ADD",
"old_path": null,
"new_path": "charts/kubernikus-system/vendor/oomkill-exporter/templates/daemonset.yaml",
"diff": "+apiVersion: apps/v1\n+kind: DaemonSet\n+metadata:\n+ name: {{ template \"name\" . }}\n+ namespace: {{ .Release.Namespace }}\n+ labels:\n+ app: {{ template \"name\" . }}\n+spec:\n+ selector:\n+ matchLabels:\n+ app: {{ template \"name\" . }}\n+ template:\n+ metadata:\n+ labels:\n+ app: {{ template \"name\" . }}\n+ annotations:\n+ prometheus.io/scrape: \"true\"\n+ prometheus.io/port: \"{{ default 9102 .Values.metrics.port }}\"\n+ spec:\n+ containers:\n+ - name: {{ template \"name\" . }}\n+ image: \"{{ .Values.image.repository }}:{{ .Values.image.tag }}\"\n+ imagePullPolicy: {{ .Values.image.pullPolicy }}\n+ args:\n+ - -logtostderr\n+ - -listen-address=:{{ default 9102 .Values.metrics.port }}\n+ - -v=0\n+ securityContext:\n+ privileged: true\n+{{ toYaml .Values.resources | indent 8 }}\n+ env:\n+ - name: DOCKER_HOST\n+ value: \"unix:///var/run/docker.sock\"\n+ volumeMounts:\n+ - name: kmsg\n+ mountPath: /dev/kmsg\n+ readOnly: true\n+ - name: docker\n+ mountPath: /var/run/docker.sock\n+ ports:\n+ - name: metrics\n+ containerPort: {{ default 9102 .Values.metrics.port }}\n+ volumes:\n+ - name: kmsg\n+ hostPath:\n+ path: /dev/kmsg\n+ - name: docker\n+ hostPath:\n+ path: /var/run/docker.sock\n"
},
{
"change_type": "ADD",
"old_path": null,
"new_path": "charts/kubernikus-system/vendor/oomkill-exporter/values.yaml",
"diff": "+image:\n+ repository: sapcc/kubernetes-oomkill-exporter\n+ tag: 0.1.0\n+ pullPolicy: IfNotPresent\n+resources:\n+ limits:\n+ cpu: 100m\n+ memory: 100Mi\n+ requests:\n+ cpu: 20m\n+ memory: 20Mi\n+metrics:\n+ port: \"9102\"\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
Add oomkill-exporter chart to kubernikus-system
|
596,240 |
04.12.2018 14:42:18
| -3,600 |
cd713455d654405bf3007a31a101d53f5ba28566
|
adds more meaningfull charts. fixes broken metric names
|
[
{
"change_type": "MODIFY",
"old_path": "charts/kubernikus-system/values.yaml",
"new_path": "charts/kubernikus-system/values.yaml",
"diff": "@@ -199,11 +199,59 @@ grafana:\nserverDashboardFiles:\nkubernikus.json: |\n{\n+ \"__inputs\": [\n+ {\n+ \"name\": \"DS_PROMETHEUS\",\n+ \"label\": \"Prometheus\",\n+ \"description\": \"\",\n+ \"type\": \"datasource\",\n+ \"pluginId\": \"prometheus\",\n+ \"pluginName\": \"Prometheus\"\n+ }\n+ ],\n+ \"__requires\": [\n+ {\n+ \"type\": \"grafana\",\n+ \"id\": \"grafana\",\n+ \"name\": \"Grafana\",\n+ \"version\": \"5.0.4\"\n+ },\n+ {\n+ \"type\": \"panel\",\n+ \"id\": \"graph\",\n+ \"name\": \"Graph\",\n+ \"version\": \"5.0.0\"\n+ },\n+ {\n+ \"type\": \"panel\",\n+ \"id\": \"natel-discrete-panel\",\n+ \"name\": \"Discrete\",\n+ \"version\": \"0.0.8-pre\"\n+ },\n+ {\n+ \"type\": \"datasource\",\n+ \"id\": \"prometheus\",\n+ \"name\": \"Prometheus\",\n+ \"version\": \"5.0.0\"\n+ },\n+ {\n+ \"type\": \"panel\",\n+ \"id\": \"singlestat\",\n+ \"name\": \"Singlestat\",\n+ \"version\": \"5.0.0\"\n+ },\n+ {\n+ \"type\": \"panel\",\n+ \"id\": \"text\",\n+ \"name\": \"Text\",\n+ \"version\": \"5.0.0\"\n+ }\n+ ],\n\"annotations\": {\n\"list\": [\n{\n\"builtIn\": 1,\n- \"datasource\": \"default\",\n+ \"datasource\": \"${DS_PROMETHEUS}\",\n\"enable\": true,\n\"hide\": true,\n\"iconColor\": \"rgba(0, 211, 255, 1)\",\n@@ -215,7 +263,8 @@ grafana:\n\"editable\": false,\n\"gnetId\": null,\n\"graphTooltip\": 0,\n- \"iteration\": 1519221161829,\n+ \"id\": null,\n+ \"iteration\": 1543925725334,\n\"links\": [],\n\"panels\": [\n{\n@@ -270,7 +319,7 @@ grafana:\n\"rgba(237, 129, 40, 0.89)\",\n\"#d44a3a\"\n],\n- \"datasource\": null,\n+ \"datasource\": \"${DS_PROMETHEUS}\",\n\"decimals\": 0,\n\"format\": \"none\",\n\"gauge\": {\n@@ -355,7 +404,7 @@ grafana:\n\"rgba(237, 129, 40, 0.89)\",\n\"#d44a3a\"\n],\n- \"datasource\": null,\n+ \"datasource\": \"${DS_PROMETHEUS}\",\n\"decimals\": 0,\n\"format\": \"none\",\n\"gauge\": {\n@@ -440,7 +489,7 @@ grafana:\n\"rgba(237, 129, 40, 0.89)\",\n\"#d44a3a\"\n],\n- \"datasource\": null,\n+ \"datasource\": \"${DS_PROMETHEUS}\",\n\"decimals\": 0,\n\"format\": \"none\",\n\"gauge\": {\n@@ -525,7 +574,7 @@ grafana:\n\"rgba(237, 129, 40, 0.89)\",\n\"#d44a3a\"\n],\n- \"datasource\": null,\n+ \"datasource\": \"${DS_PROMETHEUS}\",\n\"decimals\": 0,\n\"format\": \"none\",\n\"gauge\": {\n@@ -639,7 +688,7 @@ grafana:\n\"bars\": false,\n\"dashLength\": 10,\n\"dashes\": false,\n- \"datasource\": null,\n+ \"datasource\": \"${DS_PROMETHEUS}\",\n\"fill\": 3,\n\"gridPos\": {\n\"h\": 4,\n@@ -690,7 +739,7 @@ grafana:\n\"refId\": \"A\"\n},\n{\n- \"expr\": \"quantile_over_time(0.95, probe_duration_seconds{instance=~\\\"https://kubernikus.*\\\"}[24h]) \",\n+ \"expr\": \"quantile_over_time(0.95, probe_duration_seconds{instance=~\\\"https://kubernikus.*\\\",kubernetes_name=\\\"kubernikus-api\\\"}[24h]) \",\n\"format\": \"time_series\",\n\"hide\": false,\n\"interval\": \"\",\n@@ -742,7 +791,7 @@ grafana:\n\"bars\": false,\n\"dashLength\": 10,\n\"dashes\": false,\n- \"datasource\": null,\n+ \"datasource\": \"${DS_PROMETHEUS}\",\n\"fill\": 3,\n\"gridPos\": {\n\"h\": 4,\n@@ -853,8 +902,10 @@ grafana:\n\"text\": \"down\"\n}\n],\n- \"datasource\": null,\n+ \"crosshairColor\": \"#8F070C\",\n+ \"datasource\": \"${DS_PROMETHEUS}\",\n\"display\": \"timeline\",\n+ \"expandFromQueryS\": 86400,\n\"extendLastValue\": true,\n\"gridPos\": {\n\"h\": 4,\n@@ -890,34 +941,34 @@ grafana:\n\"showDistinctCount\": false,\n\"showLegend\": false,\n\"showLegendCounts\": false,\n- \"showLegendNames\": true,\n- \"showLegendPercent\": true,\n+ \"showLegendNames\": false,\n+ \"showLegendPercent\": false,\n\"showLegendTime\": false,\n- \"showLegendValues\": true,\n+ \"showLegendValues\": false,\n+ \"showTimeAxis\": false,\n\"showTransitionCount\": false,\n\"targets\": [\n{\n- \"expr\": \"probe_success{path=\\\"/\\\"}\",\n+ \"expr\": \"probe_success{path=\\\"/\\\", kubernetes_name!~\\\"e2e.*|kubernikus-test.*\\\"} < 1 AND ON (kubernetes_name) label_replace(kubernikus_kluster_status_phase{phase=\\\"Running\\\"}, \\\"kubernetes_name\\\", \\\"$1\\\", \\\"kluster_id\\\", \\\"(.*)\\\")\",\n\"format\": \"time_series\",\n\"hide\": false,\n+ \"instant\": false,\n\"interval\": \"\",\n\"intervalFactor\": 1,\n- \"legendFormat\": \"{{instance}}\",\n+ \"legendFormat\": \"{{kubernetes_name}}\",\n\"refId\": \"A\"\n}\n],\n\"textSize\": 12,\n- \"timeFrom\": \"24h\",\n+ \"textSizeTime\": 12,\n+ \"timeFrom\": \"5m\",\n\"timeShift\": null,\n+ \"timeTextColor\": \"#d8d9da\",\n\"title\": \"\",\n\"transparent\": true,\n\"type\": \"natel-discrete-panel\",\n+ \"units\": \"short\",\n\"valueMaps\": [\n- {\n- \"op\": \"=\",\n- \"text\": \"N/A\",\n- \"value\": \"null\"\n- },\n{\n\"op\": \"=\",\n\"text\": \"up\",\n@@ -930,8 +981,8 @@ grafana:\n},\n{\n\"op\": \"=\",\n- \"text\": \"\",\n- \"value\": \"\"\n+ \"text\": \"N/A\",\n+ \"value\": \"null\"\n}\n],\n\"valueTextColor\": \"#fff\",\n@@ -1006,7 +1057,7 @@ grafana:\n\"rgba(237, 129, 40, 0.89)\",\n\"#299c46\"\n],\n- \"datasource\": null,\n+ \"datasource\": \"${DS_PROMETHEUS}\",\n\"decimals\": 0,\n\"description\": \"Out of CPU in this many Klusters\",\n\"format\": \"none\",\n@@ -1093,7 +1144,7 @@ grafana:\n\"rgba(237, 129, 40, 0.89)\",\n\"#299c46\"\n],\n- \"datasource\": null,\n+ \"datasource\": \"${DS_PROMETHEUS}\",\n\"decimals\": 0,\n\"description\": \"Out of Memory in this many Klusters\",\n\"format\": \"none\",\n@@ -1180,7 +1231,7 @@ grafana:\n\"rgba(237, 129, 40, 0.89)\",\n\"#299c46\"\n],\n- \"datasource\": null,\n+ \"datasource\": \"${DS_PROMETHEUS}\",\n\"decimals\": 0,\n\"description\": \"Out of Podsin this many Klusters\",\n\"format\": \"none\",\n@@ -1297,7 +1348,7 @@ grafana:\n\"rgba(237, 129, 40, 0.89)\",\n\"#299c46\"\n],\n- \"datasource\": null,\n+ \"datasource\": \"${DS_PROMETHEUS}\",\n\"decimals\": 0,\n\"description\": \"Out of Volumes in this many Klusters\",\n\"format\": \"none\",\n@@ -1384,7 +1435,7 @@ grafana:\n\"rgba(237, 129, 40, 0.89)\",\n\"#299c46\"\n],\n- \"datasource\": null,\n+ \"datasource\": \"${DS_PROMETHEUS}\",\n\"decimals\": 0,\n\"description\": \"Out of Volumes in this many Klusters\",\n\"format\": \"none\",\n@@ -1477,7 +1528,7 @@ grafana:\n\"rgba(237, 129, 40, 0.89)\",\n\"#299c46\"\n],\n- \"datasource\": null,\n+ \"datasource\": \"${DS_PROMETHEUS}\",\n\"decimals\": 0,\n\"description\": \"\",\n\"format\": \"none\",\n@@ -1564,7 +1615,7 @@ grafana:\n\"rgba(237, 129, 40, 0.89)\",\n\"#299c46\"\n],\n- \"datasource\": null,\n+ \"datasource\": \"${DS_PROMETHEUS}\",\n\"decimals\": 0,\n\"description\": \"\",\n\"format\": \"none\",\n@@ -1651,7 +1702,7 @@ grafana:\n\"rgba(237, 129, 40, 0.89)\",\n\"#299c46\"\n],\n- \"datasource\": null,\n+ \"datasource\": \"${DS_PROMETHEUS}\",\n\"decimals\": 0,\n\"description\": \"\",\n\"format\": \"none\",\n@@ -1749,7 +1800,7 @@ grafana:\n\"bars\": false,\n\"dashLength\": 10,\n\"dashes\": false,\n- \"datasource\": null,\n+ \"datasource\": \"${DS_PROMETHEUS}\",\n\"decimals\": 0,\n\"fill\": 6,\n\"gridPos\": {\n@@ -1773,7 +1824,7 @@ grafana:\n\"linewidth\": 3,\n\"links\": [],\n\"nullPointMode\": \"connected\",\n- \"percentage\": true,\n+ \"percentage\": false,\n\"pointradius\": 5,\n\"points\": false,\n\"renderer\": \"flot\",\n@@ -1783,7 +1834,7 @@ grafana:\n\"steppedLine\": false,\n\"targets\": [\n{\n- \"expr\": \"sum(irate(node_cpu{mode=~\\\"system|user|iowait\\\"}[5m])) by (instance) / sum(irate(node_cpu{}[5m])) by (instance)\",\n+ \"expr\": \"sum(irate(node_cpu_seconds_total{mode=~\\\"system|user|iowait\\\"}[5m])) by (instance) / sum(irate(node_cpu_seconds_total{}[5m])) by (instance)\",\n\"format\": \"time_series\",\n\"interval\": \"\",\n\"intervalFactor\": 3,\n@@ -1833,7 +1884,7 @@ grafana:\n\"bars\": false,\n\"dashLength\": 10,\n\"dashes\": false,\n- \"datasource\": null,\n+ \"datasource\": \"${DS_PROMETHEUS}\",\n\"decimals\": 0,\n\"fill\": 6,\n\"gridPos\": {\n@@ -1857,7 +1908,7 @@ grafana:\n\"linewidth\": 3,\n\"links\": [],\n\"nullPointMode\": \"connected\",\n- \"percentage\": true,\n+ \"percentage\": false,\n\"pointradius\": 5,\n\"points\": false,\n\"renderer\": \"flot\",\n@@ -1867,7 +1918,7 @@ grafana:\n\"steppedLine\": false,\n\"targets\": [\n{\n- \"expr\": \"sum(irate(node_cpu{mode=~\\\"system|user|iowait\\\"}[5m])) by (instance) / sum(irate(node_cpu{}[5m])) by (instance)\",\n+ \"expr\": \" (node_memory_MemTotal_bytes - node_memory_MemFree_bytes - node_memory_Buffers_bytes - node_memory_Cached_bytes ) / node_memory_MemTotal_bytes\",\n\"format\": \"time_series\",\n\"interval\": \"\",\n\"intervalFactor\": 3,\n@@ -1878,7 +1929,7 @@ grafana:\n\"thresholds\": [],\n\"timeFrom\": null,\n\"timeShift\": null,\n- \"title\": \"CPU\",\n+ \"title\": \"Memory\",\n\"tooltip\": {\n\"shared\": true,\n\"sort\": 0,\n@@ -1917,7 +1968,7 @@ grafana:\n\"bars\": false,\n\"dashLength\": 10,\n\"dashes\": false,\n- \"datasource\": null,\n+ \"datasource\": \"${DS_PROMETHEUS}\",\n\"decimals\": 0,\n\"fill\": 6,\n\"gridPos\": {\n@@ -1941,17 +1992,17 @@ grafana:\n\"linewidth\": 2,\n\"links\": [],\n\"nullPointMode\": \"connected\",\n- \"percentage\": false,\n+ \"percentage\": true,\n\"pointradius\": 5,\n\"points\": false,\n\"renderer\": \"flot\",\n\"seriesOverrides\": [],\n\"spaceLength\": 10,\n- \"stack\": true,\n+ \"stack\": false,\n\"steppedLine\": false,\n\"targets\": [\n{\n- \"expr\": \"sum(label_replace(kubelet_running_pod_count, \\\"node\\\", \\\"$1\\\", \\\"instance\\\", \\\"(.*)\\\")) by (node) / sum(kube_node_status_capacity_pods) by (node) * 100\",\n+ \"expr\": \" count(kube_pod_info) by (node) / sum(kube_node_status_capacity_pods) by (node)\",\n\"format\": \"time_series\",\n\"interval\": \"\",\n\"intervalFactor\": 1,\n@@ -1986,10 +2037,10 @@ grafana:\n\"yaxes\": [\n{\n\"decimals\": null,\n- \"format\": \"percent\",\n+ \"format\": \"percentunit\",\n\"label\": \"\",\n\"logBase\": 1,\n- \"max\": \"400\",\n+ \"max\": \"1\",\n\"min\": \"0.0\",\n\"show\": false\n},\n@@ -2027,7 +2078,7 @@ grafana:\n\"rgba(237, 129, 40, 0.89)\",\n\"#d44a3a\"\n],\n- \"datasource\": null,\n+ \"datasource\": \"${DS_PROMETHEUS}\",\n\"decimals\": 0,\n\"format\": \"percent\",\n\"gauge\": {\n@@ -2114,7 +2165,7 @@ grafana:\n\"rgba(237, 129, 40, 0.89)\",\n\"#d44a3a\"\n],\n- \"datasource\": null,\n+ \"datasource\": \"${DS_PROMETHEUS}\",\n\"decimals\": null,\n\"format\": \"percentunit\",\n\"gauge\": {\n@@ -2201,7 +2252,7 @@ grafana:\n\"rgba(237, 129, 40, 0.89)\",\n\"#d44a3a\"\n],\n- \"datasource\": null,\n+ \"datasource\": \"${DS_PROMETHEUS}\",\n\"format\": \"percentunit\",\n\"gauge\": {\n\"maxValue\": 1,\n@@ -2285,7 +2336,7 @@ grafana:\n\"rgba(237, 129, 40, 0.89)\",\n\"#d44a3a\"\n],\n- \"datasource\": null,\n+ \"datasource\": \"${DS_PROMETHEUS}\",\n\"decimals\": null,\n\"format\": \"percentunit\",\n\"gauge\": {\n@@ -2369,7 +2420,7 @@ grafana:\n\"rgba(237, 129, 40, 0.89)\",\n\"#d44a3a\"\n],\n- \"datasource\": null,\n+ \"datasource\": \"${DS_PROMETHEUS}\",\n\"decimals\": null,\n\"format\": \"percentunit\",\n\"gauge\": {\n@@ -2467,7 +2518,7 @@ grafana:\n\"rgba(237, 129, 40, 0.89)\",\n\"#d44a3a\"\n],\n- \"datasource\": null,\n+ \"datasource\": \"${DS_PROMETHEUS}\",\n\"format\": \"dtdurations\",\n\"gauge\": {\n\"maxValue\": 100,\n@@ -2482,18 +2533,17 @@ grafana:\n\"x\": 0,\n\"y\": 13\n},\n+ \"hideTimeOverride\": true,\n\"id\": 71,\n\"interval\": null,\n\"links\": [],\n\"mappingType\": 1,\n\"mappingTypes\": [\n{\n- \"$$hashKey\": \"object:684\",\n\"name\": \"value to text\",\n\"value\": 1\n},\n{\n- \"$$hashKey\": \"object:685\",\n\"name\": \"range to text\",\n\"value\": 2\n}\n@@ -2521,7 +2571,6 @@ grafana:\n\"tableColumn\": \"\",\n\"targets\": [\n{\n- \"$$hashKey\": \"object:661\",\n\"expr\": \"kubernikus_kluster_boot_duration_sum / kubernikus_kluster_boot_duration_count\",\n\"format\": \"time_series\",\n\"instant\": false,\n@@ -2531,12 +2580,13 @@ grafana:\n}\n],\n\"thresholds\": \"\",\n+ \"timeFrom\": \"1m\",\n+ \"timeShift\": null,\n\"title\": \"Average Kluster Boot Time\",\n\"type\": \"singlestat\",\n\"valueFontSize\": \"80%\",\n\"valueMaps\": [\n{\n- \"$$hashKey\": \"object:687\",\n\"op\": \"=\",\n\"text\": \"N/A\",\n\"value\": \"null\"\n@@ -2549,7 +2599,7 @@ grafana:\n\"bars\": false,\n\"dashLength\": 10,\n\"dashes\": false,\n- \"datasource\": null,\n+ \"datasource\": \"${DS_PROMETHEUS}\",\n\"decimals\": 0,\n\"fill\": 1,\n\"gridPos\": {\n@@ -2584,7 +2634,6 @@ grafana:\n\"steppedLine\": false,\n\"targets\": [\n{\n- \"$$hashKey\": \"object:506\",\n\"expr\": \"sum(label_replace(rate(container_cpu_usage_seconds_total{namespace=\\\"kubernikus\\\",name=~\\\".*(etcd|manager|scheduler|apiserver).*\\\",image!~\\\".*pause.*\\\"}[1h]), \\\"container\\\", \\\"$1\\\", \\\"name\\\", \\\".*(etcd|manager|scheduler|apiserver).*\\\")) by (container) / sum(label_replace(kube_pod_container_resource_requests_cpu_cores{container=~\\\"apiserver|etcd|scheduler|controller-manager\\\"}, \\\"container\\\", \\\"$1\\\", \\\"container\\\", \\\".*(etcd|manager|scheduler|apiserver)\\\")) by (container)\",\n\"format\": \"time_series\",\n\"hide\": false,\n@@ -2594,7 +2643,6 @@ grafana:\n\"refId\": \"B\"\n},\n{\n- \"$$hashKey\": \"object:870\",\n\"expr\": \"\",\n\"format\": \"time_series\",\n\"hide\": false,\n@@ -2622,7 +2670,6 @@ grafana:\n},\n\"yaxes\": [\n{\n- \"$$hashKey\": \"object:531\",\n\"decimals\": 1,\n\"format\": \"percentunit\",\n\"label\": \"\",\n@@ -2632,7 +2679,6 @@ grafana:\n\"show\": true\n},\n{\n- \"$$hashKey\": \"object:532\",\n\"format\": \"short\",\n\"label\": null,\n\"logBase\": 1,\n@@ -2647,7 +2693,7 @@ grafana:\n\"bars\": false,\n\"dashLength\": 10,\n\"dashes\": false,\n- \"datasource\": null,\n+ \"datasource\": \"${DS_PROMETHEUS}\",\n\"decimals\": 0,\n\"fill\": 1,\n\"gridPos\": {\n@@ -2682,7 +2728,6 @@ grafana:\n\"steppedLine\": false,\n\"targets\": [\n{\n- \"$$hashKey\": \"object:1526\",\n\"expr\": \"sum(label_replace(container_memory_usage_bytes{namespace=\\\"kubernikus\\\",name=~\\\".*(etcd|manager|scheduler|apiserver).*\\\",image!~\\\".*pause.*\\\"}, \\\"container\\\", \\\"$1\\\", \\\"name\\\", \\\".*(etcd|manager|scheduler|apiserver).*\\\")) by (container) / sum(label_replace(kube_pod_container_resource_requests_memory_bytes{container=~\\\"apiserver|etcd|scheduler|controller-manager\\\"}, \\\"container\\\", \\\"$1\\\", \\\"container\\\", \\\".*(etcd|manager|scheduler|apiserver)\\\")) by (container)\",\n\"format\": \"time_series\",\n\"hide\": false,\n@@ -2711,7 +2756,6 @@ grafana:\n},\n\"yaxes\": [\n{\n- \"$$hashKey\": \"object:1887\",\n\"decimals\": null,\n\"format\": \"percentunit\",\n\"label\": \"\",\n@@ -2721,7 +2765,6 @@ grafana:\n\"show\": true\n},\n{\n- \"$$hashKey\": \"object:1888\",\n\"format\": \"short\",\n\"label\": null,\n\"logBase\": 1,\n@@ -2736,7 +2779,7 @@ grafana:\n\"bars\": false,\n\"dashLength\": 10,\n\"dashes\": false,\n- \"datasource\": null,\n+ \"datasource\": \"${DS_PROMETHEUS}\",\n\"fill\": 0,\n\"gridPos\": {\n\"h\": 7,\n@@ -2852,7 +2895,7 @@ grafana:\n\"bars\": false,\n\"dashLength\": 10,\n\"dashes\": false,\n- \"datasource\": null,\n+ \"datasource\": \"${DS_PROMETHEUS}\",\n\"fill\": 0,\n\"gridPos\": {\n\"h\": 7,\n@@ -2968,7 +3011,7 @@ grafana:\n\"bars\": false,\n\"dashLength\": 10,\n\"dashes\": false,\n- \"datasource\": null,\n+ \"datasource\": \"${DS_PROMETHEUS}\",\n\"fill\": 0,\n\"gridPos\": {\n\"h\": 7,\n@@ -3007,7 +3050,6 @@ grafana:\n\"steppedLine\": false,\n\"targets\": [\n{\n- \"$$hashKey\": \"object:444\",\n\"expr\": \"sum(go_goroutines{app=\\\"kubernikus\\\"})\",\n\"format\": \"time_series\",\n\"intervalFactor\": 1,\n@@ -3015,7 +3057,6 @@ grafana:\n\"refId\": \"D\"\n},\n{\n- \"$$hashKey\": \"object:445\",\n\"expr\": \"go_rout\",\n\"format\": \"time_series\",\n\"intervalFactor\": 1,\n@@ -3042,7 +3083,6 @@ grafana:\n},\n\"yaxes\": [\n{\n- \"$$hashKey\": \"object:517\",\n\"decimals\": null,\n\"format\": \"short\",\n\"label\": null,\n@@ -3052,7 +3092,6 @@ grafana:\n\"show\": true\n},\n{\n- \"$$hashKey\": \"object:518\",\n\"format\": \"short\",\n\"label\": null,\n\"logBase\": 1,\n@@ -3071,11 +3110,8 @@ grafana:\n\"list\": [\n{\n\"allValue\": null,\n- \"current\": {\n- \"text\": \"1.12.6\",\n- \"value\": \"1.12.6\"\n- },\n- \"datasource\": \"Prometheus\",\n+ \"current\": {},\n+ \"datasource\": \"${DS_PROMETHEUS}\",\n\"hide\": 2,\n\"includeAll\": false,\n\"label\": null,\n@@ -3094,11 +3130,8 @@ grafana:\n},\n{\n\"allValue\": null,\n- \"current\": {\n- \"text\": \"1.7.5\",\n- \"value\": \"1.7.5\"\n- },\n- \"datasource\": \"Prometheus\",\n+ \"current\": {},\n+ \"datasource\": \"${DS_PROMETHEUS}\",\n\"hide\": 2,\n\"includeAll\": false,\n\"label\": null,\n@@ -3117,11 +3150,8 @@ grafana:\n},\n{\n\"allValue\": null,\n- \"current\": {\n- \"text\": \"1576.4.0\",\n- \"value\": \"1576.4.0\"\n- },\n- \"datasource\": \"Prometheus\",\n+ \"current\": {},\n+ \"datasource\": \"${DS_PROMETHEUS}\",\n\"hide\": 2,\n\"includeAll\": false,\n\"label\": null,\n@@ -3140,11 +3170,8 @@ grafana:\n},\n{\n\"allValue\": null,\n- \"current\": {\n- \"text\": \"4.13.16\",\n- \"value\": \"4.13.16\"\n- },\n- \"datasource\": \"Prometheus\",\n+ \"current\": {},\n+ \"datasource\": \"${DS_PROMETHEUS}\",\n\"hide\": 2,\n\"includeAll\": false,\n\"label\": null,\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
adds more meaningfull charts. fixes broken metric names
|
596,226 |
04.12.2018 14:45:13
| -3,600 |
07ae8b6c5957bb56238dcd2aa2250c23d34ca49b
|
Fix e2e test imports
|
[
{
"change_type": "MODIFY",
"old_path": "test/e2e/main_test.go",
"new_path": "test/e2e/main_test.go",
"diff": "@@ -10,7 +10,7 @@ import (\n\"github.com/gophercloud/gophercloud/openstack/identity/v3/tokens\"\n\"github.com/stretchr/testify/require\"\n- \"github.com/sapcc/kubernikus/pkg/util\"\n+ \"github.com/sapcc/kubernikus/pkg/util/generator\"\n\"github.com/sapcc/kubernikus/test/e2e/framework\"\n)\n@@ -65,7 +65,7 @@ func TestMain(m *testing.M) {\n}\nfunc TestRunner(t *testing.T) {\n- klusterName := util.SimpleNameGenerator.GenerateName(\"e2e-\")\n+ klusterName := generator.SimpleNameGenerator.GenerateName(\"e2e-\")\nif kluster != nil && *kluster != \"\" {\nklusterName = *kluster\n"
},
{
"change_type": "MODIFY",
"old_path": "test/e2e/network_test.go",
"new_path": "test/e2e/network_test.go",
"diff": "@@ -15,7 +15,7 @@ import (\n\"github.com/stretchr/testify/assert\"\n\"github.com/stretchr/testify/require\"\n- \"github.com/sapcc/kubernikus/pkg/util\"\n+ \"github.com/sapcc/kubernikus/pkg/util/generator\"\n\"github.com/sapcc/kubernikus/test/e2e/framework\"\n)\n@@ -43,7 +43,7 @@ type NetworkTests struct {\nfunc (n *NetworkTests) Run(t *testing.T) {\nrunParallel(t)\n- n.Namespace = util.SimpleNameGenerator.GenerateName(\"e2e-network-\")\n+ n.Namespace = generator.SimpleNameGenerator.GenerateName(\"e2e-network-\")\nvar err error\nn.Nodes, err = n.Kubernetes.ClientSet.CoreV1().Nodes().List(meta_v1.ListOptions{})\n"
},
{
"change_type": "MODIFY",
"old_path": "test/e2e/volume_test.go",
"new_path": "test/e2e/volume_test.go",
"diff": "@@ -12,7 +12,7 @@ import (\nmeta_v1 \"k8s.io/apimachinery/pkg/apis/meta/v1\"\n\"k8s.io/apimachinery/pkg/labels\"\n- \"github.com/sapcc/kubernikus/pkg/util\"\n+ \"github.com/sapcc/kubernikus/pkg/util/generator\"\n\"github.com/sapcc/kubernikus/test/e2e/framework\"\n)\n@@ -30,7 +30,7 @@ type VolumeTests struct {\nfunc (v *VolumeTests) Run(t *testing.T) {\nrunParallel(t)\n- v.Namespace = util.SimpleNameGenerator.GenerateName(\"e2e-volumes-\")\n+ v.Namespace = generator.SimpleNameGenerator.GenerateName(\"e2e-volumes-\")\nvar err error\nv.Nodes, err = v.Kubernetes.ClientSet.CoreV1().Nodes().List(meta_v1.ListOptions{})\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
Fix e2e test imports
|
596,240 |
04.12.2018 17:19:16
| -3,600 |
e0942467145a14594e8258861291fe45484a64a4
|
emergency fix. try2
|
[
{
"change_type": "MODIFY",
"old_path": "pkg/client/openstack/kluster/client.go",
"new_path": "pkg/client/openstack/kluster/client.go",
"diff": "@@ -131,27 +131,34 @@ func (c *klusterClient) DeleteNode(id string) (err error) {\n}\nfunc (c *klusterClient) ListNodes(pool *models.NodePool) ([]Node, error) {\n+ var unfilteredNodes []Node\n+ var filteredNodes []Node\n+ var err error\n+\nprefix := fmt.Sprintf(\"%v-%v-\", c.Kluster.Spec.Name, pool.Name)\n- allNodes, err := servers.List(c.ComputeClient, servers.ListOpts{Name: prefix}).AllPages()\n+ err = servers.List(c.ComputeClient, servers.ListOpts{Name: prefix}).EachPage(func(page pagination.Page) (bool, error) {\n+ unfilteredNodes, err = ExtractServers(page)\nif err != nil {\n- return nil, err\n+ return false, err\n}\n- unfilteredNodes, err := ExtractServers(allNodes)\n+ return true, nil\n+ })\nif err != nil {\nreturn nil, err\n}\n+\n//filter nodeList https://github.com/golang/go/wiki/SliceTricks#filtering-without-allocating\n//we only keep nodes whose where the name length is matched the expected length of a name for this pool\n//otherwise we would be returning nodes from other nodepools here if the current pool name is a prefix of other pools\n- nodes := unfilteredNodes[:0]\n+ filteredNodes = unfilteredNodes[:0]\nfor _, node := range unfilteredNodes {\nif len(node.GetName()) == len(prefix)+generator.RandomLength {\n- nodes = append(nodes, node)\n+ filteredNodes = append(filteredNodes, node)\n}\n}\n- return nodes, nil\n+ return filteredNodes, nil\n}\nfunc (c *klusterClient) SetSecurityGroup(nodeID string) (err error) {\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
emergency fix. try2
|
596,240 |
04.12.2018 17:51:22
| -3,600 |
97d952eced5e99ff44839fad8ebbba7cab79bc7c
|
fixes datasource trouble
|
[
{
"change_type": "MODIFY",
"old_path": "charts/kubernikus-system/values.yaml",
"new_path": "charts/kubernikus-system/values.yaml",
"diff": "@@ -201,7 +201,7 @@ grafana:\n{\n\"__inputs\": [\n{\n- \"name\": \"DS_PROMETHEUS\",\n+ \"name\": \"default\",\n\"label\": \"Prometheus\",\n\"description\": \"\",\n\"type\": \"datasource\",\n@@ -251,7 +251,7 @@ grafana:\n\"list\": [\n{\n\"builtIn\": 1,\n- \"datasource\": \"${DS_PROMETHEUS}\",\n+ \"datasource\": \"default\",\n\"enable\": true,\n\"hide\": true,\n\"iconColor\": \"rgba(0, 211, 255, 1)\",\n@@ -319,7 +319,7 @@ grafana:\n\"rgba(237, 129, 40, 0.89)\",\n\"#d44a3a\"\n],\n- \"datasource\": \"${DS_PROMETHEUS}\",\n+ \"datasource\": null,\n\"decimals\": 0,\n\"format\": \"none\",\n\"gauge\": {\n@@ -404,7 +404,7 @@ grafana:\n\"rgba(237, 129, 40, 0.89)\",\n\"#d44a3a\"\n],\n- \"datasource\": \"${DS_PROMETHEUS}\",\n+ \"datasource\": null,\n\"decimals\": 0,\n\"format\": \"none\",\n\"gauge\": {\n@@ -489,7 +489,7 @@ grafana:\n\"rgba(237, 129, 40, 0.89)\",\n\"#d44a3a\"\n],\n- \"datasource\": \"${DS_PROMETHEUS}\",\n+ \"datasource\": null,\n\"decimals\": 0,\n\"format\": \"none\",\n\"gauge\": {\n@@ -574,7 +574,7 @@ grafana:\n\"rgba(237, 129, 40, 0.89)\",\n\"#d44a3a\"\n],\n- \"datasource\": \"${DS_PROMETHEUS}\",\n+ \"datasource\": null,\n\"decimals\": 0,\n\"format\": \"none\",\n\"gauge\": {\n@@ -688,7 +688,7 @@ grafana:\n\"bars\": false,\n\"dashLength\": 10,\n\"dashes\": false,\n- \"datasource\": \"${DS_PROMETHEUS}\",\n+ \"datasource\": null,\n\"fill\": 3,\n\"gridPos\": {\n\"h\": 4,\n@@ -791,7 +791,7 @@ grafana:\n\"bars\": false,\n\"dashLength\": 10,\n\"dashes\": false,\n- \"datasource\": \"${DS_PROMETHEUS}\",\n+ \"datasource\": null,\n\"fill\": 3,\n\"gridPos\": {\n\"h\": 4,\n@@ -903,7 +903,7 @@ grafana:\n}\n],\n\"crosshairColor\": \"#8F070C\",\n- \"datasource\": \"${DS_PROMETHEUS}\",\n+ \"datasource\": null,\n\"display\": \"timeline\",\n\"expandFromQueryS\": 86400,\n\"extendLastValue\": true,\n@@ -1057,7 +1057,7 @@ grafana:\n\"rgba(237, 129, 40, 0.89)\",\n\"#299c46\"\n],\n- \"datasource\": \"${DS_PROMETHEUS}\",\n+ \"datasource\": null,\n\"decimals\": 0,\n\"description\": \"Out of CPU in this many Klusters\",\n\"format\": \"none\",\n@@ -1144,7 +1144,7 @@ grafana:\n\"rgba(237, 129, 40, 0.89)\",\n\"#299c46\"\n],\n- \"datasource\": \"${DS_PROMETHEUS}\",\n+ \"datasource\": null,\n\"decimals\": 0,\n\"description\": \"Out of Memory in this many Klusters\",\n\"format\": \"none\",\n@@ -1231,7 +1231,7 @@ grafana:\n\"rgba(237, 129, 40, 0.89)\",\n\"#299c46\"\n],\n- \"datasource\": \"${DS_PROMETHEUS}\",\n+ \"datasource\": null,\n\"decimals\": 0,\n\"description\": \"Out of Podsin this many Klusters\",\n\"format\": \"none\",\n@@ -1348,7 +1348,7 @@ grafana:\n\"rgba(237, 129, 40, 0.89)\",\n\"#299c46\"\n],\n- \"datasource\": \"${DS_PROMETHEUS}\",\n+ \"datasource\": null,\n\"decimals\": 0,\n\"description\": \"Out of Volumes in this many Klusters\",\n\"format\": \"none\",\n@@ -1435,7 +1435,7 @@ grafana:\n\"rgba(237, 129, 40, 0.89)\",\n\"#299c46\"\n],\n- \"datasource\": \"${DS_PROMETHEUS}\",\n+ \"datasource\": null,\n\"decimals\": 0,\n\"description\": \"Out of Volumes in this many Klusters\",\n\"format\": \"none\",\n@@ -1528,7 +1528,7 @@ grafana:\n\"rgba(237, 129, 40, 0.89)\",\n\"#299c46\"\n],\n- \"datasource\": \"${DS_PROMETHEUS}\",\n+ \"datasource\": null,\n\"decimals\": 0,\n\"description\": \"\",\n\"format\": \"none\",\n@@ -1615,7 +1615,7 @@ grafana:\n\"rgba(237, 129, 40, 0.89)\",\n\"#299c46\"\n],\n- \"datasource\": \"${DS_PROMETHEUS}\",\n+ \"datasource\": null,\n\"decimals\": 0,\n\"description\": \"\",\n\"format\": \"none\",\n@@ -1702,7 +1702,7 @@ grafana:\n\"rgba(237, 129, 40, 0.89)\",\n\"#299c46\"\n],\n- \"datasource\": \"${DS_PROMETHEUS}\",\n+ \"datasource\": null,\n\"decimals\": 0,\n\"description\": \"\",\n\"format\": \"none\",\n@@ -1800,7 +1800,7 @@ grafana:\n\"bars\": false,\n\"dashLength\": 10,\n\"dashes\": false,\n- \"datasource\": \"${DS_PROMETHEUS}\",\n+ \"datasource\": null,\n\"decimals\": 0,\n\"fill\": 6,\n\"gridPos\": {\n@@ -1884,7 +1884,7 @@ grafana:\n\"bars\": false,\n\"dashLength\": 10,\n\"dashes\": false,\n- \"datasource\": \"${DS_PROMETHEUS}\",\n+ \"datasource\": null,\n\"decimals\": 0,\n\"fill\": 6,\n\"gridPos\": {\n@@ -1968,7 +1968,7 @@ grafana:\n\"bars\": false,\n\"dashLength\": 10,\n\"dashes\": false,\n- \"datasource\": \"${DS_PROMETHEUS}\",\n+ \"datasource\": null,\n\"decimals\": 0,\n\"fill\": 6,\n\"gridPos\": {\n@@ -2078,7 +2078,7 @@ grafana:\n\"rgba(237, 129, 40, 0.89)\",\n\"#d44a3a\"\n],\n- \"datasource\": \"${DS_PROMETHEUS}\",\n+ \"datasource\": null,\n\"decimals\": 0,\n\"format\": \"percent\",\n\"gauge\": {\n@@ -2165,7 +2165,7 @@ grafana:\n\"rgba(237, 129, 40, 0.89)\",\n\"#d44a3a\"\n],\n- \"datasource\": \"${DS_PROMETHEUS}\",\n+ \"datasource\": null,\n\"decimals\": null,\n\"format\": \"percentunit\",\n\"gauge\": {\n@@ -2252,7 +2252,7 @@ grafana:\n\"rgba(237, 129, 40, 0.89)\",\n\"#d44a3a\"\n],\n- \"datasource\": \"${DS_PROMETHEUS}\",\n+ \"datasource\": null,\n\"format\": \"percentunit\",\n\"gauge\": {\n\"maxValue\": 1,\n@@ -2336,7 +2336,7 @@ grafana:\n\"rgba(237, 129, 40, 0.89)\",\n\"#d44a3a\"\n],\n- \"datasource\": \"${DS_PROMETHEUS}\",\n+ \"datasource\": null,\n\"decimals\": null,\n\"format\": \"percentunit\",\n\"gauge\": {\n@@ -2420,7 +2420,7 @@ grafana:\n\"rgba(237, 129, 40, 0.89)\",\n\"#d44a3a\"\n],\n- \"datasource\": \"${DS_PROMETHEUS}\",\n+ \"datasource\": null,\n\"decimals\": null,\n\"format\": \"percentunit\",\n\"gauge\": {\n@@ -2518,7 +2518,7 @@ grafana:\n\"rgba(237, 129, 40, 0.89)\",\n\"#d44a3a\"\n],\n- \"datasource\": \"${DS_PROMETHEUS}\",\n+ \"datasource\": null,\n\"format\": \"dtdurations\",\n\"gauge\": {\n\"maxValue\": 100,\n@@ -2599,7 +2599,7 @@ grafana:\n\"bars\": false,\n\"dashLength\": 10,\n\"dashes\": false,\n- \"datasource\": \"${DS_PROMETHEUS}\",\n+ \"datasource\": null,\n\"decimals\": 0,\n\"fill\": 1,\n\"gridPos\": {\n@@ -2693,7 +2693,7 @@ grafana:\n\"bars\": false,\n\"dashLength\": 10,\n\"dashes\": false,\n- \"datasource\": \"${DS_PROMETHEUS}\",\n+ \"datasource\": null,\n\"decimals\": 0,\n\"fill\": 1,\n\"gridPos\": {\n@@ -2779,7 +2779,7 @@ grafana:\n\"bars\": false,\n\"dashLength\": 10,\n\"dashes\": false,\n- \"datasource\": \"${DS_PROMETHEUS}\",\n+ \"datasource\": null,\n\"fill\": 0,\n\"gridPos\": {\n\"h\": 7,\n@@ -2895,7 +2895,7 @@ grafana:\n\"bars\": false,\n\"dashLength\": 10,\n\"dashes\": false,\n- \"datasource\": \"${DS_PROMETHEUS}\",\n+ \"datasource\": null,\n\"fill\": 0,\n\"gridPos\": {\n\"h\": 7,\n@@ -3011,7 +3011,7 @@ grafana:\n\"bars\": false,\n\"dashLength\": 10,\n\"dashes\": false,\n- \"datasource\": \"${DS_PROMETHEUS}\",\n+ \"datasource\": null,\n\"fill\": 0,\n\"gridPos\": {\n\"h\": 7,\n@@ -3111,7 +3111,7 @@ grafana:\n{\n\"allValue\": null,\n\"current\": {},\n- \"datasource\": \"${DS_PROMETHEUS}\",\n+ \"datasource\": null,\n\"hide\": 2,\n\"includeAll\": false,\n\"label\": null,\n@@ -3131,7 +3131,7 @@ grafana:\n{\n\"allValue\": null,\n\"current\": {},\n- \"datasource\": \"${DS_PROMETHEUS}\",\n+ \"datasource\": null,\n\"hide\": 2,\n\"includeAll\": false,\n\"label\": null,\n@@ -3151,7 +3151,7 @@ grafana:\n{\n\"allValue\": null,\n\"current\": {},\n- \"datasource\": \"${DS_PROMETHEUS}\",\n+ \"datasource\": null,\n\"hide\": 2,\n\"includeAll\": false,\n\"label\": null,\n@@ -3171,7 +3171,7 @@ grafana:\n{\n\"allValue\": null,\n\"current\": {},\n- \"datasource\": \"${DS_PROMETHEUS}\",\n+ \"datasource\": null,\n\"hide\": 2,\n\"includeAll\": false,\n\"label\": null,\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
fixes datasource trouble
|
596,226 |
05.12.2018 14:33:07
| -3,600 |
4b1e913d8eb24c034e455bd5aab55c53d0d1057e
|
Run etcd backup e2e test at the very end
|
[
{
"change_type": "MODIFY",
"old_path": "test/e2e/main_test.go",
"new_path": "test/e2e/main_test.go",
"diff": "@@ -175,6 +175,12 @@ func TestRunner(t *testing.T) {\n}\nt.Run(\"Smoke\", func(t *testing.T) {\n+ volumeTests := &VolumeTests{Kubernetes: kubernetes}\n+ t.Run(\"Volumes\", volumeTests.Run)\n+\n+ networkTests := &NetworkTests{Kubernetes: kubernetes}\n+ t.Run(\"Network\", networkTests.Run)\n+\nif os.Getenv(\"CP_KUBERNIKUS_URL\") != \"\" {\nkubernetesControlPlane, err := framework.NewKubernetesFramework(kubernikusControlPlane, os.Getenv(\"CP_KLUSTER\"))\nrequire.NoError(t, err, \"Must be able to create a control plane kubernetes client\")\n@@ -192,12 +198,6 @@ func TestRunner(t *testing.T) {\n}\nt.Run(\"EtcdBackupTests\", etcdBackupTests.Run)\n}\n-\n- volumeTests := &VolumeTests{Kubernetes: kubernetes}\n- t.Run(\"Volumes\", volumeTests.Run)\n-\n- networkTests := &NetworkTests{Kubernetes: kubernetes}\n- t.Run(\"Network\", networkTests.Run)\n})\n}\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
Run etcd backup e2e test at the very end
|
596,226 |
07.12.2018 14:37:30
| -3,600 |
1f19a05cc01365b616039fcdb94a7b5389d2c709
|
Bump to oomkill-exporter 0.2.0, use rolling update in daemonset
|
[
{
"change_type": "MODIFY",
"old_path": "charts/kubernikus-system/vendor/oomkill-exporter/templates/daemonset.yaml",
"new_path": "charts/kubernikus-system/vendor/oomkill-exporter/templates/daemonset.yaml",
"diff": "@@ -6,6 +6,8 @@ metadata:\nlabels:\napp: {{ template \"name\" . }}\nspec:\n+ updateStrategy:\n+ type: RollingUpdate\nselector:\nmatchLabels:\napp: {{ template \"name\" . }}\n"
},
{
"change_type": "MODIFY",
"old_path": "charts/kubernikus-system/vendor/oomkill-exporter/values.yaml",
"new_path": "charts/kubernikus-system/vendor/oomkill-exporter/values.yaml",
"diff": "image:\nrepository: sapcc/kubernetes-oomkill-exporter\n- tag: 0.1.0\n+ tag: 0.2.0\npullPolicy: IfNotPresent\nresources:\nlimits:\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
Bump to oomkill-exporter 0.2.0, use rolling update in daemonset
|
596,226 |
27.11.2018 14:42:14
| -3,600 |
9a0549cd7ad0ef658b66c0f44bed0778d9b8d060
|
Added eventexporter chart
|
[
{
"change_type": "MODIFY",
"old_path": "charts/kubernikus-system/requirements.lock",
"new_path": "charts/kubernikus-system/requirements.lock",
"diff": "@@ -17,5 +17,8 @@ dependencies:\n- name: oomkill-exporter\nrepository: file://vendor/oomkill-exporter\nversion: 0.1.0\n-digest: sha256:93d33808c11f78a7c2d82aa81ebfb86e576c65d186e5d35cc0ed62e34aba6289\n-generated: 2018-12-04T13:39:32.115337967+01:00\n+- name: eventexporter\n+ repository: file://vendor/eventexporter\n+ version: 0.1.0\n+digest: sha256:767e204ee157475630116592bcd2dbe45983b6ab763d9310c73568b4ba101c5f\n+generated: 2018-12-05T16:30:27.997611663+01:00\n"
},
{
"change_type": "MODIFY",
"old_path": "charts/kubernikus-system/requirements.yaml",
"new_path": "charts/kubernikus-system/requirements.yaml",
"diff": "@@ -17,3 +17,6 @@ dependencies:\n- name: oomkill-exporter\nrepository: file://vendor/oomkill-exporter\nversion: 0.1.0\n+ - name: eventexporter\n+ repository: file://vendor/eventexporter\n+ version: 0.1.0\n"
},
{
"change_type": "ADD",
"old_path": null,
"new_path": "charts/kubernikus-system/vendor/eventexporter/Chart.yaml",
"diff": "+apiVersion: v1\n+description: Eventexporter for Kubernetes\n+name: eventexporter\n+version: 0.1.0\n"
},
{
"change_type": "ADD",
"old_path": null,
"new_path": "charts/kubernikus-system/vendor/eventexporter/templates/_helpers.tpl",
"diff": "+{{/* vim: set filetype=mustache: */}}\n+{{/*\n+Expand the name of the chart.\n+*/}}\n+{{- define \"name\" -}}\n+{{- default .Chart.Name .Values.nameOverride | trunc 63 | trimSuffix \"-\" -}}\n+{{- end -}}\n"
},
{
"change_type": "ADD",
"old_path": null,
"new_path": "charts/kubernikus-system/vendor/eventexporter/templates/config.yaml",
"diff": "+apiVersion: v1\n+kind: ServiceAccount\n+metadata:\n+ name: {{ template \"name\" . }}\n+ namespace: {{ .Release.Namespace }}\n+---\n+apiVersion: rbac.authorization.k8s.io/v1beta1\n+kind: ClusterRole\n+metadata:\n+ name: {{ template \"name\" . }}\n+rules:\n+- apiGroups: [\"\"]\n+ resources: [\"events\"]\n+ verbs: [\"get\", \"watch\", \"list\"]\n+---\n+apiVersion: rbac.authorization.k8s.io/v1beta1\n+kind: ClusterRoleBinding\n+metadata:\n+ name: {{ template \"name\" . }}\n+roleRef:\n+ apiGroup: rbac.authorization.k8s.io\n+ kind: ClusterRole\n+ name: {{ template \"name\" . }}\n+subjects:\n+- kind: ServiceAccount\n+ name: {{ template \"name\" . }}\n+ namespace: {{ .Release.Namespace }}\n+---\n+apiVersion: v1\n+kind: ConfigMap\n+metadata:\n+ name: {{ template \"name\" . }}\n+ namespace: {{ .Release.Namespace }}\n+data:\n+ config.yaml: |-\n+ metrics:\n+ - name: volume_mount_error\n+ event_matcher:\n+ - key: InvolvedObject.Kind\n+ expr: Pod\n+ - key: Message\n+ expr: Unable to mount volumes for pod.*\n+ - key: Reason\n+ expr: FailedMount\n+ - key: Type\n+ expr: Warning\n+ labels:\n+ node: Source.Host\n"
},
{
"change_type": "ADD",
"old_path": null,
"new_path": "charts/kubernikus-system/vendor/eventexporter/templates/deployment.yaml",
"diff": "+apiVersion: apps/v1beta2\n+kind: Deployment\n+metadata:\n+ name: {{ template \"name\" . }}\n+ namespace: {{ .Release.Namespace }}\n+ labels:\n+ app: {{ template \"name\" . }}\n+spec:\n+ replicas: 1\n+ selector:\n+ matchLabels:\n+ app: {{ template \"name\" . }}\n+ template:\n+ metadata:\n+ labels:\n+ app: {{ template \"name\" . }}\n+ annotations:\n+ prometheus.io/scrape: \"true\"\n+ prometheus.io/port: \"9102\"\n+ spec:\n+ containers:\n+ - name: {{ template \"name\" . }}\n+ image: \"{{ .Values.image.repository }}:{{ .Values.image.tag }}\"\n+ imagePullPolicy: {{ .Values.image.pullPolicy }}\n+ volumeMounts:\n+ - name: config-volume\n+ mountPath: /etc/eventexporter\n+ ports:\n+ - name: metrics\n+ containerPort: \"9102\"\n+ serviceAccount: {{ template \"name\" . }}\n+ volumes:\n+ - name: config-volume\n+ configMap:\n+ name: {{ template \"name\" . }}\n"
},
{
"change_type": "ADD",
"old_path": null,
"new_path": "charts/kubernikus-system/vendor/eventexporter/values.yaml",
"diff": "+image:\n+ repository: sapcc/kubernetes-eventexporter\n+ tag: 0.1.0\n+ pullPolicy: IfNotPresent\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
Added eventexporter chart
|
596,226 |
27.11.2018 17:36:34
| -3,600 |
5d8560df269870317798456f12bf91fb715d0ca9
|
Move config to values.yaml, resources to separate files
|
[
{
"change_type": "ADD",
"old_path": null,
"new_path": "charts/kubernikus-system/vendor/eventexporter/templates/clusterrole.yaml",
"diff": "+apiVersion: rbac.authorization.k8s.io/v1beta1\n+kind: ClusterRole\n+metadata:\n+ name: {{ template \"name\" . }}\n+rules:\n+- apiGroups: [\"\"]\n+ resources: [\"events\"]\n+ verbs: [\"get\", \"watch\", \"list\"]\n"
},
{
"change_type": "ADD",
"old_path": null,
"new_path": "charts/kubernikus-system/vendor/eventexporter/templates/clusterrolebinding.yaml",
"diff": "+apiVersion: rbac.authorization.k8s.io/v1beta1\n+kind: ClusterRoleBinding\n+metadata:\n+ name: {{ template \"name\" . }}\n+roleRef:\n+ apiGroup: rbac.authorization.k8s.io\n+ kind: ClusterRole\n+ name: {{ template \"name\" . }}\n+subjects:\n+- kind: ServiceAccount\n+ name: {{ template \"name\" . }}\n+ namespace: {{ .Release.Namespace }}\n"
},
{
"change_type": "DELETE",
"old_path": "charts/kubernikus-system/vendor/eventexporter/templates/config.yaml",
"new_path": null,
"diff": "-apiVersion: v1\n-kind: ServiceAccount\n-metadata:\n- name: {{ template \"name\" . }}\n- namespace: {{ .Release.Namespace }}\n----\n-apiVersion: rbac.authorization.k8s.io/v1beta1\n-kind: ClusterRole\n-metadata:\n- name: {{ template \"name\" . }}\n-rules:\n-- apiGroups: [\"\"]\n- resources: [\"events\"]\n- verbs: [\"get\", \"watch\", \"list\"]\n----\n-apiVersion: rbac.authorization.k8s.io/v1beta1\n-kind: ClusterRoleBinding\n-metadata:\n- name: {{ template \"name\" . }}\n-roleRef:\n- apiGroup: rbac.authorization.k8s.io\n- kind: ClusterRole\n- name: {{ template \"name\" . }}\n-subjects:\n-- kind: ServiceAccount\n- name: {{ template \"name\" . }}\n- namespace: {{ .Release.Namespace }}\n----\n-apiVersion: v1\n-kind: ConfigMap\n-metadata:\n- name: {{ template \"name\" . }}\n- namespace: {{ .Release.Namespace }}\n-data:\n- config.yaml: |-\n- metrics:\n- - name: volume_mount_error\n- event_matcher:\n- - key: InvolvedObject.Kind\n- expr: Pod\n- - key: Message\n- expr: Unable to mount volumes for pod.*\n- - key: Reason\n- expr: FailedMount\n- - key: Type\n- expr: Warning\n- labels:\n- node: Source.Host\n"
},
{
"change_type": "ADD",
"old_path": null,
"new_path": "charts/kubernikus-system/vendor/eventexporter/templates/configmap.yaml",
"diff": "+apiVersion: v1\n+kind: ConfigMap\n+metadata:\n+ name: {{ template \"name\" . }}\n+ namespace: {{ .Release.Namespace }}\n+data:\n+ {{- if .Values.config }}\n+{{ toYaml .Values.config | indent 2 }}\n+ {{ end }}\n"
},
{
"change_type": "MODIFY",
"old_path": "charts/kubernikus-system/vendor/eventexporter/templates/deployment.yaml",
"new_path": "charts/kubernikus-system/vendor/eventexporter/templates/deployment.yaml",
"diff": "@@ -16,18 +16,21 @@ spec:\napp: {{ template \"name\" . }}\nannotations:\nprometheus.io/scrape: \"true\"\n- prometheus.io/port: \"9102\"\n+ prometheus.io/port: \"{{ default 9102 .Values.metrics.port }}\"\nspec:\ncontainers:\n- name: {{ template \"name\" . }}\nimage: \"{{ .Values.image.repository }}:{{ .Values.image.tag }}\"\nimagePullPolicy: {{ .Values.image.pullPolicy }}\n+ args:\n+ - -logtostderr\n+ - -listen-address \":{{ default 9102 .Values.metrics.port }}\"\nvolumeMounts:\n- name: config-volume\nmountPath: /etc/eventexporter\nports:\n- name: metrics\n- containerPort: \"9102\"\n+ containerPort: \"{{ default 9102 .Values.metrics.port }}\"\nserviceAccount: {{ template \"name\" . }}\nvolumes:\n- name: config-volume\n"
},
{
"change_type": "ADD",
"old_path": null,
"new_path": "charts/kubernikus-system/vendor/eventexporter/templates/serviceaccount.yaml",
"diff": "+apiVersion: v1\n+kind: ServiceAccount\n+metadata:\n+ name: {{ template \"name\" . }}\n+ namespace: {{ .Release.Namespace }}\n"
},
{
"change_type": "MODIFY",
"old_path": "charts/kubernikus-system/vendor/eventexporter/values.yaml",
"new_path": "charts/kubernikus-system/vendor/eventexporter/values.yaml",
"diff": "@@ -2,3 +2,22 @@ image:\nrepository: sapcc/kubernetes-eventexporter\ntag: 0.1.0\npullPolicy: IfNotPresent\n+\n+metrics:\n+ port: \"9102\"\n+\n+config:\n+ config.yaml: |-\n+ metrics:\n+ - name: volume_mount_error\n+ event_matcher:\n+ - key: InvolvedObject.Kind\n+ expr: Pod\n+ - key: Message\n+ expr: Unable to mount volumes for pod.*\n+ - key: Reason\n+ expr: FailedMount\n+ - key: Type\n+ expr: Warning\n+ labels:\n+ node: Source.Host\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
Move config to values.yaml, resources to separate files
|
596,226 |
28.11.2018 09:27:38
| -3,600 |
4ddef8439740fbc94c27e7eb31698dd1ae8e6b5d
|
Fix eventexporter chart
|
[
{
"change_type": "MODIFY",
"old_path": "charts/kubernikus-system/vendor/eventexporter/templates/configmap.yaml",
"new_path": "charts/kubernikus-system/vendor/eventexporter/templates/configmap.yaml",
"diff": "@@ -4,6 +4,6 @@ metadata:\nname: {{ template \"name\" . }}\nnamespace: {{ .Release.Namespace }}\ndata:\n- {{- if .Values.config }}\n-{{ toYaml .Values.config | indent 2 }}\n+ {{- if .Values.metrics.config }}\n+{{ toYaml .Values.metrics.config | indent 2 }}\n{{ end }}\n"
},
{
"change_type": "MODIFY",
"old_path": "charts/kubernikus-system/vendor/eventexporter/templates/deployment.yaml",
"new_path": "charts/kubernikus-system/vendor/eventexporter/templates/deployment.yaml",
"diff": "@@ -24,13 +24,13 @@ spec:\nimagePullPolicy: {{ .Values.image.pullPolicy }}\nargs:\n- -logtostderr\n- - -listen-address \":{{ default 9102 .Values.metrics.port }}\"\n+ - -listen-address=:{{ default 9102 .Values.metrics.port }}\nvolumeMounts:\n- name: config-volume\nmountPath: /etc/eventexporter\nports:\n- name: metrics\n- containerPort: \"{{ default 9102 .Values.metrics.port }}\"\n+ containerPort: {{ default 9102 .Values.metrics.port }}\nserviceAccount: {{ template \"name\" . }}\nvolumes:\n- name: config-volume\n"
},
{
"change_type": "MODIFY",
"old_path": "charts/kubernikus-system/vendor/eventexporter/values.yaml",
"new_path": "charts/kubernikus-system/vendor/eventexporter/values.yaml",
"diff": "@@ -5,7 +5,6 @@ image:\nmetrics:\nport: \"9102\"\n-\nconfig:\nconfig.yaml: |-\nmetrics:\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
Fix eventexporter chart
|
596,226 |
17.10.2018 13:30:27
| -7,200 |
9e74359921ca628875db1ea30327b0692eb7f993
|
Fix fluent-bit configmap template
|
[
{
"change_type": "MODIFY",
"old_path": "charts/kubernikus-system/templates/fluent-bit-configmap.yaml",
"new_path": "charts/kubernikus-system/templates/fluent-bit-configmap.yaml",
"diff": "@@ -4,7 +4,6 @@ metadata:\nname: fluent-bit-config\nlabels:\napp: fluent-bit-fluent-bit\n- chart: fluent-bit-0.12.0\nrelease: fluent-bit\ndata:\nfluent-bit.conf: |-\n@@ -57,11 +56,11 @@ data:\nKube_Token_File /var/run/secrets/kubernetes.io/serviceaccount/token\nK8S-Logging.Parser On\n-{{ if .Values.filter.additionalValues }}\n+{{ if index .Values \"fluent-bit\" \"filter\" \"additionalValues\" }}\n[FILTER]\nName record_modifier\nMatch *\n-{{- range .Values.filter.additionalValues }}\n+{{- range index .Values \"fluent-bit\" \"filter\" \"additionalValues\" }}\nRecord {{ .key }} {{ .value }}\n{{- end }}\n{{- end }}\n@@ -69,15 +68,15 @@ data:\n[OUTPUT]\nName es\nMatch *\n- Host {{ .Values.backend.es.host }}\n+ Host {{ index .Values \"fluent-bit\" \"backend\" \"es\" \"host\" }}\nPort 443\nLogstash_Format On\nRetry_Limit False\nType flb_type\nTime_Key @timestamp\nLogstash_Prefix kubernikus\n- HTTP_User {{ .Values.backend.es.http_user }}\n- HTTP_Passwd {{ .Values.backend.es.http_passwd }}\n+ HTTP_User {{ index .Values \"fluent-bit\" \"backend\" \"es\" \"http_user\" }}\n+ HTTP_Passwd {{ index .Values \"fluent-bit\" \"backend\" \"es\" \"http_passwd\" }}\ntls on\ntls.verify on\ntls.debug 1\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
Fix fluent-bit configmap template
|
596,226 |
16.11.2018 17:07:43
| -3,600 |
79b05aa564c9a27d86e48bcb35be7ecb1500564e
|
Add condition to dependency, enable fluent-bit only when host value is set
|
[
{
"change_type": "MODIFY",
"old_path": "charts/kubernikus-system/requirements.lock",
"new_path": "charts/kubernikus-system/requirements.lock",
"diff": "@@ -23,5 +23,5 @@ dependencies:\n- name: fluent-bit\nrepository: https://kubernetes-charts.storage.googleapis.com/\nversion: 0.16.3\n-digest: sha256:9bcb28ff32e654185c2af4bd2d2b76cc283e31f31ec7fafa2d7112745cc7f18d\n-generated: 2018-12-12T11:20:19.430704267+01:00\n+digest: sha256:68e7833ea7849cfd2022b0d837a7d1066f128bf40d58dfef48b450611e5829f0\n+generated: 2018-12-12T11:22:28.462074993+01:00\n"
},
{
"change_type": "MODIFY",
"old_path": "charts/kubernikus-system/requirements.yaml",
"new_path": "charts/kubernikus-system/requirements.yaml",
"diff": "@@ -23,3 +23,4 @@ dependencies:\n- name: fluent-bit\nrepository: https://kubernetes-charts.storage.googleapis.com/\nversion: 0.16.3\n+ condition: fluent-bit.backend.es.host\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
Add condition to dependency, enable fluent-bit only when host value is set
|
596,226 |
12.12.2018 17:28:02
| -3,600 |
1d07bc4950bd2437c14728ce24390a19d0015cae
|
Fix fluent-bit config
|
[
{
"change_type": "MODIFY",
"old_path": "charts/kubernikus-system/requirements.yaml",
"new_path": "charts/kubernikus-system/requirements.yaml",
"diff": "@@ -23,4 +23,4 @@ dependencies:\n- name: fluent-bit\nrepository: https://kubernetes-charts.storage.googleapis.com/\nversion: 0.16.3\n- condition: fluent-bit.backend.es.host\n+ condition: fluent-bit.backend.es.host,fluent-bit.backend.es.http_user,fluent-bit.backend.es.http_passwd\n"
},
{
"change_type": "MODIFY",
"old_path": "charts/kubernikus-system/templates/fluent-bit-configmap.yaml",
"new_path": "charts/kubernikus-system/templates/fluent-bit-configmap.yaml",
"diff": "@@ -20,7 +20,7 @@ data:\nTag kube.*\nRefresh_Interval 5\nMem_Buf_Limit 5MB\n- Skip_Long_Lines On\n+ Skip_Long_Lines Off\n[INPUT]\nName systemd\n@@ -48,13 +48,18 @@ data:\nRename _COMM cmd\nRename _HOSTNAME hostname\n+ [FILTER]\n+ Name record_modifier\n+ Match kube.*\n+ Remove_key time\n+\n[FILTER]\nName kubernetes\nMatch kube.*\nKube_URL https://kubernetes.default.svc:443\n+ tls.verify Off\nKube_CA_File /var/run/secrets/kubernetes.io/serviceaccount/ca.crt\nKube_Token_File /var/run/secrets/kubernetes.io/serviceaccount/token\n- K8S-Logging.Parser On\n{{ if index .Values \"fluent-bit\" \"filter\" \"additionalValues\" }}\n[FILTER]\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
Fix fluent-bit config
|
596,226 |
13.12.2018 08:26:01
| -3,600 |
a1e6cf0916af5734ceaefff2f66f6ae1f34ff339
|
Update dependency lockfile
|
[
{
"change_type": "MODIFY",
"old_path": "charts/kubernikus-system/requirements.lock",
"new_path": "charts/kubernikus-system/requirements.lock",
"diff": "@@ -23,5 +23,5 @@ dependencies:\n- name: fluent-bit\nrepository: https://kubernetes-charts.storage.googleapis.com/\nversion: 0.16.3\n-digest: sha256:68e7833ea7849cfd2022b0d837a7d1066f128bf40d58dfef48b450611e5829f0\n-generated: 2018-12-12T11:26:21.467928357+01:00\n+digest: sha256:514d79144c98135353a816479dac18edaf7d754a2122b403a9a6d7a2af1b9fca\n+generated: 2018-12-13T08:25:25.42064407+01:00\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
Update dependency lockfile
|
596,226 |
13.12.2018 10:21:44
| -3,600 |
48549c8ca9e11f8681f6e4f91dbad17042da368d
|
Add manual versioning to fluent-bit
|
[
{
"change_type": "MODIFY",
"old_path": "charts/kubernikus-system/values.yaml",
"new_path": "charts/kubernikus-system/values.yaml",
"diff": "@@ -3256,7 +3256,9 @@ fluent-bit:\nenv: []\n- podAnnotations: {}\n+ podAnnotations:\n+ # manual versioning, raise if configmap changes\n+ versioning: 1\nexistingConfigMap: fluent-bit-config\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
Add manual versioning to fluent-bit
|
596,226 |
14.12.2018 10:34:48
| -3,600 |
0279335709f606b91f33b69ae38fb2fec15bf7dc
|
Add fluent-bit chart to vendor
|
[
{
"change_type": "ADD",
"old_path": null,
"new_path": "charts/kubernikus-system/vendor/fluent-bit/Chart.yaml",
"diff": "+appVersion: 0.14.8\n+description: Fast and Lightweight Log/Data Forwarder for Linux, BSD and OSX\n+home: http://fluentbit.io\n+icon: http://fluentbit.io/assets/img/logo1-default.png\n+keywords:\n+- logging\n+- monitoring\n+- fluent\n+- fluentd\n+maintainers:\n+- email: Kevin.Fox@pnnl.gov\n+ name: kfox1111\n+- email: eduardo@treasure-data.com\n+ name: edsiper\n+name: fluent-bit\n+sources:\n+- http://fluentbit.io\n+version: 0.16.3\n"
},
{
"change_type": "ADD",
"old_path": null,
"new_path": "charts/kubernikus-system/vendor/fluent-bit/OWNERS",
"diff": "+approvers:\n+- kfox1111\n+- edsiper\n+reviewers:\n+- kfox1111\n+- edsiper\n"
},
{
"change_type": "ADD",
"old_path": null,
"new_path": "charts/kubernikus-system/vendor/fluent-bit/README.md",
"diff": "+# Fluent-Bit Chart\n+\n+[Fluent Bit](http://fluentbit.io/) is an open source and multi-platform Log Forwarder.\n+\n+## Chart Details\n+\n+This chart will do the following:\n+\n+* Install a configmap for Fluent Bit\n+* Install a daemonset that provisions Fluent Bit [per-host architecture]\n+\n+## Installing the Chart\n+\n+To install the chart with the release name `my-release`:\n+\n+```bash\n+$ helm install --name my-release stable/fluent-bit\n+```\n+\n+When installing this chart on [Minikube](https://kubernetes.io/docs/getting-started-guides/minikube/), it's required to specify that so the DaemonSet will be able to mount the log files properly, make sure to append the _--set on\\_minikube=true_ option at the end of the _helm_ command, e.g:\n+\n+```bash\n+$ helm install --name my-release stable/fluent-bit --set on_minikube=true\n+```\n+\n+## Configuration\n+\n+The following table lists the configurable parameters of the Fluent-Bit chart and the default values.\n+\n+| Parameter | Description | Default |\n+| ----------------------- | ---------------------------------- | ----------------------- |\n+| **Backend Selection** |\n+| `backend.type` | Set the backend to which Fluent-Bit should flush the information it gathers | `forward` |\n+| **Forward Backend** |\n+| `backend.forward.host` | Target host where Fluent-Bit or Fluentd are listening for Forward messages | `fluentd` |\n+| `backend.forward.port` | TCP Port of the target service | `24284` |\n+| `backend.forward.shared_key` | A key string known by the remote Fluentd used for authorization. | `` |\n+| `backend.forward.tls` | Enable or disable TLS support | `off` |\n+| `backend.forward.tls_verify` | Force certificate validation | `on` |\n+| `backend.forward.tls_debug` | Set TLS debug verbosity level. It accept the following values: 0-4 | `1` |\n+| **ElasticSearch Backend** |\n+| `backend.es.host` | IP address or hostname of the target Elasticsearch instance | `elasticsearch` |\n+| `backend.es.port` | TCP port of the target Elasticsearch instance. | `9200` |\n+| `backend.es.index` | Elastic Index name | `kubernetes_cluster` |\n+| `backend.es.type` | Elastic Type name | `flb_type` |\n+| `backend.es.time_key` | Elastic Time Key | `@timestamp` |\n+| `backend.es.logstash_prefix` | Index Prefix. If Logstash_Prefix is equals to 'mydata' your index will become 'mydata-YYYY.MM.DD'. | `kubernetes_cluster` |\n+| `backend.es.http_user` | Optional username credential for Elastic X-Pack access. | `` |\n+| `backend.es.http_passwd:` | Password for user defined in HTTP_User. | `` |\n+| `backend.es.tls` | Enable or disable TLS support | `off` |\n+| `backend.es.tls_verify` | Force certificate validation | `on` |\n+| `backend.es.tls_ca` | TLS CA certificate for the Elastic instance (in PEM format). Specify if tls: on. | `` |\n+| `backend.es.tls_debug` | Set TLS debug verbosity level. It accept the following values: 0-4 | `1` |\n+| **HTTP Backend** |\n+| `backend.http.host` | IP address or hostname of the target HTTP Server | `127.0.0.1` |\n+| `backend.http.port` | TCP port of the target HTTP Server | `80` |\n+| `backend.http.uri` | Specify an optional HTTP URI for the target web server, e.g: /something | `\"/\"`\n+| `backend.http.http_user` | Optional username credential for Basic Authentication. | `` |\n+| `backend.http.http_passwd:` | Password for user defined in HTTP_User. | `` |\n+| `backend.http.format` | Specify the data format to be used in the HTTP request body, by default it uses msgpack, optionally it can be set to json. | `msgpack` |\n+| `backend.http.tls` | Enable or disable TLS support | `off` |\n+| `backend.http.tls_verify` | Force certificate validation | `on` |\n+| `backend.http.tls_debug` | Set TLS debug verbosity level. It accept the following values: 0-4 | `1` |\n+| **Splunk Backend** |\n+| `backend.splunk.host` | IP address or hostname of the target Splunk Server | `127.0.0.1` |\n+| `backend.splunk.port` | TCP port of the target Splunk Server | `8088` |\n+| `backend.splunk.token` | Specify the Authentication Token for the HTTP Event Collector interface. | `` |\n+| `backend.splunk.send_raw` | If enabled, record keys and values are set in the main map. | `off` |\n+| `backend.splunk.tls` | Enable or disable TLS support | `on` |\n+| `backend.splunk.tls_verify` | Force TLS certificate validation | `off` |\n+| `backend.splunk.tls_debug` | Set TLS debug verbosity level. It accept the following values: 0-4 | `1` |\n+| `backend.splunk.message_key` | Tag applied to all incoming logs | `kubernetes` |\n+| **Parsers** |\n+| `parsers.enabled` | Enable custom parsers | `false` |\n+| `parsers.regex` | List of regex parsers | `NULL` |\n+| `parsers.json` | List of json parsers | `NULL` |\n+| **General** |\n+| `annotations` | Optional deamonset set annotations | `NULL` |\n+| `podAnnotations` | Optional pod annotations | `NULL` |\n+| `existingConfigMap` | ConfigMap override | `` |\n+| `extraInputs` | Add extra Input sections to config | `` |\n+| `extraFilters` | Add extra Filter sections to config | `` |\n+| `extraOutputs` | Add extra Output sections to config | `` |\n+| `extraVolumeMounts` | Mount an extra volume, required to mount ssl certificates when elasticsearch has tls enabled | |\n+| `extraVolume` | Extra volume | |\n+| `filter.kubeURL` | Optional custom configmaps | `https://kubernetes.default.svc:443` |\n+| `filter.kubeCAFile` | Optional custom configmaps | `/var/run/secrets/kubernetes.io/serviceaccount/ca.crt` |\n+| `filter.kubeTokenFile` | Optional custom configmaps | `/var/run/secrets/kubernetes.io/serviceaccount/token` |\n+| `filter.kubeTag` | Optional top-level tag for matching in filter | `kube` |\n+| `image.fluent_bit.repository` | Image | `fluent/fluent-bit` |\n+| `image.fluent_bit.tag` | Image tag | `0.14.8` |\n+| `image.pullPolicy` | Image pull policy | `Always` |\n+| `image.pullSecrets` | Specify image pull secrets | `nil` |\n+| `input.tail.memBufLimit` | Specify Mem_Buf_Limit in tail input | `5MB` |\n+| `rbac.create` | Specifies whether RBAC resources should be created. | `true` |\n+| `serviceAccount.create` | Specifies whether a ServiceAccount should be created. | `true` |\n+| `serviceAccount.name` | The name of the ServiceAccount to use. | `NULL` |\n+| `resources.limits.cpu` | CPU limit | `100m` |\n+| `resources.limits.memory` | Memory limit | `500Mi` |\n+| `resources.requests.cpu` | CPU request | `100m` |\n+| `resources.requests.memory` | Memory request | `200Mi` |\n+| `tolerations` | Optional daemonset tolerations | `NULL` |\n+| `nodeSelector` | Node labels for fluent-bit pod assignment | `NULL` |\n+| `metrics.enabled` | Specifies whether a service for metrics should be exposed | `false` |\n+| `metrics.service.annotations` | Optional metrics service annotations | `NULL` |\n+| `metrics.service.port` | Port on where metrics should be exposed | `2020` |\n+| `metrics.service.type` | Service type for metrics | `ClusterIP` |\n+| `trackOffsets` | Specify whether to track the file offsets for tailing docker logs. This allows fluent-bit to pick up where it left after pod restarts but requires access to a `hostPath` | `false` |\n+| | | |\n+\n+\n+Specify each parameter using the `--set key=value[,key=value]` argument to `helm install`.\n+\n+Alternatively, a YAML file that specifies the values for the parameters can be provided while installing the chart. For example,\n+\n+```bash\n+$ helm install --name my-release -f values.yaml stable/fluent-bit\n+```\n+\n+> **Tip**: You can use the default [values.yaml](values.yaml)\n"
},
{
"change_type": "ADD",
"old_path": null,
"new_path": "charts/kubernikus-system/vendor/fluent-bit/templates/NOTES.txt",
"diff": "+fluent-bit is now running.\n+\n+{{- if eq .Values.backend.type \"forward\" }}\n+\n+It will forward all container logs to the svc named {{ .Values.backend.forward.host }} on port: {{ .Values.backend.forward.port }}\n+{{- else if eq .Values.backend.type \"es\" }}\n+\n+It will forward all container logs to the svc named {{ .Values.backend.es.host }} on port: {{ .Values.backend.es.port }}\n+{{- else if eq .Values.backend.type \"http\" }}\n+\n+It will forward all container logs to the svc named {{ .Values.backend.http.host }} on port: {{ .Values.backend.http.port }}\n+{{- else if eq .Values.backend.type \"splunk\" }}\n+\n+It will forward all container logs to the svc named {{ .Values.backend.splunk.host }} on port: {{ .Values.backend.splunk.port }}\n+{{- end }}\n"
},
{
"change_type": "ADD",
"old_path": null,
"new_path": "charts/kubernikus-system/vendor/fluent-bit/templates/_helpers.tpl",
"diff": "+{{/* vim: set filetype=mustache: */}}\n+{{/*\n+Expand the name of the chart.\n+*/}}\n+{{- define \"fluent-bit.name\" -}}\n+{{- default .Chart.Name .Values.nameOverride | trunc 63 | trimSuffix \"-\" -}}\n+{{- end -}}\n+\n+{{/*\n+Create a default fully qualified app name.\n+We truncate at 63 chars because some Kubernetes name fields are limited to this (by the DNS naming spec).\n+*/}}\n+{{- define \"fluent-bit.fullname\" -}}\n+{{- $name := default .Chart.Name .Values.nameOverride -}}\n+{{- printf \"%s-%s\" .Release.Name $name | trunc 63 | trimSuffix \"-\" -}}\n+{{- end -}}\n+\n+{{/*\n+Return the appropriate apiVersion for RBAC APIs.\n+*/}}\n+{{- define \"rbac.apiVersion\" -}}\n+{{- if .Capabilities.APIVersions.Has \"v1\" -}}\n+rbac.authorization.k8s.io/v1\n+{{- else if .Capabilities.APIVersions.Has \"v1beta1\" -}}\n+rbac.authorization.k8s.io/v1beta1\n+{{- else -}}\n+rbac.authorization.k8s.io/v1alpha1\n+{{- end -}}\n+{{- end -}}\n+\n+{{/*\n+Create the name of the service account to use\n+*/}}\n+{{- define \"fluent-bit.serviceAccountName\" -}}\n+{{- if .Values.serviceAccount.create -}}\n+ {{ default (include \"fluent-bit.fullname\" .) .Values.serviceAccount.name }}\n+{{- else -}}\n+ {{ default \"default\" .Values.serviceAccount.name }}\n+{{- end -}}\n+{{- end -}}\n+\n"
},
{
"change_type": "ADD",
"old_path": null,
"new_path": "charts/kubernikus-system/vendor/fluent-bit/templates/cluster-role.yaml",
"diff": "+{{- if .Values.rbac.create -}}\n+apiVersion: {{ template \"rbac.apiVersion\" . }}\n+kind: ClusterRole\n+metadata:\n+ labels:\n+ app: {{ template \"fluent-bit.name\" . }}\n+ chart: {{ .Chart.Name }}-{{ .Chart.Version }}\n+ heritage: {{ .Release.Service }}\n+ release: {{ .Release.Name }}\n+ name: {{ template \"fluent-bit.fullname\" . }}\n+rules:\n+ - apiGroups:\n+ - \"\"\n+ resources:\n+ - pods\n+ verbs:\n+ - get\n+{{- end -}}\n"
},
{
"change_type": "ADD",
"old_path": null,
"new_path": "charts/kubernikus-system/vendor/fluent-bit/templates/cluster-rolebinding.yaml",
"diff": "+{{- if .Values.rbac.create -}}\n+apiVersion: {{ template \"rbac.apiVersion\" . }}\n+kind: ClusterRoleBinding\n+metadata:\n+ labels:\n+ app: {{ template \"fluent-bit.name\" . }}\n+ chart: {{ .Chart.Name }}-{{ .Chart.Version }}\n+ heritage: {{ .Release.Service }}\n+ release: {{ .Release.Name }}\n+ name: {{ template \"fluent-bit.fullname\" . }}\n+roleRef:\n+ apiGroup: rbac.authorization.k8s.io\n+ kind: ClusterRole\n+ name: {{ template \"fluent-bit.fullname\" . }}\n+subjects:\n+ - kind: ServiceAccount\n+ name: {{ template \"fluent-bit.serviceAccountName\" . }}\n+ namespace: {{ .Release.Namespace }}\n+{{- end -}}\n"
},
{
"change_type": "ADD",
"old_path": null,
"new_path": "charts/kubernikus-system/vendor/fluent-bit/templates/config.yaml",
"diff": "+{{- if (empty .Values.existingConfigMap) -}}\n+apiVersion: v1\n+kind: ConfigMap\n+metadata:\n+ name: {{ template \"fluent-bit.fullname\" . }}-config\n+ labels:\n+ app: {{ template \"fluent-bit.fullname\" . }}\n+ chart: {{ .Chart.Name }}-{{ .Chart.Version }}\n+ heritage: {{ .Release.Service }}\n+ release: {{ .Release.Name }}\n+data:\n+ fluent-bit.conf: |-\n+ [SERVICE]\n+ Flush 1\n+ Daemon Off\n+ Log_Level info\n+ Parsers_File parsers.conf\n+{{- if .Values.parsers.enabled }}\n+ Parsers_File parsers_custom.conf\n+{{- end }}\n+{{- if .Values.metrics.enabled }}\n+ HTTP_Server On\n+ HTTP_Listen 0.0.0.0\n+ HTTP_Port 2020\n+{{- end }}\n+\n+ [INPUT]\n+ Name tail\n+ Path /var/log/containers/*.log\n+ Parser docker\n+ Tag {{ .Values.filter.kubeTag }}.*\n+ Refresh_Interval 5\n+ Mem_Buf_Limit {{ .Values.input.tail.memBufLimit }}\n+ Skip_Long_Lines On\n+{{- if .Values.trackOffsets }}\n+ DB /tail-db/tail-containers-state.db\n+ DB.Sync Normal\n+{{- end }}\n+{{ .Values.extraInputs | indent 8 }}\n+\n+ [FILTER]\n+ Name kubernetes\n+ Match {{ .Values.filter.kubeTag }}.*\n+ Kube_URL {{ .Values.filter.kubeURL }}\n+ Kube_CA_File {{ .Values.filter.kubeCAFile }}\n+ Kube_Token_File {{ .Values.filter.kubeTokenFile }}\n+{{- if .Values.filter.mergeJSONLog }}\n+ Merge_Log On\n+{{- end }}\n+{{- if .Values.filter.enableParser }}\n+ K8S-Logging.Parser On\n+{{- end }}\n+{{- if .Values.filter.enableExclude }}\n+ K8S-Logging.Exclude On\n+{{- end }}\n+{{ .Values.extraFilters | indent 8 }}\n+\n+{{ if eq .Values.backend.type \"test\" }}\n+ [OUTPUT]\n+ Name file\n+ Match *\n+ Path /tmp/fluent-bit.log\n+{{ else if eq .Values.backend.type \"forward\" }}\n+ [OUTPUT]\n+ Name forward\n+ Match *\n+ Host {{ .Values.backend.forward.host }}\n+ Port {{ .Values.backend.forward.port }}\n+ Retry_Limit False\n+{{- if .Values.backend.forward.shared_key }}\n+ Shared_Key {{ .Values.backend.forward.shared_key }}\n+{{- end }}\n+{{ else if eq .Values.backend.type \"es\" }}\n+ [OUTPUT]\n+ Name es\n+ Match *\n+ Host {{ .Values.backend.es.host }}\n+ Port {{ .Values.backend.es.port }}\n+ Logstash_Format On\n+ Retry_Limit False\n+ Type {{ .Values.backend.es.type }}\n+{{- if .Values.backend.es.time_key }}\n+ Time_Key {{ .Values.backend.es.time_key }}\n+{{- end }}\n+{{- if .Values.backend.es.logstash_prefix }}\n+ Logstash_Prefix {{ .Values.backend.es.logstash_prefix }}\n+{{ else if .Values.backend.es.index }}\n+ Index {{ .Values.backend.es.index }}\n+{{- end }}\n+{{- if .Values.backend.es.http_user }}\n+ HTTP_User {{ .Values.backend.es.http_user }}\n+ HTTP_Passwd {{ .Values.backend.es.http_passwd }}\n+{{- end }}\n+{{if eq .Values.backend.es.tls \"on\" }}\n+ tls {{ .Values.backend.es.tls }}\n+ tls.verify {{ .Values.backend.es.tls_verify }}\n+ tls.debug {{ .Values.backend.es.tls_debug }}\n+{{- if .Values.backend.es.tls_ca }}\n+ tls.ca_file /secure/es-tls-ca.crt\n+{{- end }}\n+{{- end }}\n+{{ else if eq .Values.backend.type \"splunk\" }}\n+ [OUTPUT]\n+ Name splunk\n+ Match *\n+ Host {{ .Values.backend.splunk.host }}\n+ Port {{ .Values.backend.splunk.port }}\n+ Splunk_Token {{ .Values.backend.splunk.token }}\n+ Splunk_Send_Raw {{ .Values.backend.splunk.send_raw}}\n+ TLS {{ .Values.backend.splunk.tls }}\n+ TLS.Verify {{ .Values.backend.splunk.tls_verify }}\n+ tls.debug {{ .Values.backend.splunk.tls_debug }}\n+ Message_Key {{ .Values.backend.splunk.message_key }}\n+{{ else if eq .Values.backend.type \"http\" }}\n+ [OUTPUT]\n+ Name http\n+ Match *\n+ Host {{ .Values.backend.http.host }}\n+ Port {{ .Values.backend.http.port }}\n+ URI {{ .Values.backend.http.uri }}\n+{{- if .Values.backend.http.http_user }}\n+ HTTP_User {{ .Values.backend.http.http_user }}\n+ HTTP_Passwd {{ .Values.backend.http.http_passwd }}\n+{{- end }}\n+ tls {{ .Values.backend.http.tls }}\n+ tls.verify {{ .Values.backend.http.tls_verify }}\n+ tls.debug {{ .Values.backend.http.tls_debug }}\n+{{- if .Values.backend.http.proxy }}\n+ Proxy {{ .Values.backend.http.proxy }}\n+{{- end }}\n+ Format {{ .Values.backend.http.format }}\n+{{- end }}\n+{{ .Values.extraOutputs | indent 8 }}\n+\n+ parsers.conf: |-\n+{{- if .Values.parsers.regex }}\n+{{- range .Values.parsers.regex }}\n+ [PARSER]\n+ Name {{ .name }}\n+ Format regex\n+ Regex {{ .regex }}\n+{{- if .timeKey }}\n+ Time_Key {{ .timeKey }}\n+{{- end }}\n+{{- if .timeFormat }}\n+ Time_Format {{ .timeFormat }}\n+{{- end }}\n+{{ end }}\n+{{- end }}\n+{{- if .Values.parsers.json }}\n+{{- range .Values.parsers.json }}\n+ [PARSER]\n+ Name {{ .name }}\n+ Format json\n+{{- if .timeKeep }}\n+ Time_Keep {{ .timeKeep }}\n+{{- end }}\n+{{- if .timeKey }}\n+ Time_Key {{ .timeKey }}\n+{{- end }}\n+{{- if .timeFormat }}\n+ Time_Format {{ .timeFormat }}\n+{{- end }}\n+{{- if .decodeFieldAs }}\n+ Decode_Field_As {{ .decodeFieldAs }} {{ .decodeField | default \"log\" }}\n+{{- end }}\n+{{ end }}\n+{{- end }}\n+\n+{{- end -}}\n"
},
{
"change_type": "ADD",
"old_path": null,
"new_path": "charts/kubernikus-system/vendor/fluent-bit/templates/daemonset.yaml",
"diff": "+apiVersion: extensions/v1beta1\n+kind: DaemonSet\n+metadata:\n+ name: fluent-bit\n+ labels:\n+ app: {{ template \"fluent-bit.fullname\" . }}\n+ chart: {{ .Chart.Name }}-{{ .Chart.Version }}\n+ heritage: {{ .Release.Service }}\n+ release: {{ .Release.Name }}\n+spec:\n+ updateStrategy:\n+ type: RollingUpdate\n+ template:\n+ metadata:\n+ labels:\n+ app: {{ template \"fluent-bit.fullname\" . }}\n+ release: {{ .Release.Name }}\n+ annotations:\n+ checksum/config: {{ include (print $.Template.BasePath \"/config.yaml\") . | sha256sum }}\n+{{- if .Values.podAnnotations }}\n+{{ toYaml .Values.podAnnotations | indent 8 }}\n+{{- end }}\n+ spec:\n+{{- if .Values.image.pullSecrets }}\n+ imagePullSecrets:\n+{{ toYaml .Values.image.pullSecrets | indent 8 }}\n+{{- end }}\n+ serviceAccountName: {{ template \"fluent-bit.serviceAccountName\" . }}\n+ containers:\n+ - name: fluent-bit\n+ image: \"{{ .Values.image.fluent_bit.repository }}:{{ .Values.image.fluent_bit.tag }}\"\n+ imagePullPolicy: \"{{ .Values.image.pullPolicy }}\"\n+ env:\n+{{ toYaml .Values.env | indent 10 }}\n+ resources:\n+{{ toYaml .Values.resources | indent 10 }}\n+{{- if .Values.metrics.enabled }}\n+ ports:\n+ - name: metrics\n+ containerPort: 2020\n+ protocol: TCP\n+{{- end }}\n+ volumeMounts:\n+ - name: varlog\n+ mountPath: /var/log\n+ - name: varlibdockercontainers\n+ mountPath: /var/lib/docker/containers\n+ readOnly: true\n+ - name: config\n+ mountPath: /fluent-bit/etc/fluent-bit.conf\n+ subPath: fluent-bit.conf\n+{{- if .Values.parsers.enabled }}\n+ - name: config\n+ mountPath: /fluent-bit/etc/parsers_custom.conf\n+ subPath: parsers.conf\n+{{- end }}\n+{{- if .Values.backend.es.tls_ca }}\n+ - name: es-tls-secret\n+ mountPath: /secure/es-tls-ca.crt\n+ subPath: es-tls-ca.crt\n+{{- end }}\n+{{- if .Values.trackOffsets }}\n+ - name: tail-db\n+ mountPath: /tail-db\n+{{- end }}\n+{{- if .Values.extraVolumeMounts }}\n+{{ toYaml .Values.extraVolumeMounts | indent 8 }}\n+{{- end }}\n+{{ if .Values.on_minikube }}\n+ - name: mnt\n+ mountPath: /mnt\n+ readOnly: true\n+{{ end }}\n+ terminationGracePeriodSeconds: 10\n+ {{- if .Values.nodeSelector }}\n+ nodeSelector:\n+{{ toYaml .Values.nodeSelector | indent 8 }}\n+ {{- end }}\n+ {{- if .Values.tolerations }}\n+ tolerations:\n+{{ toYaml .Values.tolerations | indent 8 }}\n+ {{- end }}\n+ volumes:\n+ - name: varlog\n+ hostPath:\n+ path: /var/log\n+ - name: varlibdockercontainers\n+ hostPath:\n+ path: /var/lib/docker/containers\n+{{- if .Values.backend.es.tls_ca }}\n+ - name: es-tls-secret\n+ secret:\n+ secretName: \"{{ template \"fluent-bit.fullname\" . }}-es-tls-secret\"\n+{{- end }}\n+{{- if .Values.trackOffsets }}\n+ - name: tail-db\n+ hostPath:\n+ path: /var/lib/fluent-bit\n+ type: DirectoryOrCreate\n+{{- end }}\n+ - name: config\n+ configMap:\n+ name: {{ if .Values.existingConfigMap }}{{ .Values.existingConfigMap }}{{- else }}{{ template \"fluent-bit.fullname\" . }}-config{{- end }}\n+{{- if .Values.extraVolumes }}\n+{{ toYaml .Values.extraVolumes | indent 6 }}\n+{{- end }}\n+{{ if .Values.on_minikube }}\n+ - name: mnt\n+ hostPath:\n+ path: /mnt\n+{{ end }}\n"
},
{
"change_type": "ADD",
"old_path": null,
"new_path": "charts/kubernikus-system/vendor/fluent-bit/templates/secret.yaml",
"diff": "+apiVersion: v1\n+kind: Secret\n+metadata:\n+ name: \"{{ template \"fluent-bit.fullname\" . }}-es-tls-secret\"\n+ labels:\n+ app: {{ template \"fluent-bit.fullname\" . }}\n+ chart: {{ .Chart.Name }}-{{ .Chart.Version }}\n+ heritage: {{ .Release.Service }}\n+ release: {{ .Release.Name }}\n+type: Opaque\n+data:\n+ es-tls-ca.crt: {{ .Values.backend.es.tls_ca | b64enc | quote }}\n"
},
{
"change_type": "ADD",
"old_path": null,
"new_path": "charts/kubernikus-system/vendor/fluent-bit/templates/service.yaml",
"diff": "+{{- if .Values.metrics.enabled }}\n+apiVersion: v1\n+kind: Service\n+metadata:\n+{{- if .Values.metrics.service.annotations }}\n+ annotations:\n+{{ toYaml .Values.metrics.service.annotations | indent 4 }}\n+{{- end }}\n+ name: {{ template \"fluent-bit.fullname\" . }}-metrics\n+ labels:\n+ app: {{ template \"fluent-bit.fullname\" . }}\n+ chart: {{ .Chart.Name }}-{{ .Chart.Version }}\n+ heritage: {{ .Release.Service }}\n+ release: {{ .Release.Name }}\n+spec:\n+ type: {{ .Values.metrics.service.type}}\n+ sessionAffinity: None\n+ ports:\n+ - port: {{ .Values.metrics.service.port }}\n+ targetPort: metrics\n+ name: metrics\n+ selector:\n+ app: {{ template \"fluent-bit.fullname\" . }}\n+ release: {{ .Release.Name }}\n+{{- end }}\n"
},
{
"change_type": "ADD",
"old_path": null,
"new_path": "charts/kubernikus-system/vendor/fluent-bit/templates/serviceaccount.yaml",
"diff": "+{{- if .Values.serviceAccount.create -}}\n+apiVersion: v1\n+kind: ServiceAccount\n+metadata:\n+ labels:\n+ app: {{ template \"fluent-bit.name\" . }}\n+ chart: {{ .Chart.Name }}-{{ .Chart.Version }}\n+ heritage: {{ .Release.Service }}\n+ release: {{ .Release.Name }}\n+ name: {{ template \"fluent-bit.serviceAccountName\" . }}\n+{{- end -}}\n"
},
{
"change_type": "ADD",
"old_path": null,
"new_path": "charts/kubernikus-system/vendor/fluent-bit/values.yaml",
"diff": "+# Minikube stores its logs in a separate directory.\n+# enable if started in minikube.\n+on_minikube: false\n+\n+image:\n+ fluent_bit:\n+ repository: fluent/fluent-bit\n+ tag: 0.14.8\n+ pullPolicy: Always\n+\n+# When enabled, exposes json and prometheus metrics on {{ .Release.Name }}-metrics service\n+metrics:\n+ enabled: false\n+ service:\n+ annotations: {}\n+ # In order for Prometheus to consume metrics automatically use the following annotations:\n+ # prometheus.io/path: \"/api/v1/metrics/prometheus\"\n+ # prometheus.io/port: \"2020\"\n+ # prometheus.io/scrape: \"true\"\n+ port: 2020\n+ type: ClusterIP\n+\n+# When enabled, fluent-bit will keep track of tailing offsets across pod restarts.\n+trackOffsets: false\n+\n+backend:\n+ type: forward\n+ forward:\n+ host: fluentd\n+ port: 24284\n+ tls: \"off\"\n+ tls_verify: \"on\"\n+ tls_debug: 1\n+ shared_key:\n+ es:\n+ host: elasticsearch\n+ port: 9200\n+ # Elastic Index Name\n+ index: kubernetes_cluster\n+ type: flb_type\n+ logstash_prefix: kubernetes_cluster\n+ time_key: \"@timestamp\"\n+ # Optional username credential for Elastic X-Pack access\n+ http_user:\n+ # Password for user defined in HTTP_User\n+ http_passwd:\n+ # Optional TLS encryption to ElasticSearch instance\n+ tls: \"off\"\n+ tls_verify: \"on\"\n+ # TLS certificate for the Elastic (in PEM format). Use if tls=on and tls_verify=on.\n+ tls_ca: \"\"\n+ # TLS debugging levels = 1-4\n+ tls_debug: 1\n+ splunk:\n+ host: 127.0.0.1\n+ port: 8088\n+ token: \"\"\n+ send_raw: \"on\"\n+ tls: \"on\"\n+ tls_verify: \"off\"\n+ tls_debug: 1\n+ message_key: \"kubernetes\"\n+\n+ ##\n+ ## Ref: http://fluentbit.io/documentation/current/output/http.html\n+ ##\n+ http:\n+ host: 127.0.0.1\n+ port: 80\n+ uri: \"/\"\n+ http_user:\n+ http_passwd:\n+ tls: \"off\"\n+ tls_verify: \"on\"\n+ tls_debug: 1\n+ ## Specify the data format to be used in the HTTP request body\n+ ## Can be either 'msgpack' or 'json'\n+ format: msgpack\n+\n+parsers:\n+ enabled: false\n+ ## List the respective parsers in key: value format per entry\n+ ## Regex required fields are name and regex. JSON required field\n+ ## is name.\n+ regex: []\n+ json: []\n+\n+env: []\n+\n+## Annotations to add to the DaemonSet's Pods\n+podAnnotations: {}\n+\n+## ConfigMap override where fullname is {{.Release.Name}}-{{.Values.existingConfigMap}}\n+## Defining existingConfigMap will cause templates/config.yaml\n+## to NOT generate a ConfigMap resource\n+##\n+existingConfigMap: \"\"\n+\n+## Add extra input sources\n+extraInputs: |-\n+## Add extra filters\n+extraFilters: |-\n+## Add extra outputs\n+extraOutputs: |-\n+\n+## Extra volumes containing additional files required for fluent-bit to work\n+## (eg. CA certificates)\n+## Ref: https://kubernetes.io/docs/concepts/storage/volumes/\n+##\n+extraVolumes: []\n+\n+## Extra volume mounts for the fluent-bit pod.\n+## Ref: https://kubernetes.io/docs/tasks/configure-pod-container/configure-volume-storage/\n+##\n+extraVolumeMounts: []\n+\n+resources:\n+ limits:\n+ memory: 100Mi\n+ requests:\n+ cpu: 100m\n+ memory: 100Mi\n+\n+## Node tolerations for fluent-bit scheduling to nodes with taints\n+## Ref: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/\n+##\n+tolerations: []\n+# - key: \"key\"\n+# operator: \"Equal|Exists\"\n+# value: \"value\"\n+# effect: \"NoSchedule|PreferNoSchedule|NoExecute(1.6 only)\"\n+\n+## Node labels for fluent-bit pod assignment\n+## Ref: https://kubernetes.io/docs/user-guide/node-selection/\n+##\n+nodeSelector: {}\n+\n+input:\n+ tail:\n+ memBufLimit: 5MB\n+\n+filter:\n+ kubeURL: https://kubernetes.default.svc:443\n+ kubeCAFile: /var/run/secrets/kubernetes.io/serviceaccount/ca.crt\n+ kubeTokenFile: /var/run/secrets/kubernetes.io/serviceaccount/token\n+ kubeTag: kube\n+# If true, check to see if the log field content is a JSON string map, if so,\n+# it append the map fields as part of the log structure.\n+# mergeJSONLog: true\n+\n+# If true, enable the use of monitoring for a pod annotation of\n+# fluentbit.io/parser: parser_name. parser_name must be the name\n+# of a parser contained within parsers.conf\n+# enableParser: true\n+\n+# If true, enable the use of monitoring for a pod annotation of\n+# fluentbit.io/exclude: true. If present, discard logs from that pod.\n+# enableExclude: true\n+\n+rbac:\n+ # Specifies whether RBAC resources should be created\n+ create: true\n+\n+serviceAccount:\n+ # Specifies whether a ServiceAccount should be created\n+ create: true\n+ # The name of the ServiceAccount to use.\n+ # If not set and create is true, a name is generated using the fullname template\n+ name:\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
Add fluent-bit chart to vendor
|
596,226 |
14.12.2018 10:38:13
| -3,600 |
fa43a5876a986a88e3ce48f60c428f5102cb461f
|
Fix fluent-bit chart template rbac api version
|
[
{
"change_type": "MODIFY",
"old_path": "charts/kubernikus-system/requirements.lock",
"new_path": "charts/kubernikus-system/requirements.lock",
"diff": "@@ -21,7 +21,7 @@ dependencies:\nrepository: file://vendor/eventexporter\nversion: 0.1.0\n- name: fluent-bit\n- repository: https://kubernetes-charts.storage.googleapis.com/\n+ repository: file://vendor/fluent-bit\nversion: 0.16.3\n-digest: sha256:514d79144c98135353a816479dac18edaf7d754a2122b403a9a6d7a2af1b9fca\n-generated: 2018-12-13T08:25:25.42064407+01:00\n+digest: sha256:63489822dd8231f38148e9bcd5003aa46e30daaff72927b19295e1a4d1f900cd\n+generated: 2018-12-14T10:36:14.519474995+01:00\n"
},
{
"change_type": "MODIFY",
"old_path": "charts/kubernikus-system/requirements.yaml",
"new_path": "charts/kubernikus-system/requirements.yaml",
"diff": "@@ -21,6 +21,6 @@ dependencies:\nrepository: file://vendor/eventexporter\nversion: 0.1.0\n- name: fluent-bit\n- repository: https://kubernetes-charts.storage.googleapis.com/\n+ repository: file://vendor/fluent-bit\nversion: 0.16.3\ncondition: fluent-bit.backend.es.host,fluent-bit.backend.es.http_user,fluent-bit.backend.es.http_passwd\n"
},
{
"change_type": "MODIFY",
"old_path": "charts/kubernikus-system/values.yaml",
"new_path": "charts/kubernikus-system/values.yaml",
"diff": "@@ -3280,7 +3280,6 @@ fluent-bit:\nrbac:\ncreate: true\n- apiVersion: rbac.authorization.k8s.io/v1beta1\nserviceAccount:\ncreate: true\n"
},
{
"change_type": "MODIFY",
"old_path": "charts/kubernikus-system/vendor/fluent-bit/templates/_helpers.tpl",
"new_path": "charts/kubernikus-system/vendor/fluent-bit/templates/_helpers.tpl",
"diff": "@@ -19,9 +19,9 @@ We truncate at 63 chars because some Kubernetes name fields are limited to this\nReturn the appropriate apiVersion for RBAC APIs.\n*/}}\n{{- define \"rbac.apiVersion\" -}}\n-{{- if .Capabilities.APIVersions.Has \"v1\" -}}\n+{{- if .Capabilities.APIVersions.Has \"rbac.authorization.k8s.io/v1\" -}}\nrbac.authorization.k8s.io/v1\n-{{- else if .Capabilities.APIVersions.Has \"v1beta1\" -}}\n+{{- else if .Capabilities.APIVersions.Has \"rbac.authorization.k8s.io/v1beta1\" -}}\nrbac.authorization.k8s.io/v1beta1\n{{- else -}}\nrbac.authorization.k8s.io/v1alpha1\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
Fix fluent-bit chart template rbac api version
|
596,226 |
17.12.2018 09:35:16
| -3,600 |
b04d5bf807faa981fb49af5d726f411e6b41fdf6
|
Go back to fluent-bit upstream and update chart
|
[
{
"change_type": "MODIFY",
"old_path": "charts/kubernikus-system/requirements.lock",
"new_path": "charts/kubernikus-system/requirements.lock",
"diff": "@@ -22,6 +22,6 @@ dependencies:\nversion: 0.1.0\n- name: fluent-bit\nrepository: https://kubernetes-charts.storage.googleapis.com/\n- version: 0.16.3\n-digest: sha256:514d79144c98135353a816479dac18edaf7d754a2122b403a9a6d7a2af1b9fca\n-generated: 2018-12-13T08:25:25.42064407+01:00\n+ version: 1.1.1\n+digest: sha256:c4731cb8cc9f10db8e81718524495d7596c061867090dc4120e3bb8969846eee\n+generated: 2018-12-17T09:28:23.256060573+01:00\n"
},
{
"change_type": "MODIFY",
"old_path": "charts/kubernikus-system/requirements.yaml",
"new_path": "charts/kubernikus-system/requirements.yaml",
"diff": "@@ -22,5 +22,5 @@ dependencies:\nversion: 0.1.0\n- name: fluent-bit\nrepository: https://kubernetes-charts.storage.googleapis.com/\n- version: 0.16.3\n+ version: 1.1.1\ncondition: fluent-bit.backend.es.host,fluent-bit.backend.es.http_user,fluent-bit.backend.es.http_passwd\n"
},
{
"change_type": "MODIFY",
"old_path": "charts/kubernikus-system/values.yaml",
"new_path": "charts/kubernikus-system/values.yaml",
"diff": "@@ -3280,7 +3280,6 @@ fluent-bit:\nrbac:\ncreate: true\n- apiVersion: rbac.authorization.k8s.io/v1beta1\nserviceAccount:\ncreate: true\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
Go back to fluent-bit upstream and update chart
|
596,226 |
17.12.2018 09:50:08
| -3,600 |
6771a877d95715c3a7fd2ded2896186b0193c35b
|
Raise fluent-bit resource limits
|
[
{
"change_type": "MODIFY",
"old_path": "charts/kubernikus-system/values.yaml",
"new_path": "charts/kubernikus-system/values.yaml",
"diff": "@@ -3269,10 +3269,10 @@ fluent-bit:\nresources:\nlimits:\ncpu: 500m\n- memory: 100Mi\n+ memory: 200Mi\nrequests:\ncpu: 100m\n- memory: 50Mi\n+ memory: 100Mi\ntolerations: []\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
Raise fluent-bit resource limits
|
596,240 |
17.12.2018 14:54:20
| -3,600 |
83185726b1499963c97110e40d02fbdd4faab8d3
|
add grafana banners
|
[
{
"change_type": "ADD",
"old_path": "assets/grafana-banners-kubernikus.png",
"new_path": "assets/grafana-banners-kubernikus.png",
"diff": "Binary files /dev/null and b/assets/grafana-banners-kubernikus.png differ\n"
},
{
"change_type": "ADD",
"old_path": "assets/grafana-banners-openstack.png",
"new_path": "assets/grafana-banners-openstack.png",
"diff": "Binary files /dev/null and b/assets/grafana-banners-openstack.png differ\n"
},
{
"change_type": "ADD",
"old_path": "assets/grafana-banners-scaleout.png",
"new_path": "assets/grafana-banners-scaleout.png",
"diff": "Binary files /dev/null and b/assets/grafana-banners-scaleout.png differ\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
add grafana banners
|
596,240 |
18.12.2018 15:01:03
| -3,600 |
5f9ec1fe5cb32a168ad8cb1738ed12781ebd24b0
|
escalate VolumeAttachment criticality
|
[
{
"change_type": "MODIFY",
"old_path": "charts/kubernikus-system/charts/prometheus/kubernetes.alerts",
"new_path": "charts/kubernikus-system/charts/prometheus/kubernetes.alerts",
"diff": "@@ -110,7 +110,7 @@ groups:\nlabels:\ntier: kks\nservice: node\n- severity: warning\n+ severity: critical\ncontext: pvc\nplaybook: docs/support/playbook/kubernikus/volume_attachment_failed.html\nannotations:\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
escalate VolumeAttachment criticality
|
596,226 |
18.12.2018 13:42:00
| -3,600 |
bc4db18b245e51f829ddf579090023311901d83e
|
Adds chart template rendering and api version testing
|
[
{
"change_type": "MODIFY",
"old_path": "Makefile",
"new_path": "Makefile",
"diff": "@@ -131,6 +131,9 @@ else\ngrep -v \"CONT\\|PAUSE\"\nendif\n+.PHONY: test-charts\n+test-charts:\n+ docker run -ti --rm -v $(shell pwd)/charts:/charts -v $(shell pwd)/test/charts:/test --entrypoint \"/test/charts.sh\" alpine/helm:2.10.0\ninclude code-generate.mk\ncode-gen: client-gen informer-gen lister-gen deepcopy-gen\n"
},
{
"change_type": "ADD",
"old_path": null,
"new_path": "test/charts/charts.sh",
"diff": "+#!/bin/ash\n+\n+#set -o xtrace\n+\n+# get minimum supported api version with `kubectl explain resource`\n+# in ash no arrays are supported, using grep on env var instead\n+k8s_min_version=\"1.7\"\n+resources=`echo -e \"ClusterRole:rbac.authorization.k8s.io/v1alpha1\\n\\\n+ClusterRole:rbac.authorization.k8s.io/v1beta1\\n\\\n+ClusterRoleBinding:rbac.authorization.k8s.io/v1beta1\\n\\\n+ClusterRoleBinding:rbac.authorization.k8s.io/v1alpha1\\n\\\n+Config:v1\\n\\\n+ConfigMap:v1\\n\\\n+DaemonSet:extensions/v1beta1\\n\\\n+Deployment:extensions/v1beta1\\n\\\n+Ingress:extensions/v1beta1\\n\\\n+PersistentVolumeClaim:v1\\n\\\n+Role:rbac.authorization.k8s.io/v1beta1\\n\\\n+RoleBinding:rbac.authorization.k8s.io/v1beta1\\n\\\n+Secret:v1\\n\\\n+Service:v1\\n\\\n+ServiceAccount:v1\"`\n+\n+helm init --client-only\n+helm repo add bugroger-charts https://raw.githubusercontent.com/BugRoger/charts/repo\n+\n+for chart in /charts/*; do\n+ if [ -d \"$chart\" ]; then\n+ echo \"Rendering chart in $chart ...\"\n+ cd $chart\n+ helm dependency build\n+ cat values.yaml /test/dummy-values.yaml > /tmp/values.yaml\n+ helm template --debug -f /tmp/values.yaml . > /tmp/chart.yaml\n+ retval=$?\n+ rm -f $chart/charts/*.tgz\n+ if [ $retval -ne 0 ]; then\n+ echo \"Rendering of template failed.\"\n+ exit $retval\n+ fi\n+ echo \"Done.\"\n+\n+ echo \"Checking API versions ...\"\n+ while IFS= read -r line <&3; do\n+ if echo \"$line\" | grep \"^---$\" > /dev/null; then\n+ unset api_real kind_real\n+ continue\n+ fi\n+ api_tmp=`echo $line | grep \"^apiVersion: .*$\" | awk -F': ' '{print $2}' | sed 's/\\\"//g'`\n+ kind_tmp=`echo $line | grep \"^kind: .*$\" | awk -F': ' '{print $2}' | sed 's/\\\"//g'`\n+ if [[ ! -z \"$api_tmp\" ]]; then\n+ api_real=$api_tmp\n+ fi\n+ if [[ ! -z \"$kind_tmp\" ]]; then\n+ kind_real=$kind_tmp\n+ fi\n+ if [[ ! -z \"$api_real\" && ! -z \"$kind_real\" ]]; then\n+ if ! echo \"$resources\" | grep \"^$kind_real:$api_real$\" > /dev/null; then\n+ echo \"kind: $kind_real apiVersion: $api_real not matching minimum version requirements ($k8s_min_version)!\"\n+ exit 1\n+ fi\n+ unset api_real kind_real\n+ fi\n+ done 3< \"/tmp/chart.yaml\"\n+ echo \"Done.\"\n+ fi\n+done\n"
},
{
"change_type": "ADD",
"old_path": null,
"new_path": "test/charts/dummy-values.yaml",
"diff": "+domain: xyz.com\n+bootstrapToken: xyz\n+nodePassword: xyz\n+certsSecretName: xyz\n+standalone: true\n+openstack:\n+ auth_user_id: xyz\n+ auth_user_password: xyz\n+ auth_domain: xyz\n+ auth_project: xyz\n+ auth_project_domain: xyz\n+ authURL: http://xyz.com\n+ username: xyz\n+ password: xyz\n+ domainName: xyz.com\n+ lbSubnetID: xyz\n+ lbFloatingNetworkID: xyz\n+ routerID: xyz\n+ projectID: xy\n+ region: xy-xy-1\n+api:\n+ tls_crt: xyz\n+ apiserverHost: xyz.com\n+ wormholeHost: xyz.com\n+version:\n+ kubernikus: xyz\n+etcd:\n+ openstack:\n+ authURL: http://xyz.com\n+ username: xyz\n+ password: xyz\n+ domainName: xyz.com\n+ projectID: xyz\n+vice-president:\n+ vice:\n+ cert: xyz\n+ key: xyz\n+ intermediate: xyz\n+ first_name: xyz\n+ last_name: xyz\n+ email: xyz@xyz.com\n+ country: xyz\n+ province: xyz\n+ locality: xyz\n+ organization: xyz\n+ organizational_unit: xyz\n+ default_challenge: xyz\n+global:\n+ domain: xyz.com\n+fluent-bit:\n+ backend:\n+ es:\n+ host: xyz.com\n+authentication:\n+ enabled: true\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
Adds chart template rendering and api version testing
|
596,226 |
21.12.2018 14:44:05
| -3,600 |
d0af431212f025babbf3032f8e33a7e49d37751a
|
Update fluent-bit chart and docker image
|
[
{
"change_type": "MODIFY",
"old_path": "charts/kubernikus-system/requirements.lock",
"new_path": "charts/kubernikus-system/requirements.lock",
"diff": "@@ -22,6 +22,6 @@ dependencies:\nversion: 0.1.0\n- name: fluent-bit\nrepository: https://kubernetes-charts.storage.googleapis.com/\n- version: 1.1.1\n-digest: sha256:c4731cb8cc9f10db8e81718524495d7596c061867090dc4120e3bb8969846eee\n-generated: 2018-12-17T09:28:23.256060573+01:00\n+ version: 1.2.0\n+digest: sha256:8d69e427b1f85b68ae97cf1ffbd6f563bf6f5642667cd00b104bbdd02d67007b\n+generated: 2018-12-21T14:42:09.575053176+01:00\n"
},
{
"change_type": "MODIFY",
"old_path": "charts/kubernikus-system/requirements.yaml",
"new_path": "charts/kubernikus-system/requirements.yaml",
"diff": "@@ -22,5 +22,5 @@ dependencies:\nversion: 0.1.0\n- name: fluent-bit\nrepository: https://kubernetes-charts.storage.googleapis.com/\n- version: 1.1.1\n+ version: 1.2.0\ncondition: fluent-bit.backend.es.host,fluent-bit.backend.es.http_user,fluent-bit.backend.es.http_passwd\n"
},
{
"change_type": "MODIFY",
"old_path": "charts/kubernikus-system/values.yaml",
"new_path": "charts/kubernikus-system/values.yaml",
"diff": "@@ -3230,7 +3230,7 @@ fluent-bit:\nimage:\nfluent_bit:\nrepository: fluent/fluent-bit\n- tag: 0.14.9\n+ tag: 1.0.1\npullPolicy: IfNotPresent\nmetrics:\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
Update fluent-bit chart and docker image
|
596,226 |
02.01.2019 11:38:45
| -3,600 |
b6f1fa5b6bc8de86a1d3f7bb745f3d650c048333
|
Fix ListNodes when nodes are already deleted
|
[
{
"change_type": "MODIFY",
"old_path": "pkg/client/openstack/kluster/client.go",
"new_path": "pkg/client/openstack/kluster/client.go",
"diff": "@@ -137,10 +137,12 @@ func (c *klusterClient) ListNodes(pool *models.NodePool) ([]Node, error) {\nprefix := fmt.Sprintf(\"%v-%v-\", c.Kluster.Spec.Name, pool.Name)\nerr = servers.List(c.ComputeClient, servers.ListOpts{Name: prefix}).EachPage(func(page pagination.Page) (bool, error) {\n+ if page != nil {\nunfilteredNodes, err = ExtractServers(page)\nif err != nil {\nreturn false, err\n}\n+ }\nreturn true, nil\n})\nif err != nil {\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
Fix ListNodes when nodes are already deleted
|
596,226 |
07.01.2019 13:17:30
| -3,600 |
e5f375ca02ae14a6a04d252d717626e8c4505041
|
Export mount success events
|
[
{
"change_type": "MODIFY",
"old_path": "charts/kubernikus-system/vendor/eventexporter/values.yaml",
"new_path": "charts/kubernikus-system/vendor/eventexporter/values.yaml",
"diff": "@@ -8,7 +8,7 @@ metrics:\nconfig:\nconfig.yaml: |-\nmetrics:\n- - name: volume_mount_error\n+ - name: volume_mount_error_total\nevent_matcher:\n- key: InvolvedObject.Kind\nexpr: Pod\n@@ -20,3 +20,13 @@ metrics:\nexpr: Warning\nlabels:\nnode: Source.Host\n+ - name: volume_mount_success_total\n+ event_matcher:\n+ - key: InvolvedObject.Kind\n+ expr: Pod\n+ - key: Message\n+ expr: MountVolume.SetUp succeeded for volume .pvc-.*\n+ - key: Reason\n+ expr: SuccessfulMountVolume\n+ labels:\n+ node: Source.Host\n\\ No newline at end of file\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
Export mount success events
|
596,226 |
07.01.2019 13:48:34
| -3,600 |
cf685c6a84e1155f9cbd6ab8e7d6e684162f0743
|
Include successful mounts in volume mount alert
|
[
{
"change_type": "MODIFY",
"old_path": "charts/kubernikus-system/charts/prometheus/kubernetes.alerts",
"new_path": "charts/kubernikus-system/charts/prometheus/kubernetes.alerts",
"diff": "@@ -105,8 +105,8 @@ groups:\ndescription: The pod {{ $labels.namespace }}/{{ $labels.pod_name }} killed several times in short succession. This could be due to wrong resource limits.\n- alert: VolumeAttachmentFailed\n- expr: sum(increase(volume_mount_error[3h]) / 36) by (node) > 1\n- for: 30m\n+ expr: sum(increase(volume_mount_error_total[30m])) by (node) > 10 unless sum(increase(volume_mount_success_total[30m])) by (node) > 1\n+ for: 5m\nlabels:\ntier: kks\nservice: node\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
Include successful mounts in volume mount alert
|
596,226 |
07.01.2019 14:06:03
| -3,600 |
f125e12113f20447c3545eab3bbf26b228c76136
|
Add configmap checksum to deployment
|
[
{
"change_type": "MODIFY",
"old_path": "charts/kubernikus-system/vendor/eventexporter/templates/deployment.yaml",
"new_path": "charts/kubernikus-system/vendor/eventexporter/templates/deployment.yaml",
"diff": "@@ -17,6 +17,7 @@ spec:\nannotations:\nprometheus.io/scrape: \"true\"\nprometheus.io/port: \"{{ default 9102 .Values.metrics.port }}\"\n+ checksum/config: {{ include (print $.Template.BasePath \"/configmap.yaml\") . | sha256sum }}\nspec:\ncontainers:\n- name: {{ template \"name\" . }}\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
Add configmap checksum to deployment
|
596,226 |
11.01.2019 14:35:13
| -3,600 |
5dba6f4833b5d00f53bebb5bd4f04762011fd368
|
Add server group/building block e2e tests
|
[
{
"change_type": "MODIFY",
"old_path": "test/e2e/cleanup_test.go",
"new_path": "test/e2e/cleanup_test.go",
"diff": "package main\nimport (\n+ \"strings\"\n\"testing\"\n\"time\"\n+ \"github.com/gophercloud/gophercloud\"\n+ \"github.com/gophercloud/gophercloud/openstack\"\nblockstorage_quota \"github.com/gophercloud/gophercloud/openstack/blockstorage/extensions/quotasets\"\ncompute_quota \"github.com/gophercloud/gophercloud/openstack/compute/v2/extensions/quotasets\"\n+ \"github.com/gophercloud/gophercloud/openstack/compute/v2/extensions/servergroups\"\n\"github.com/gophercloud/gophercloud/openstack/identity/v3/tokens\"\n\"github.com/stretchr/testify/assert\"\n\"github.com/stretchr/testify/require\"\n@@ -34,6 +38,7 @@ func (s *CleanupTests) Run(t *testing.T) {\nif s.Reuse == false {\nt.Run(\"QuotaPostFlightCheck\", s.QuotaPostFlightCheck)\n+ t.Run(\"ServerGroupsGotDeleted\", s.ServerGroupsGotDeleted)\n}\n}\n}\n@@ -75,3 +80,22 @@ func (s *CleanupTests) QuotaPostFlightCheck(t *testing.T) {\nassert.True(t, storage.Volumes.InUse == 0, \"There should be no Volume left in use\")\nassert.True(t, storage.Gigabytes.InUse == 0, \"There should be no Storage left in use\")\n}\n+\n+func (s *CleanupTests) ServerGroupsGotDeleted(t *testing.T) {\n+ computeClient, err := openstack.NewComputeV2(s.OpenStack.Provider, gophercloud.EndpointOpts{})\n+ require.NoError(t, err, \"There should be no error creating compute client\")\n+\n+ allPages, err := servergroups.List(computeClient).AllPages()\n+ require.NoError(t, err, \"There should be no error listing server groups\")\n+\n+ allGroups, err := servergroups.ExtractServerGroups(allPages)\n+ require.NoError(t, err, \"There should be no error extracting server groups\")\n+\n+ count := 0\n+ for _, sg := range allGroups {\n+ if strings.HasPrefix(sg.Name, \"e2e-\") {\n+ count++\n+ }\n+ }\n+ require.Equal(t, 0, count, \"There should be no server groups left\")\n+}\n"
},
{
"change_type": "MODIFY",
"old_path": "test/e2e/etcdbr_test.go",
"new_path": "test/e2e/etcdbr_test.go",
"diff": "@@ -53,7 +53,7 @@ func (e *EtcdBackupTests) WaitForBackupRestore(t *testing.T) {\nrv := pod.GetResourceVersion()\nrequire.NotEmpty(t, rv, \"ResourceVersion should not be empty\")\n- cmd := fmt.Sprintf(\"rm -rf %s/*\", EtcdDataDir)\n+ cmd := fmt.Sprintf(\"mv %s %s.bak\", EtcdDataDir, EtcdDataDir)\n_, _, err = e.KubernetesControlPlane.ExecCommandInContainerWithFullOutput(e.Namespace, podName, \"backup\", \"/bin/sh\", \"-c\", cmd)\nrequire.NoError(t, err, \"Deletion of etcd data failed: %s\", err)\n"
},
{
"change_type": "MODIFY",
"old_path": "test/e2e/main_test.go",
"new_path": "test/e2e/main_test.go",
"diff": "@@ -171,7 +171,7 @@ func TestRunner(t *testing.T) {\napiTests := &APITests{kubernikus, klusterName}\nt.Run(\"API\", apiTests.Run)\n- nodeTests := &NodeTests{kubernetes, kubernikus, SmokeTestNodeCount, klusterName}\n+ nodeTests := &NodeTests{kubernetes, kubernikus, openstack, SmokeTestNodeCount, klusterName}\nif !t.Run(\"Nodes\", nodeTests.Run) {\nreturn\n}\n"
},
{
"change_type": "MODIFY",
"old_path": "test/e2e/node_test.go",
"new_path": "test/e2e/node_test.go",
"diff": "@@ -8,6 +8,10 @@ import (\n\"testing\"\n\"time\"\n+ \"github.com/gophercloud/gophercloud\"\n+ \"github.com/gophercloud/gophercloud/openstack\"\n+ \"github.com/gophercloud/gophercloud/openstack/compute/v2/servers\"\n+ \"github.com/gophercloud/gophercloud/openstack/identity/v3/tokens\"\n\"github.com/stretchr/testify/assert\"\n\"github.com/stretchr/testify/require\"\n\"k8s.io/api/core/v1\"\n@@ -33,6 +37,7 @@ const (\ntype NodeTests struct {\nKubernetes *framework.Kubernetes\nKubernikus *framework.Kubernikus\n+ OpenStack *framework.OpenStack\nExpectedNodeCount int\nKlusterName string\n}\n@@ -46,7 +51,8 @@ func (k *NodeTests) Run(t *testing.T) {\nt.Run(\"Healthy\", k.StateHealthy) &&\nt.Run(\"Ready\", k.ConditionReady) &&\nt.Run(\"Labeled\", k.Labeled) &&\n- t.Run(\"Sufficient\", k.Sufficient)\n+ t.Run(\"Sufficient\", k.Sufficient) &&\n+ t.Run(\"SameBuildingBlock\", k.SameBuildingBlock)\n}\nfunc (k *NodeTests) StateRunning(t *testing.T) {\n@@ -187,3 +193,39 @@ func (k *NodeTests) checkCondition(t *testing.T, conditionType v1.NodeConditionT\nreturn count, err\n}\n+\n+func (k *NodeTests) SameBuildingBlock(t *testing.T) {\n+ if k.ExpectedNodeCount < 2 {\n+ return\n+ }\n+\n+ computeClient, err := openstack.NewComputeV2(k.OpenStack.Provider, gophercloud.EndpointOpts{})\n+ require.NoError(t, err, \"There should be no error creating compute client\")\n+\n+ project, err := tokens.Get(k.OpenStack.Identity, k.OpenStack.Provider.Token()).ExtractProject()\n+ require.NoError(t, err, \"There should be no error while extracting the project\")\n+\n+ serversListOpts := servers.ListOpts{\n+ Name: \"e2e-\",\n+ TenantID: project.ID,\n+ }\n+\n+ allPages, err := servers.List(computeClient, serversListOpts).AllPages()\n+ require.NoError(t, err, \"There should be no error while listing all servers\")\n+\n+ var s []struct {\n+ BuildingBlock string `json:\"OS-EXT-SRV-ATTR:host\"`\n+ }\n+ err = servers.ExtractServersInto(allPages, &s)\n+ require.NoError(t, err, \"There should be no error extracting server info\")\n+\n+ bb := \"\"\n+ for _, bbs := range s {\n+ require.NotEmpty(t, bbs.BuildingBlock, \"Node building block should not be empty\")\n+ if bb == \"\" {\n+ bb = string(bbs.BuildingBlock)\n+ } else {\n+ require.Equal(t, bb, bbs.BuildingBlock, \"Nodes should be on the same building block\")\n+ }\n+ }\n+}\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
Add server group/building block e2e tests
|
596,240 |
17.01.2019 10:53:09
| -3,600 |
0e19af9003ab9331be3634cd21c6c7d03d237e90
|
adds more launchctl workers
|
[
{
"change_type": "MODIFY",
"old_path": "pkg/controller/operator.go",
"new_path": "pkg/controller/operator.go",
"diff": "@@ -172,7 +172,7 @@ func NewKubernikusOperator(options *KubernikusOperatorOptions, logger log.Logger\ncase \"groundctl\":\no.Config.Kubernikus.Controllers[\"groundctl\"] = NewGroundController(10, o.Factories, o.Clients, recorder, o.Config, logger)\ncase \"launchctl\":\n- o.Config.Kubernikus.Controllers[\"launchctl\"] = launch.NewController(1, o.Factories, o.Clients, recorder, logger)\n+ o.Config.Kubernikus.Controllers[\"launchctl\"] = launch.NewController(10, o.Factories, o.Clients, recorder, logger)\ncase \"routegc\":\no.Config.Kubernikus.Controllers[\"routegc\"] = routegc.New(60*time.Second, o.Factories, logger)\ncase \"deorbiter\":\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
adds more launchctl workers
|
596,226 |
17.01.2019 14:55:45
| -3,600 |
f10216794ad1fb6f954e355131613da6b2c1922e
|
Update eventexporter chart and config
|
[
{
"change_type": "MODIFY",
"old_path": "charts/kubernikus-system/vendor/eventexporter/templates/clusterrole.yaml",
"new_path": "charts/kubernikus-system/vendor/eventexporter/templates/clusterrole.yaml",
"diff": "@@ -6,3 +6,6 @@ rules:\n- apiGroups: [\"\"]\nresources: [\"events\"]\nverbs: [\"get\", \"watch\", \"list\"]\n+- apiGroups: [\"\"]\n+ resources: [\"pods\"]\n+ verbs: [\"get\"]\n"
},
{
"change_type": "MODIFY",
"old_path": "charts/kubernikus-system/vendor/eventexporter/templates/deployment.yaml",
"new_path": "charts/kubernikus-system/vendor/eventexporter/templates/deployment.yaml",
"diff": "@@ -24,14 +24,17 @@ spec:\nimage: \"{{ .Values.image.repository }}:{{ .Values.image.tag }}\"\nimagePullPolicy: {{ .Values.image.pullPolicy }}\nargs:\n+ - -discard=60s\n- -logtostderr\n- -listen-address=:{{ default 9102 .Values.metrics.port }}\n+ - -v=0\nvolumeMounts:\n- name: config-volume\nmountPath: /etc/eventexporter\nports:\n- name: metrics\ncontainerPort: {{ default 9102 .Values.metrics.port }}\n+{{ toYaml .Values.resources | indent 8 }}\nserviceAccount: {{ template \"name\" . }}\nvolumes:\n- name: config-volume\n"
},
{
"change_type": "MODIFY",
"old_path": "charts/kubernikus-system/vendor/eventexporter/values.yaml",
"new_path": "charts/kubernikus-system/vendor/eventexporter/values.yaml",
"diff": "image:\nrepository: sapcc/kubernetes-eventexporter\n- tag: 0.1.0\n+ tag: 0.2.0\npullPolicy: IfNotPresent\n+resources:\n+ requests:\n+ cpu: 50m\n+ memory: 20Mi\n+ limits:\n+ cpu: 200m\n+ memory: 100Mi\n+\nmetrics:\nport: \"9102\"\nconfig:\n@@ -12,14 +20,14 @@ metrics:\nevent_matcher:\n- key: InvolvedObject.Kind\nexpr: Pod\n- - key: Message\n- expr: Unable to mount volumes for pod.*\n- key: Reason\n- expr: FailedMount\n+ expr: FailedAttachVolume\n- key: Type\nexpr: Warning\n+ - key: Source.Component\n+ expr: attachdetach.*\nlabels:\n- node: Source.Host\n+ node: Object.Spec.NodeName\n- name: volume_mount_success_total\nevent_matcher:\n- key: InvolvedObject.Kind\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
Update eventexporter chart and config
|
596,226 |
29.01.2019 09:00:44
| -3,600 |
f770388b94cb98c6f6eafb7e3940e8b03f022dfc
|
Add alert and chart testing to travis
|
[
{
"change_type": "MODIFY",
"old_path": ".travis.yml",
"new_path": ".travis.yml",
"diff": "@@ -15,7 +15,7 @@ before_script:\n- sudo apt-get -y -o Dpkg::Options::=\"--force-confnew\" install docker-ce\nscript:\n- - make pull build gh-pages\n+ - make pull build gh-pages test-alerts test-charts\nafter_success:\n- if [ \"$TRAVIS_BRANCH\" == \"master\" ] || [ \"$TRAVIS_BRANCH\" == \"feature\" ]; then\n"
},
{
"change_type": "MODIFY",
"old_path": "Makefile",
"new_path": "Makefile",
"diff": "@@ -88,6 +88,12 @@ gh-pages:\ndocker cp gh-pages:/public/kubernikus gh-pages\ndocker rm gh-pages\n+tests-image:\n+ docker build $(BUILD_ARGS) -t sapcc/kubernikus-tests:$(VERSION) --cache-from=sapcc/kubernikus-tests:latest ./contrib/kubernikus-tests\n+ docker tag sapcc/kubernikus-tests:$(VERSION) sapcc/kubernikus-tests:latest\n+ docker push sapcc/kubernikus-tests:$(VERSION)\n+ docker push sapcc/kubernikus-tests:latest\n+\npkg/api/rest/operations/kubernikus_api.go: swagger.yml\nifneq (,$(wildcard $(SWAGGER_BIN)))\n$(SWAGGER_BIN) generate server --name kubernikus --target pkg/api --model-package models \\\n@@ -137,7 +143,11 @@ endif\n.PHONY: test-charts\ntest-charts:\n- docker run -ti --rm -v $(shell pwd)/charts:/charts -v $(shell pwd)/test/charts:/test --entrypoint \"/test/charts.sh\" alpine/helm:2.10.0\n+ docker run -ti --rm -v $(shell pwd):/go/src/github.com/sapcc/kubernikus --entrypoint \"/go/src/github.com/sapcc/kubernikus/test/charts/charts.sh\" sapcc/kubernikus-tests:latest\n+\n+.PHONY: test-alerts\n+test-alerts:\n+ docker run -ti --rm -v $(shell pwd):/go/src/github.com/sapcc/kubernikus --entrypoint \"/go/src/github.com/sapcc/kubernikus/test/alerts/alerts.sh\" sapcc/kubernikus-tests:latest\ninclude code-generate.mk\ncode-gen: client-gen informer-gen lister-gen deepcopy-gen\n"
},
{
"change_type": "ADD",
"old_path": null,
"new_path": "contrib/kubernikus-tests/Dockerfile",
"diff": "+FROM golang:1.11.5-alpine3.8\n+\n+WORKDIR /go/src/github.com/sapcc/kubernikus/\n+\n+RUN apk add --no-cache make git curl bash\n+\n+RUN curl -Lf https://storage.googleapis.com/kubernetes-helm/helm-v2.10.0-linux-amd64.tar.gz \\\n+ | tar --strip-components=1 -C /usr/local/bin -zxv \\\n+ && helm version -c\n+\n+RUN curl -Lf https://github.com/prometheus/prometheus/releases/download/v2.4.2/prometheus-2.4.2.linux-amd64.tar.gz \\\n+ | tar --strip-components=1 -C /usr/local/bin -zxv prometheus-2.4.2.linux-amd64/promtool \\\n+ && promtool --version\n+\n+RUN curl -Lfo /usr/local/bin/yq https://github.com/mikefarah/yq/releases/download/2.2.1/yq_linux_amd64 \\\n+ && chmod +x /usr/local/bin/yq\n"
},
{
"change_type": "ADD",
"old_path": null,
"new_path": "test/alerts/alerts.sh",
"diff": "+#!/bin/bash\n+\n+TMP_PROMETHEUS_CONFIG=/tmp/prometheus.yaml\n+TMP_VALUES=/tmp/values.yaml\n+TMP_CONFIGMAP=/tmp/config.yaml.bak\n+\n+helm init --client-only\n+mkdir -p /var/run/secrets/kubernetes.io/serviceaccount/\n+touch /var/run/secrets/kubernetes.io/serviceaccount/token\n+cd charts/kubernikus-system/charts/prometheus\n+cat values.yaml ../../../../test/alerts/dummy-values.yaml > ${TMP_VALUES}\n+cp templates/config.yaml ${TMP_CONFIGMAP}\n+sed -i \"s/kubernikus-system\\/charts\\///g\" templates/config.yaml\n+\n+echo \"Checking prometheus rules ...\"\n+promtool check rules *.rules\n+if [ $? -ne 0 ]; then\n+ echo \"Checking of prometheus rules failed.\"\n+ exit 1\n+fi\n+\n+echo \"Checking prometheus alerts ...\"\n+helm template -f ${TMP_VALUES} . | yq r - data[prometheus.yaml] > ${TMP_PROMETHEUS_CONFIG}\n+if [ ! -s \"${TMP_PROMETHEUS_CONFIG}\" ]\n+then\n+ echo \"Prometheus config is empty, exiting.\"\n+ exit 1\n+fi\n+promtool check config ${TMP_PROMETHEUS_CONFIG}\n+if [ $? -ne 0 ]; then\n+ echo \"Checking of prometheus config failed.\"\n+ exit 1\n+fi\n+\n+cp -f ${TMP_CONFIGMAP} templates/config.yaml\n+rm -f ${TMP_PROMETHEUS_CONFIG} ${TMP_VALUES}\n"
},
{
"change_type": "ADD",
"old_path": null,
"new_path": "test/alerts/dummy-values.yaml",
"diff": "+global:\n+ domain: xyz.com\n+ region: xy-xy-1\n+ cluster: xy\n+ cluster_type: xy\n\\ No newline at end of file\n"
},
{
"change_type": "MODIFY",
"old_path": "test/charts/charts.sh",
"new_path": "test/charts/charts.sh",
"diff": "@@ -24,19 +24,27 @@ ServiceAccount:v1\"`\nhelm init --client-only\nhelm repo add bugroger-charts https://raw.githubusercontent.com/BugRoger/charts/repo\n-for chart in /charts/*; do\n+pwd=$(pwd)\n+for chart in $pwd/charts/*; do\nif [ -d \"$chart\" ]; then\necho \"Rendering chart in $chart ...\"\ncd $chart\n- helm dependency build\n- cat values.yaml /test/dummy-values.yaml > /tmp/values.yaml\n+ # fix cross device move of overlay fs\n+ if [ -d \"./charts\" ]; then\n+ cp -a ./charts ./charts.bak\n+ rm -rf ./charts\n+ mv ./charts.bak ./charts\n+ fi\n+ helm dependency build --debug\n+ cat values.yaml ../../test/charts/dummy-values.yaml > /tmp/values.yaml\nhelm template --debug -f /tmp/values.yaml . > /tmp/chart.yaml\nretval=$?\n- rm -f $chart/charts/*.tgz\n+ rm -f ./charts/*.tgz\nif [ $retval -ne 0 ]; then\necho \"Rendering of template failed.\"\nexit $retval\nfi\n+ cd ..\necho \"Done.\"\necho \"Checking API versions ...\"\n"
},
{
"change_type": "MODIFY",
"old_path": "test/charts/dummy-values.yaml",
"new_path": "test/charts/dummy-values.yaml",
"diff": "@@ -47,6 +47,9 @@ vice-president:\ndefault_challenge: xyz\nglobal:\ndomain: xyz.com\n+ region: xy-xy-1\n+ cluster: xy\n+ cluster_type: xy\nfluent-bit:\nbackend:\nes:\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
Add alert and chart testing to travis (#372)
|
596,226 |
09.01.2019 14:12:15
| -3,600 |
87c52900918f18eb4f8c58c1766cc1a14eb7d064
|
Export volume multi attach error events
|
[
{
"change_type": "MODIFY",
"old_path": "charts/kubernikus-system/vendor/eventexporter/values.yaml",
"new_path": "charts/kubernikus-system/vendor/eventexporter/values.yaml",
"diff": "@@ -38,3 +38,13 @@ metrics:\nexpr: SuccessfulMountVolume\nlabels:\nnode: Source.Host\n+ - name: volume_multi_attach_error_total\n+ event_matcher:\n+ - key: InvolvedObject.Kind\n+ expr: Pod\n+ - key: Message\n+ expr: Multi-Attach error for volume.*\n+ - key: Reason\n+ expr: FailedAttachVolume\n+ labels:\n+ node: InvolvedObject.Name\n\\ No newline at end of file\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
Export volume multi attach error events
|
596,226 |
29.01.2019 09:04:09
| -3,600 |
79487f5c99e09213fa192de7e17353206ef169ac
|
Fix attach volume event export on 1.7, update eventexporter to 0.3.1
|
[
{
"change_type": "MODIFY",
"old_path": "charts/kubernikus-system/vendor/eventexporter/values.yaml",
"new_path": "charts/kubernikus-system/vendor/eventexporter/values.yaml",
"diff": "image:\nrepository: sapcc/kubernetes-eventexporter\n- tag: 0.2.0\n+ tag: 0.3.1\npullPolicy: IfNotPresent\nresources:\n@@ -21,7 +21,7 @@ metrics:\n- key: InvolvedObject.Kind\nexpr: Pod\n- key: Reason\n- expr: FailedAttachVolume\n+ expr: (FailedAttachVolume|FailedMount)\n- key: Type\nexpr: Warning\n- key: Source.Component\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
Fix attach volume event export on 1.7, update eventexporter to 0.3.1
|
596,226 |
15.01.2019 16:13:12
| -3,600 |
a34ba8f5b7ce901422024a7e3579febc26fe1830
|
Make kubelet container ready for rook ceph operator
|
[
{
"change_type": "MODIFY",
"old_path": "pkg/templates/node_1.10.go",
"new_path": "pkg/templates/node_1.10.go",
"diff": "@@ -83,9 +83,13 @@ systemd:\n--volume var-lib-cni,kind=host,source=/var/lib/cni \\\n--volume var-log,kind=host,source=/var/log \\\n--volume etc-machine-id,kind=host,source=/etc/machine-id,readOnly=true \\\n+ --volume modprobe,kind=host,source=/usr/sbin/modprobe \\\n+ --volume lib-modules,kind=host,source=/lib/modules \\\n--mount volume=var-lib-cni,target=/var/lib/cni \\\n--mount volume=var-log,target=/var/log \\\n--mount volume=etc-machine-id,target=/etc/machine-id \\\n+ --mount volume=modprobe,target=/usr/sbin/modprobe \\\n+ --mount volume=lib-modules,target=/lib/modules \\\n--insecure-options=image\"\nEnvironment=\"KUBELET_IMAGE_TAG=v1.10.11\"\nEnvironment=\"KUBELET_IMAGE_URL=docker://sapcc/hyperkube\"\n@@ -111,6 +115,7 @@ systemd:\n{{- if .NodeTaints }}\n--register-with-taints={{ .NodeTaints | join \",\" }} \\\n{{- end }}\n+ --volume-plugin-dir=/var/lib/kubelet/volumeplugins \\\n--exit-on-lock-contention\nExecStop=-/usr/bin/rkt stop --uuid-file=/var/run/kubelet-pod.uuid\nRestart=always\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
Make kubelet container ready for rook ceph operator
|
596,226 |
29.01.2019 09:46:03
| -3,600 |
c20f2be2db85c2edb32718297ce544b0b1ffe954
|
Removed unneeded volume mount
|
[
{
"change_type": "MODIFY",
"old_path": "pkg/templates/node_1.10.go",
"new_path": "pkg/templates/node_1.10.go",
"diff": "@@ -84,12 +84,10 @@ systemd:\n--volume var-log,kind=host,source=/var/log \\\n--volume etc-machine-id,kind=host,source=/etc/machine-id,readOnly=true \\\n--volume modprobe,kind=host,source=/usr/sbin/modprobe \\\n- --volume lib-modules,kind=host,source=/lib/modules \\\n--mount volume=var-lib-cni,target=/var/lib/cni \\\n--mount volume=var-log,target=/var/log \\\n--mount volume=etc-machine-id,target=/etc/machine-id \\\n--mount volume=modprobe,target=/usr/sbin/modprobe \\\n- --mount volume=lib-modules,target=/lib/modules \\\n--insecure-options=image\"\nEnvironment=\"KUBELET_IMAGE_TAG=v1.10.11\"\nEnvironment=\"KUBELET_IMAGE_URL=docker://sapcc/hyperkube\"\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
Removed unneeded volume mount
|
596,226 |
30.01.2019 17:42:47
| -3,600 |
2ef92e845cf681310a36fca7ab48be89d747988b
|
Delay check for current core os version in e2e tests
|
[
{
"change_type": "MODIFY",
"old_path": "test/e2e/node_test.go",
"new_path": "test/e2e/node_test.go",
"diff": "@@ -45,7 +45,7 @@ type NodeTests struct {\nfunc (k *NodeTests) Run(t *testing.T) {\n_ = t.Run(\"Created\", k.StateRunning) &&\nt.Run(\"Registered\", k.Registered) &&\n- //t.Run(\"LatestStableContainerLinux\", k.LatestStableContainerLinux) &&\n+ t.Run(\"LatestStableContainerLinux\", k.LatestStableContainerLinux) &&\nt.Run(\"Schedulable\", k.StateSchedulable) &&\nt.Run(\"NetworkUnavailable\", k.ConditionNetworkUnavailable) &&\nt.Run(\"Healthy\", k.StateHealthy) &&\n@@ -113,7 +113,6 @@ func (k *NodeTests) Registered(t *testing.T) {\n}\nfunc (k NodeTests) LatestStableContainerLinux(t *testing.T) {\n-\nnodes, err := k.Kubernetes.ClientSet.CoreV1().Nodes().List(meta_v1.ListOptions{})\nif !assert.NoError(t, err) {\nreturn\n@@ -126,18 +125,37 @@ func (k NodeTests) LatestStableContainerLinux(t *testing.T) {\nreturn\n}\n+ version := \"\"\n+ var date time.Time\nscanner := bufio.NewScanner(resp.Body)\nfor scanner.Scan() {\nkeyval := strings.Split(scanner.Text(), \"=\")\n+\nif len(keyval) == 2 && keyval[0] == \"COREOS_VERSION\" {\n- for _, node := range nodes.Items {\n- assert.Contains(t, node.Status.NodeInfo.OSImage, keyval[1], \"Node %s is not on latest version\", node.Name)\n+ version = keyval[1]\n+ if !assert.NotEmpty(t, version, \"Failed to detect latest stable Container Linux version\") {\n+ return\n+ }\n+ }\n+\n+ if len(keyval) == 2 && keyval[0] == \"COREOS_BUILD_ID\" {\n+ date, err = time.Parse(\"2006-01-02\", keyval[1][1:11])\n+ if !assert.NoError(t, err) {\n+ return\n+ }\n+ if !assert.NotEmpty(t, date, \"Could not get release date\") {\n+ return\n}\n+ // check if release is at least 72 old, otherwise image might not be up-to-date\n+ if time.Since(date).Hours() < 72 {\nreturn\n}\n}\n- t.Error(\"Failed to detect latest stable Container Linux version\")\n+ }\n+ for _, node := range nodes.Items {\n+ assert.Contains(t, node.Status.NodeInfo.OSImage, version, \"Node %s is not on latest version\", node.Name)\n+ }\n}\nfunc (k *NodeTests) Sufficient(t *testing.T) {\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
Delay check for current core os version in e2e tests
|
596,226 |
11.02.2019 11:43:40
| -3,600 |
31b8b7e02370433a2f2aad44f00a91da4500a23d
|
Fix exporter charts resources
|
[
{
"change_type": "MODIFY",
"old_path": "charts/kubernikus-system/vendor/eventexporter/templates/deployment.yaml",
"new_path": "charts/kubernikus-system/vendor/eventexporter/templates/deployment.yaml",
"diff": "@@ -34,7 +34,8 @@ spec:\nports:\n- name: metrics\ncontainerPort: {{ default 9102 .Values.metrics.port }}\n-{{ toYaml .Values.resources | indent 8 }}\n+ resources:\n+{{ toYaml .Values.resources | indent 10 }}\nserviceAccount: {{ template \"name\" . }}\nvolumes:\n- name: config-volume\n"
},
{
"change_type": "MODIFY",
"old_path": "charts/kubernikus-system/vendor/oomkill-exporter/templates/daemonset.yaml",
"new_path": "charts/kubernikus-system/vendor/oomkill-exporter/templates/daemonset.yaml",
"diff": "@@ -29,7 +29,8 @@ spec:\n- -v=0\nsecurityContext:\nprivileged: true\n-{{ toYaml .Values.resources | indent 8 }}\n+ resources:\n+{{ toYaml .Values.resources | indent 10 }}\nenv:\n- name: DOCKER_HOST\nvalue: \"unix:///var/run/docker.sock\"\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
Fix exporter charts resources
|
596,229 |
14.02.2019 13:58:35
| -3,600 |
71325612e18bb12fa570193a404de1bb5c8ac6bc
|
backward compatibility for node pool name validation
|
[
{
"change_type": "MODIFY",
"old_path": "pkg/api/models/node_pool.go",
"new_path": "pkg/api/models/node_pool.go",
"diff": "@@ -33,7 +33,7 @@ type NodePool struct {\n// name\n// Required: true\n// Max Length: 20\n- // Pattern: ^[a-z]([a-z0-9]*)?$\n+ // Pattern: ^[a-z0-9]([-\\.a-z0-9]*)?$\nName string `json:\"name\"`\n// size\n@@ -103,7 +103,7 @@ func (m *NodePool) validateName(formats strfmt.Registry) error {\nreturn err\n}\n- if err := validate.Pattern(\"name\", \"body\", string(m.Name), `^[a-z]([a-z0-9]*)?$`); err != nil {\n+ if err := validate.Pattern(\"name\", \"body\", string(m.Name), `^[a-z0-9]([-\\.a-z0-9]*)?$`); err != nil {\nreturn err\n}\n"
},
{
"change_type": "MODIFY",
"old_path": "pkg/api/spec/embedded_spec.go",
"new_path": "pkg/api/spec/embedded_spec.go",
"diff": "@@ -511,7 +511,7 @@ func init() {\n\"name\": {\n\"type\": \"string\",\n\"maxLength\": 20,\n- \"pattern\": \"^[a-z]([a-z0-9]*)?$\",\n+ \"pattern\": \"^[a-z0-9]([-\\\\.a-z0-9]*)?$\",\n\"x-nullable\": false\n},\n\"size\": {\n@@ -1282,7 +1282,7 @@ func init() {\n\"name\": {\n\"type\": \"string\",\n\"maxLength\": 20,\n- \"pattern\": \"^[a-z]([a-z0-9]*)?$\",\n+ \"pattern\": \"^[a-z0-9]([-\\\\.a-z0-9]*)?$\",\n\"x-nullable\": false\n},\n\"size\": {\n"
},
{
"change_type": "MODIFY",
"old_path": "swagger.yml",
"new_path": "swagger.yml",
"diff": "@@ -411,7 +411,7 @@ definitions:\nname:\nx-nullable: false\ntype: string\n- pattern: '^[a-z]([a-z0-9]*)?$'\n+ pattern: '^[a-z0-9]([-\\.a-z0-9]*)?$'\nmaxLength: 20\nsize:\nx-nullable: false\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
backward compatibility for node pool name validation (#378)
|
596,226 |
14.02.2019 14:17:25
| -3,600 |
b2c9d983769df07a2addbe20dff35a4de9de2e26
|
Make volume attachment alerts less picky
|
[
{
"change_type": "MODIFY",
"old_path": "charts/kubernikus-system/charts/prometheus/kubernetes.alerts",
"new_path": "charts/kubernikus-system/charts/prometheus/kubernetes.alerts",
"diff": "@@ -107,7 +107,7 @@ groups:\ndescription: The pod {{ $labels.namespace }}/{{ $labels.pod_name }} killed several times in short succession. This could be due to wrong resource limits.\n- alert: VolumeAttachmentFailed\n- expr: sum(increase(volume_mount_error_total[30m])) by (node) > 5 unless sum(increase(volume_mount_success_total[30m])) by (node) > 1\n+ expr: sum(increase(volume_mount_error_total[3h])) by (node) > 75 unless sum(increase(volume_mount_success_total[3h])) by (node) > 1\nfor: 5m\nlabels:\ntier: kks\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
Make volume attachment alerts less picky
|
596,240 |
15.02.2019 15:00:48
| -3,600 |
cc6ca471e3b9bcf33a9391c09013eb361e363bc7
|
adds validations and defaults also to nodepool update
|
[
{
"change_type": "MODIFY",
"old_path": "pkg/api/handlers/create_cluster.go",
"new_path": "pkg/api/handlers/create_cluster.go",
"diff": "@@ -38,38 +38,36 @@ func (d *createCluster) Handle(params operations.CreateClusterParams, principal\n}\nvar metadata *models.OpenstackMetadata\n- spec.Name = name\n- for i, pool := range spec.NodePools {\n- // Set default image\n- if pool.Image == \"\" {\n- spec.NodePools[i].Image = DEFAULT_IMAGE\n- }\n-\n- // Set default AvailabilityZone\n- if pool.AvailabilityZone == \"\" {\n- if metadata == nil {\n- m, err := fetchOpenstackMetadata(params.HTTPRequest, principal)\n+ var defaultAVZ string\n+ if len(spec.NodePools) > 0 {\n+ m, err := FetchOpenstackMetadataFunc(params.HTTPRequest, principal)\nif err != nil {\nreturn NewErrorResponse(&operations.CreateClusterDefault{}, 500, err.Error())\n}\nmetadata = m\n- }\n+\navz, err := getDefaultAvailabilityZone(metadata)\nif err != nil {\nreturn NewErrorResponse(&operations.CreateClusterDefault{}, 500, err.Error())\n}\n- spec.NodePools[i].AvailabilityZone = avz\n- } else {\n- if metadata == nil {\n- m, err := fetchOpenstackMetadata(params.HTTPRequest, principal)\n- if err != nil {\n- return NewErrorResponse(&operations.CreateClusterDefault{}, 500, err.Error())\n+ defaultAVZ = avz\n}\n- metadata = m\n+\n+ spec.Name = name\n+ for i, pool := range spec.NodePools {\n+ // Set default image\n+ if pool.Image == \"\" {\n+ spec.NodePools[i].Image = DEFAULT_IMAGE\n}\n- if err := validateAavailabilityZone(pool.AvailabilityZone, metadata); err != nil {\n- return NewErrorResponse(&operations.CreateClusterDefault{}, 409, \"Availability Zone %s is invalid: %s\", pool.AvailabilityZone, err)\n+\n+ // Set default AvailabilityZone\n+ if pool.AvailabilityZone == \"\" {\n+ spec.NodePools[i].AvailabilityZone = defaultAVZ\n}\n+\n+ // Validate AVZ\n+ if err := validateAavailabilityZone(spec.NodePools[i].AvailabilityZone, metadata); err != nil {\n+ return NewErrorResponse(&operations.CreateClusterDefault{}, 409, \"Availability Zone %s is invalid: %s\", spec.NodePools[i].AvailabilityZone, err)\n}\n}\n"
},
{
"change_type": "MODIFY",
"old_path": "pkg/api/handlers/update_cluster.go",
"new_path": "pkg/api/handlers/update_cluster.go",
"diff": "package handlers\nimport (\n+ \"fmt\"\n+ \"strings\"\n+\n\"github.com/go-openapi/runtime/middleware\"\napierrors \"k8s.io/apimachinery/pkg/api/errors\"\n@@ -19,9 +22,17 @@ type updateCluster struct {\n}\nfunc (d *updateCluster) Handle(params operations.UpdateClusterParams, principal *models.Principal) middleware.Responder {\n+ metadata, err := FetchOpenstackMetadataFunc(params.HTTPRequest, principal)\n+ if err != nil {\n+ return NewErrorResponse(&operations.UpdateClusterDefault{}, 500, err.Error())\n+ }\n- kluster, err := editCluster(d.Kubernikus.Kubernikus().Klusters(d.Namespace), principal, params.Name, func(kluster *v1.Kluster) error {\n+ defaultAVZ, err := getDefaultAvailabilityZone(metadata)\n+ if err != nil {\n+ return NewErrorResponse(&operations.UpdateClusterDefault{}, 500, err.Error())\n+ }\n+ kluster, err := editCluster(d.Kubernikus.Kubernikus().Klusters(d.Namespace), principal, params.Name, func(kluster *v1.Kluster) error {\n// find the deleted nodepools\ndeletedNodePoolNames, err := detectNodePoolChanges(kluster.Spec.NodePools, params.Body.Spec.NodePools)\nif err != nil {\n@@ -59,6 +70,17 @@ func (d *updateCluster) Handle(params operations.UpdateClusterParams, principal\n}\n}\n+ for i, paramPool := range nodePools {\n+ // Set default AvailabilityZone\n+ if paramPool.AvailabilityZone == \"\" {\n+ nodePools[i].AvailabilityZone = defaultAVZ\n+ }\n+\n+ if err := validateAavailabilityZone(nodePools[i].AvailabilityZone, metadata); err != nil {\n+ return fmt.Errorf(\"Availability Zone %s is invalid: %s\", nodePools[i].AvailabilityZone, err)\n+ }\n+ }\n+\n// Update nodepool\nkluster.Spec.NodePools = nodePools\nkluster.Spec.SSHPublicKey = params.Body.Spec.SSHPublicKey\n@@ -69,11 +91,18 @@ func (d *updateCluster) Handle(params operations.UpdateClusterParams, principal\nreturn nil\n})\n+\nif err != nil {\nif apierrors.IsNotFound(err) {\nreturn NewErrorResponse(&operations.UpdateClusterDefault{}, 404, \"Not found\")\n}\n+\n+ if strings.HasPrefix(err.Error(), \"Availability Zone\") {\n+ return NewErrorResponse(&operations.UpdateClusterDefault{}, 409, err.Error())\n+ }\n+\nreturn NewErrorResponse(&operations.UpdateClusterDefault{}, 500, err.Error())\n}\n+\nreturn operations.NewUpdateClusterOK().WithPayload(klusterFromCRD(kluster))\n}\n"
},
{
"change_type": "MODIFY",
"old_path": "pkg/api/handlers/util.go",
"new_path": "pkg/api/handlers/util.go",
"diff": "@@ -21,6 +21,7 @@ import (\nvar (\nDEFAULT_IMAGE = spec.MustDefaultString(\"NodePool\", \"image\")\n+ FetchOpenstackMetadataFunc = fetchOpenstackMetadata\n)\nfunc accountSelector(principal *models.Principal) labels.Selector {\n"
},
{
"change_type": "MODIFY",
"old_path": "pkg/api/rest/api_test.go",
"new_path": "pkg/api/rest/api_test.go",
"diff": "@@ -18,6 +18,7 @@ import (\napipkg \"github.com/sapcc/kubernikus/pkg/api\"\n\"github.com/sapcc/kubernikus/pkg/api/auth\"\n+ \"github.com/sapcc/kubernikus/pkg/api/handlers\"\n\"github.com/sapcc/kubernikus/pkg/api/models\"\n\"github.com/sapcc/kubernikus/pkg/api/rest/operations\"\n\"github.com/sapcc/kubernikus/pkg/api/spec\"\n@@ -162,6 +163,16 @@ func TestClusterShow(t *testing.T) {\nfunc TestClusterUpdate(t *testing.T) {\nhandler, rt := createTestHandler(t)\n+\n+ handlers.FetchOpenstackMetadataFunc = func(request *http.Request, principal *models.Principal) (*models.OpenstackMetadata, error) {\n+ return &models.OpenstackMetadata{\n+ AvailabilityZones: []models.AvailabilityZone{\n+ models.AvailabilityZone{\"us-west-1a\"},\n+ models.AvailabilityZone{\"us-east-1a\"},\n+ },\n+ }, nil\n+ }\n+\nkluster := kubernikusv1.Kluster{\nObjectMeta: metav1.ObjectMeta{\nName: fmt.Sprintf(\"%s-%s\", \"nase\", ACCOUNT),\n@@ -183,7 +194,6 @@ func TestClusterUpdate(t *testing.T) {\n},\nNodePools: []models.NodePool{\n{\n- AvailabilityZone: \"us-west-1a\",\nFlavor: \"flavour\",\nImage: \"image\",\nName: \"poolname\",\n@@ -261,5 +271,4 @@ func TestClusterUpdate(t *testing.T) {\n//assert nodepool was updated\nassert.Equal(t, updateObject.Spec.NodePools, apiResponse.Spec.NodePools)\n-\n}\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
adds validations and defaults also to nodepool update
|
596,240 |
15.02.2019 15:11:50
| -3,600 |
66bfc34ae6c47ad62761ab74b5af28c38aabf32a
|
more linter nits
|
[
{
"change_type": "MODIFY",
"old_path": "pkg/api/rest/api_test.go",
"new_path": "pkg/api/rest/api_test.go",
"diff": "@@ -167,8 +167,8 @@ func TestClusterUpdate(t *testing.T) {\nhandlers.FetchOpenstackMetadataFunc = func(request *http.Request, principal *models.Principal) (*models.OpenstackMetadata, error) {\nreturn &models.OpenstackMetadata{\nAvailabilityZones: []models.AvailabilityZone{\n- models.AvailabilityZone{\"us-west-1a\"},\n- models.AvailabilityZone{\"us-east-1a\"},\n+ {\"us-west-1a\"},\n+ {\"us-east-1a\"},\n},\n}, nil\n}\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
more linter nits
|
596,226 |
18.02.2019 10:24:56
| -3,600 |
2a4454fa01a9d53dc14485814fe1331e9c4b55e1
|
Bump etcdbrctl to version 0.5.0
|
[
{
"change_type": "MODIFY",
"old_path": "charts/kube-master/charts/etcd/values.yaml",
"new_path": "charts/kube-master/charts/etcd/values.yaml",
"diff": "@@ -22,7 +22,7 @@ resources:\nbackup:\nimage:\nrepository: sapcc/etcdbrctl\n- tag: 0.4.1\n+ tag: 0.5.0\npullPolicy: IfNotPresent\nconfig:\n# do a full-backup every hour\n@@ -40,5 +40,5 @@ backup:\ncpu: 100m\nmemory: 128Mi\nlimits:\n- cpu: 300m\n- memory: 1Gi\n+ cpu: 500m\n+ memory: 1.5Gi\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
Bump etcdbrctl to version 0.5.0 (#382)
|
596,226 |
19.02.2019 11:18:06
| -3,600 |
d3e9fae51862f4c4faf1edafdf9f44feb8b1d785
|
Enable node certificate rotation
|
[
{
"change_type": "MODIFY",
"old_path": "pkg/controller/ground/bootstrap.go",
"new_path": "pkg/controller/ground/bootstrap.go",
"diff": "@@ -34,6 +34,9 @@ func SeedKluster(clients config.Clients, factories config.Factories, kluster *v1\nif err := SeedAutoApproveNodeBootstrapTokens(kubernetes); err != nil {\nreturn err\n}\n+ if err := SeedAutoRenewalNodeCertificates(kubernetes); err != nil {\n+ return err\n+ }\nif err := SeedKubernikusAdmin(kubernetes); err != nil {\nreturn err\n}\n@@ -259,3 +262,39 @@ func SeedAutoApproveNodeBootstrapTokens(client clientset.Interface) error {\n},\n})\n}\n+\n+func SeedAutoRenewalNodeCertificates(client clientset.Interface) error {\n+ err := bootstrap.CreateOrUpdateClusterRole(client, &rbac.ClusterRole{\n+ ObjectMeta: metav1.ObjectMeta{\n+ Name: \"system:certificates.k8s.io:certificatesigningrequests:selfnodeclient\",\n+ },\n+ Rules: []rbac.PolicyRule{\n+ {\n+ Verbs: []string{\"create\"},\n+ APIGroups: []string{\"certificates.k8s.io\"},\n+ Resources: []string{\"certificatesigningrequests/selfnodeclient\"},\n+ },\n+ },\n+ })\n+ if err != nil {\n+ return err\n+ }\n+\n+ return bootstrap.CreateOrUpdateClusterRoleBinding(client, &rbac.ClusterRoleBinding{\n+ ObjectMeta: metav1.ObjectMeta{\n+ Name: \"kubernikus:auto-approve-renewals-for-nodes\",\n+ },\n+ RoleRef: rbac.RoleRef{\n+ APIGroup: rbac.GroupName,\n+ Kind: \"ClusterRole\",\n+ Name: \"system:certificates.k8s.io:certificatesigningrequests:selfnodeclient\",\n+ },\n+ Subjects: []rbac.Subject{\n+ {\n+ APIGroup: rbac.GroupName,\n+ Kind: \"Group\",\n+ Name: \"system:nodes\",\n+ },\n+ },\n+ })\n+}\n"
},
{
"change_type": "MODIFY",
"old_path": "pkg/templates/node_1.10.go",
"new_path": "pkg/templates/node_1.10.go",
"diff": "@@ -114,7 +114,8 @@ systemd:\n--register-with-taints={{ .NodeTaints | join \",\" }} \\\n{{- end }}\n--volume-plugin-dir=/var/lib/kubelet/volumeplugins \\\n- --exit-on-lock-contention\n+ --exit-on-lock-contention \\\n+ --rotate-certificates\nExecStop=-/usr/bin/rkt stop --uuid-file=/var/run/kubelet-pod.uuid\nRestart=always\nRestartSec=10\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
Enable node certificate rotation
|
596,240 |
20.02.2019 11:19:32
| -3,600 |
e878444ed5275d48883b0ed97e560f16713e1541
|
Generalize Migrations
This commit makes the migration mechanism more flexible. It passes the complete `config.Clients` and `config.Factories` structs.
This allows to use the full functionality of all clients. Concrete use is the seeding of StorageClasses or RBAC rules to already existing clusters.
|
[
{
"change_type": "MODIFY",
"old_path": "pkg/controller/migration/controller.go",
"new_path": "pkg/controller/migration/controller.go",
"diff": "@@ -18,8 +18,7 @@ const (\n)\ntype MigrationReconciler struct {\n- config.Clients\n-\n+ Clients config.Clients\nFactories config.Factories\nRecorder record.EventRecorder\nLogger log.Logger\n@@ -40,11 +39,11 @@ func (mr *MigrationReconciler) Reconcile(kluster *v1.Kluster) (bool, error) {\n//We only care about klusters with pending migrations\nif !migration.MigrationsPending(kluster) {\n// Ensure the kluster migration status is up to date\n- return false, util.UpdateKlusterMigrationStatus(mr.Kubernikus.Kubernikus(), kluster, false)\n+ return false, util.UpdateKlusterMigrationStatus(mr.Clients.Kubernikus.Kubernikus(), kluster, false)\n}\n//Ensure pending migrations are reflected in the status\n- if err := util.UpdateKlusterMigrationStatus(mr.Kubernikus.Kubernikus(), kluster, true); err != nil {\n+ if err := util.UpdateKlusterMigrationStatus(mr.Clients.Kubernikus.Kubernikus(), kluster, true); err != nil {\nreturn false, err\n}\n@@ -53,7 +52,7 @@ func (mr *MigrationReconciler) Reconcile(kluster *v1.Kluster) (bool, error) {\nreturn false, nil\n}\n- err := migration.Migrate(kluster, mr.Kubernetes, mr.Kubernikus, mr.Factories.Openstack)\n+ err := migration.Migrate(kluster, mr.Clients, mr.Factories)\nmr.Logger.Log(\n\"msg\", \"Migrating spec\",\n\"kluster\", kluster.Name,\n@@ -67,7 +66,7 @@ func (mr *MigrationReconciler) Reconcile(kluster *v1.Kluster) (bool, error) {\nreturn false, err\n}\n//Clear the klusters migration status as migrations are applied successfully\n- util.UpdateKlusterMigrationStatus(mr.Kubernikus.Kubernikus(), kluster, false)\n+ util.UpdateKlusterMigrationStatus(mr.Clients.Kubernikus.Kubernikus(), kluster, false)\nreturn false, nil\n}\n"
},
{
"change_type": "MODIFY",
"old_path": "pkg/migration/1_init.go",
"new_path": "pkg/migration/1_init.go",
"diff": "package migration\nimport (\n- \"k8s.io/client-go/kubernetes\"\n-\n- \"github.com/sapcc/kubernikus/pkg/apis/kubernikus/v1\"\n- \"github.com/sapcc/kubernikus/pkg/client/openstack\"\n+ v1 \"github.com/sapcc/kubernikus/pkg/apis/kubernikus/v1\"\n+ \"github.com/sapcc/kubernikus/pkg/controller/config\"\n)\n//Init is the first migration that only sets the SpecVersion to 1\n-func Init(rawKluster []byte, current *v1.Kluster, client kubernetes.Interface, openstackFactory openstack.SharedOpenstackClientFactory) (err error) {\n+func Init(rawKluster []byte, current *v1.Kluster, clients config.Clients, factories config.Factories) (err error) {\nreturn nil\n}\n"
},
{
"change_type": "MODIFY",
"old_path": "pkg/migration/2_add_aggregation_layer_certificates.go",
"new_path": "pkg/migration/2_add_aggregation_layer_certificates.go",
"diff": "@@ -2,15 +2,15 @@ package migration\nimport (\nmetav1 \"k8s.io/apimachinery/pkg/apis/meta/v1\"\n- \"k8s.io/client-go/kubernetes\"\n- \"github.com/sapcc/kubernikus/pkg/apis/kubernikus/v1\"\n- \"github.com/sapcc/kubernikus/pkg/client/openstack\"\n+ v1 \"github.com/sapcc/kubernikus/pkg/apis/kubernikus/v1\"\n+ \"github.com/sapcc/kubernikus/pkg/controller/config\"\n\"github.com/sapcc/kubernikus/pkg/util\"\n)\n-func AddAggregationLayerCertificates(rawKluster []byte, kluster *v1.Kluster, client kubernetes.Interface, openstackFactory openstack.SharedOpenstackClientFactory) (err error) {\n- apiSecret, err := client.CoreV1().Secrets(kluster.Namespace).Get(kluster.GetName(), metav1.GetOptions{})\n+func AddAggregationLayerCertificates(rawKluster []byte, kluster *v1.Kluster, clients config.Clients, factories config.Factories) (err error) {\n+\n+ apiSecret, err := clients.Kubernetes.CoreV1().Secrets(kluster.Namespace).Get(kluster.GetName(), metav1.GetOptions{})\nif err != nil {\nreturn err\n}\n@@ -36,7 +36,7 @@ func AddAggregationLayerCertificates(rawKluster []byte, kluster *v1.Kluster, cli\n}\napiSecret.Data = secretData\n- _, err = client.CoreV1().Secrets(kluster.Namespace).Update(apiSecret)\n+ _, err = clients.Kubernetes.CoreV1().Secrets(kluster.Namespace).Update(apiSecret)\nreturn err\n}\n"
},
{
"change_type": "MODIFY",
"old_path": "pkg/migration/3_etcdbr_create_storage_container.go",
"new_path": "pkg/migration/3_etcdbr_create_storage_container.go",
"diff": "@@ -4,15 +4,14 @@ import (\n\"errors\"\nmetav1 \"k8s.io/apimachinery/pkg/apis/meta/v1\"\n- \"k8s.io/client-go/kubernetes\"\n- \"github.com/sapcc/kubernikus/pkg/apis/kubernikus/v1\"\n- \"github.com/sapcc/kubernikus/pkg/client/openstack\"\n+ v1 \"github.com/sapcc/kubernikus/pkg/apis/kubernikus/v1\"\n+ \"github.com/sapcc/kubernikus/pkg/controller/config\"\netcd_util \"github.com/sapcc/kubernikus/pkg/util/etcd\"\n)\n-func CreateEtcdBackupStorageContainer(rawKluster []byte, current *v1.Kluster, client kubernetes.Interface, openstackFactory openstack.SharedOpenstackClientFactory) (err error) {\n- secret, err := client.CoreV1().Secrets(current.GetNamespace()).Get(current.GetName(), metav1.GetOptions{})\n+func CreateEtcdBackupStorageContainer(rawKluster []byte, current *v1.Kluster, clients config.Clients, factories config.Factories) (err error) {\n+ secret, err := clients.Kubernetes.CoreV1().Secrets(current.GetNamespace()).Get(current.GetName(), metav1.GetOptions{})\nif err != nil {\nreturn err\n}\n@@ -27,7 +26,7 @@ func CreateEtcdBackupStorageContainer(rawKluster []byte, current *v1.Kluster, cl\nreturn errors.New(\"openstack domain name secret not set\")\n}\n- adminClient, err := openstackFactory.AdminClient()\n+ adminClient, err := factories.Openstack.AdminClient()\nif err != nil {\nreturn err\n}\n"
},
{
"change_type": "MODIFY",
"old_path": "pkg/migration/4_migrate_kluster_secret.go",
"new_path": "pkg/migration/4_migrate_kluster_secret.go",
"diff": "@@ -2,29 +2,28 @@ package migration\nimport (\nmeta_v1 \"k8s.io/apimachinery/pkg/apis/meta/v1\"\n- \"k8s.io/client-go/kubernetes\"\nv1 \"github.com/sapcc/kubernikus/pkg/apis/kubernikus/v1\"\n- \"github.com/sapcc/kubernikus/pkg/client/openstack\"\n+ \"github.com/sapcc/kubernikus/pkg/controller/config\"\n\"github.com/sapcc/kubernikus/pkg/util\"\n)\n-func MigrateKlusterSecret(rawKluster []byte, current *v1.Kluster, client kubernetes.Interface, openstackFactory openstack.SharedOpenstackClientFactory) (err error) {\n+func MigrateKlusterSecret(rawKluster []byte, current *v1.Kluster, clients config.Clients, factories config.Factories) (err error) {\n- oldSecret, err := client.CoreV1().Secrets(current.Namespace).Get(current.Name, meta_v1.GetOptions{})\n+ oldSecret, err := clients.Kubernetes.CoreV1().Secrets(current.Namespace).Get(current.Name, meta_v1.GetOptions{})\nif err != nil {\nreturn err\n}\n- if _, err := util.EnsureKlusterSecret(client, current); err != nil {\n+ if _, err := util.EnsureKlusterSecret(clients.Kubernetes, current); err != nil {\nreturn err\n}\n- newSecret, err := client.CoreV1().Secrets(current.Namespace).Get(current.Name+\"-secret\", meta_v1.GetOptions{})\n+ newSecret, err := clients.Kubernetes.CoreV1().Secrets(current.Namespace).Get(current.Name+\"-secret\", meta_v1.GetOptions{})\nif err != nil {\nreturn err\n}\nnewSecret.Data = oldSecret.Data\n- _, err = client.CoreV1().Secrets(current.Namespace).Update(newSecret)\n+ _, err = clients.Kubernetes.CoreV1().Secrets(current.Namespace).Update(newSecret)\nreturn err\n}\n"
},
{
"change_type": "MODIFY",
"old_path": "pkg/migration/5_insert_avz_into_nodepools.go",
"new_path": "pkg/migration/5_insert_avz_into_nodepools.go",
"diff": "@@ -3,15 +3,13 @@ package migration\nimport (\n\"fmt\"\n- \"k8s.io/client-go/kubernetes\"\n-\nv1 \"github.com/sapcc/kubernikus/pkg/apis/kubernikus/v1\"\n- \"github.com/sapcc/kubernikus/pkg/client/openstack\"\n+ \"github.com/sapcc/kubernikus/pkg/controller/config\"\n\"github.com/sapcc/kubernikus/pkg/util\"\n)\n-func InsertAVZIntoNodePools(rawKluster []byte, current *v1.Kluster, client kubernetes.Interface, openstackFactory openstack.SharedOpenstackClientFactory) (err error) {\n- secret, err := util.KlusterSecret(client, current)\n+func InsertAVZIntoNodePools(rawKluster []byte, current *v1.Kluster, clients config.Clients, factories config.Factories) (err error) {\n+ secret, err := util.KlusterSecret(clients.Kubernetes, current)\nif err != nil {\nreturn err\n}\n"
},
{
"change_type": "MODIFY",
"old_path": "pkg/migration/migration.go",
"new_path": "pkg/migration/migration.go",
"diff": "@@ -3,11 +3,8 @@ package migration\nimport (\n\"fmt\"\n- \"k8s.io/client-go/kubernetes\"\n-\n- \"github.com/sapcc/kubernikus/pkg/apis/kubernikus/v1\"\n- \"github.com/sapcc/kubernikus/pkg/client/openstack\"\n- kubernikus \"github.com/sapcc/kubernikus/pkg/generated/clientset\"\n+ v1 \"github.com/sapcc/kubernikus/pkg/apis/kubernikus/v1\"\n+ \"github.com/sapcc/kubernikus/pkg/controller/config\"\nkubernikusfake \"github.com/sapcc/kubernikus/pkg/generated/clientset/fake\"\n)\n@@ -18,7 +15,7 @@ var defaultRegistry Registry\n// The function is expected to modify the kluster accordingly, changed object is persisted\n// automatically after the handler returns with no error.\n// The kubernetes client can be used to modify other things (e.g. kluster secret)\n-type Migration func(klusterRaw []byte, kluster *v1.Kluster, client kubernetes.Interface, openstack_factory openstack.SharedOpenstackClientFactory) (err error)\n+type Migration func(klusterRaw []byte, kluster *v1.Kluster, clients config.Clients, factories config.Factories) (err error)\n//Latest returns to latest spec version available\nfunc Latest() int {\n@@ -31,8 +28,8 @@ func MigrationsPending(kluster *v1.Kluster) bool {\n}\n//Migrate a kluster to the most recent spec version\n-func Migrate(k *v1.Kluster, client kubernetes.Interface, kubernikus_client kubernikus.Interface, openstack_factory openstack.SharedOpenstackClientFactory) error {\n- return defaultRegistry.Migrate(k, client, kubernikus_client, openstack_factory)\n+func Migrate(k *v1.Kluster, clients config.Clients, factories config.Factories) error {\n+ return defaultRegistry.Migrate(k, clients, factories)\n}\n//Registry manages an ordered list of migration steps\n@@ -53,7 +50,7 @@ func (r Registry) MigrationsPending(kluster *v1.Kluster) bool {\nreturn int(kluster.Status.SpecVersion) < r.Latest()\n}\n-func (r *Registry) Migrate(k *v1.Kluster, client kubernetes.Interface, kubernikus_client kubernikus.Interface, openstack_factory openstack.SharedOpenstackClientFactory) error {\n+func (r *Registry) Migrate(k *v1.Kluster, clients config.Clients, factories config.Factories) error {\nklusterVersion := int(k.Status.SpecVersion)\nif klusterVersion >= r.Latest() {\nreturn nil\n@@ -64,28 +61,28 @@ func (r *Registry) Migrate(k *v1.Kluster, client kubernetes.Interface, kuberniku\nfor idx := klusterVersion; idx < r.Latest(); idx++ {\nmigration := r.migrations[idx]\nversion := idx + 1\n- if kluster, err = migrateKluster(kluster, version, migration, client, kubernikus_client, openstack_factory); err != nil {\n+ if kluster, err = migrateKluster(kluster, version, migration, clients, factories); err != nil {\nreturn fmt.Errorf(\"Error running migration %d: %s\", version, err)\n}\n}\nreturn nil\n}\n-func migrateKluster(kluster *v1.Kluster, version int, migration Migration, client kubernetes.Interface, kubernikus_client kubernikus.Interface, openstack_factory openstack.SharedOpenstackClientFactory) (*v1.Kluster, error) {\n+func migrateKluster(kluster *v1.Kluster, version int, migration Migration, clients config.Clients, factories config.Factories) (*v1.Kluster, error) {\nvar rawData []byte\nvar err error\n//TODO: Don't import fake pkg outside of test code\n- if _, ok := kubernikus_client.(*kubernikusfake.Clientset); !ok {\n- request := kubernikus_client.Kubernikus().RESTClient().Get().Namespace(kluster.Namespace).Resource(\"klusters\").Name(kluster.Name)\n+ if _, ok := clients.Kubernikus.(*kubernikusfake.Clientset); !ok {\n+ request := clients.Kubernikus.Kubernikus().RESTClient().Get().Namespace(kluster.Namespace).Resource(\"klusters\").Name(kluster.Name)\nif rawData, err = request.DoRaw(); err != nil {\nreturn nil, err\n}\n}\n- if err = migration(rawData, kluster, client, openstack_factory); err != nil {\n+ if err = migration(rawData, kluster, clients, factories); err != nil {\nreturn nil, err\n}\nkluster.Status.SpecVersion = int64(version)\n- return kubernikus_client.Kubernikus().Klusters(kluster.Namespace).Update(kluster)\n+ return clients.Kubernikus.Kubernikus().Klusters(kluster.Namespace).Update(kluster)\n}\n"
},
{
"change_type": "MODIFY",
"old_path": "pkg/migration/migration_test.go",
"new_path": "pkg/migration/migration_test.go",
"diff": "@@ -6,36 +6,37 @@ import (\n\"github.com/stretchr/testify/assert\"\nmetav1 \"k8s.io/apimachinery/pkg/apis/meta/v1\"\n- \"k8s.io/client-go/kubernetes\"\n\"k8s.io/client-go/kubernetes/fake\"\n\"github.com/sapcc/kubernikus/pkg/api/models\"\n- \"github.com/sapcc/kubernikus/pkg/apis/kubernikus/v1\"\n- \"github.com/sapcc/kubernikus/pkg/client/openstack\"\n+ v1 \"github.com/sapcc/kubernikus/pkg/apis/kubernikus/v1\"\n+ \"github.com/sapcc/kubernikus/pkg/controller/config\"\nkubernikusfake \"github.com/sapcc/kubernikus/pkg/generated/clientset/fake\"\n)\nconst NAMESPACE = \"test\"\nfunc TestInitialMigration(t *testing.T) {\n-\nkluster := &v1.Kluster{\nObjectMeta: metav1.ObjectMeta{\nNamespace: NAMESPACE,\nName: \"test\",\n},\n}\n- cs := fake.NewSimpleClientset()\n- kcs := kubernikusfake.NewSimpleClientset(kluster)\n+\n+ clients := config.Clients{\n+ Kubernetes: fake.NewSimpleClientset(),\n+ Kubernikus: kubernikusfake.NewSimpleClientset(kluster),\n+ }\nvar registry Registry\n- registry.AddMigration(func(_ []byte, kluster *v1.Kluster, _ kubernetes.Interface, _ openstack.SharedOpenstackClientFactory) error {\n+ registry.AddMigration(func(_ []byte, kluster *v1.Kluster, _ config.Clients, _ config.Factories) error {\nkluster.Spec.Name = \"executed\"\nreturn nil\n})\n- if assert.NoError(t, registry.Migrate(kluster, cs, kcs, nil)) {\n- kluster, _ = kcs.Kubernikus().Klusters(NAMESPACE).Get(\"test\", metav1.GetOptions{})\n+ if assert.NoError(t, registry.Migrate(kluster, clients, config.Factories{})) {\n+ kluster, _ = clients.Kubernikus.Kubernikus().Klusters(NAMESPACE).Get(\"test\", metav1.GetOptions{})\nassert.Equal(t, 1, int(kluster.Status.SpecVersion))\nassert.Equal(t, \"executed\", kluster.Spec.Name)\n}\n@@ -52,27 +53,30 @@ func TestMigration(t *testing.T) {\nSpecVersion: 1,\n},\n}\n- cs := fake.NewSimpleClientset()\n- kcs := kubernikusfake.NewSimpleClientset(kluster)\n+\n+ clients := config.Clients{\n+ Kubernetes: fake.NewSimpleClientset(),\n+ Kubernikus: kubernikusfake.NewSimpleClientset(kluster),\n+ }\nvar registry Registry\n- registry.AddMigration(func(_ []byte, kluster *v1.Kluster, _ kubernetes.Interface, _ openstack.SharedOpenstackClientFactory) error {\n+ registry.AddMigration(func(_ []byte, kluster *v1.Kluster, _ config.Clients, _ config.Factories) error {\nt.Error(\"First migration should be skipped\")\nreturn nil\n})\n- registry.AddMigration(func(_ []byte, kluster *v1.Kluster, _ kubernetes.Interface, _ openstack.SharedOpenstackClientFactory) error {\n+ registry.AddMigration(func(_ []byte, kluster *v1.Kluster, _ config.Clients, _ config.Factories) error {\nkluster.Spec.Name = kluster.Spec.Name + \"2\"\nreturn nil\n})\n- registry.AddMigration(func(_ []byte, kluster *v1.Kluster, _ kubernetes.Interface, _ openstack.SharedOpenstackClientFactory) error {\n+ registry.AddMigration(func(_ []byte, kluster *v1.Kluster, _ config.Clients, _ config.Factories) error {\nkluster.Spec.Name = kluster.Spec.Name + \"3\"\nreturn nil\n})\n- if assert.NoError(t, registry.Migrate(kluster, cs, kcs, nil)) {\n- kluster, _ = kcs.Kubernikus().Klusters(NAMESPACE).Get(\"test\", metav1.GetOptions{})\n+ if assert.NoError(t, registry.Migrate(kluster, clients, config.Factories{})) {\n+ kluster, _ = clients.Kubernikus.Kubernikus().Klusters(NAMESPACE).Get(\"test\", metav1.GetOptions{})\nassert.Equal(t, 3, int(kluster.Status.SpecVersion))\nassert.Equal(t, \"23\", kluster.Spec.Name)\n}\n@@ -88,17 +92,20 @@ func TestMigrationError(t *testing.T) {\nName: \"Before\",\n},\n}\n- cs := fake.NewSimpleClientset()\n- kcs := kubernikusfake.NewSimpleClientset(kluster)\n+\n+ clients := config.Clients{\n+ Kubernetes: fake.NewSimpleClientset(),\n+ Kubernikus: kubernikusfake.NewSimpleClientset(kluster),\n+ }\nvar registry Registry\n- registry.AddMigration(func(_ []byte, kluster *v1.Kluster, _ kubernetes.Interface, _ openstack.SharedOpenstackClientFactory) error {\n+ registry.AddMigration(func(_ []byte, kluster *v1.Kluster, _ config.Clients, _ config.Factories) error {\nkluster.Spec.Name = \"After\"\nreturn errors.New(\"migration failed\")\n})\n- if assert.Error(t, registry.Migrate(kluster, cs, kcs, nil)) {\n- kluster, _ = kcs.Kubernikus().Klusters(NAMESPACE).Get(\"test\", metav1.GetOptions{})\n+ if assert.Error(t, registry.Migrate(kluster, clients, config.Factories{})) {\n+ kluster, _ = clients.Kubernikus.Kubernikus().Klusters(NAMESPACE).Get(\"test\", metav1.GetOptions{})\nassert.Equal(t, 0, int(kluster.Status.SpecVersion))\nassert.Equal(t, \"Before\", kluster.Spec.Name)\n}\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
Generalize Migrations (#386)
This commit makes the migration mechanism more flexible. It passes the complete `config.Clients` and `config.Factories` structs.
This allows to use the full functionality of all clients. Concrete use is the seeding of StorageClasses or RBAC rules to already existing clusters.
|
596,240 |
20.02.2019 13:04:13
| -3,600 |
3431f21848ddc136c4e31b6d84d16fb0a759564e
|
adds migration to seed multi-avz storage classes to existing clusters
|
[
{
"change_type": "ADD",
"old_path": null,
"new_path": "pkg/migration/6_seed_storage_classes.go",
"diff": "+package migration\n+\n+import (\n+ v1 \"github.com/sapcc/kubernikus/pkg/apis/kubernikus/v1\"\n+ \"github.com/sapcc/kubernikus/pkg/controller/config\"\n+ \"github.com/sapcc/kubernikus/pkg/controller/ground\"\n+)\n+\n+func SeedCinderStorageClasses(rawKluster []byte, current *v1.Kluster, clients config.Clients, factories config.Factories) (err error) {\n+ kubernetes, err := clients.Satellites.ClientFor(current)\n+ if err != nil {\n+ return err\n+ }\n+\n+ openstack, err := factories.Openstack.ProjectAdminClientFor(current.Spec.Openstack.ProjectID)\n+ if err != nil {\n+ return err\n+ }\n+\n+ return ground.SeedCinderStorageClasses(kubernetes, openstack)\n+}\n"
},
{
"change_type": "MODIFY",
"old_path": "pkg/migration/register.go",
"new_path": "pkg/migration/register.go",
"diff": "@@ -12,6 +12,7 @@ func init() {\nCreateEtcdBackupStorageContainer,\nMigrateKlusterSecret,\nInsertAVZIntoNodePools,\n+ SeedCinderStorageClasses,\n// <-- Insert new migrations at the end only!\n}\n}\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
adds migration to seed multi-avz storage classes to existing clusters (#387)
|
596,240 |
26.02.2019 17:51:43
| -3,600 |
b547f815737375a0d920a80fae8aacfd4a3b0558
|
do not try to delete pending pvcs
|
[
{
"change_type": "MODIFY",
"old_path": "pkg/controller/deorbit/deorbiter.go",
"new_path": "pkg/controller/deorbit/deorbiter.go",
"diff": "@@ -89,6 +89,10 @@ func (d *ConcreteDeorbiter) DeletePersistentVolumeClaims() (deleted []core_v1.Pe\n}\nfor _, pvc := range pvcs.Items {\n+ if pvc.Status.Phase != core_v1.ClaimBound || pvc.Spec.VolumeName == \"\" {\n+ continue\n+ }\n+\npv, err := d.Client.Core().PersistentVolumes().Get(pvc.Spec.VolumeName, meta_v1.GetOptions{})\nif err != nil {\nreturn deleted, err\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
do not try to delete pending pvcs
|
596,240 |
26.02.2019 18:10:02
| -3,600 |
031418d0aea7e4fbc4d59f1e710c7347656e2b30
|
extends and fixes deorbiting for pending pvcs
|
[
{
"change_type": "MODIFY",
"old_path": "pkg/controller/deorbit/deorbiter_test.go",
"new_path": "pkg/controller/deorbit/deorbiter_test.go",
"diff": "@@ -67,6 +67,9 @@ var (\nSpec: core_v1.PersistentVolumeClaimSpec{\nVolumeName: \"pv-cinder0\",\n},\n+ Status: core_v1.PersistentVolumeClaimStatus{\n+ Phase: core_v1.ClaimBound,\n+ },\n}\npvcCinder1 = &core_v1.PersistentVolumeClaim{\n@@ -76,6 +79,19 @@ var (\nSpec: core_v1.PersistentVolumeClaimSpec{\nVolumeName: \"pv-cinder1\",\n},\n+ Status: core_v1.PersistentVolumeClaimStatus{\n+ Phase: core_v1.ClaimBound,\n+ },\n+ }\n+\n+ pvcCinder2 = &core_v1.PersistentVolumeClaim{\n+ ObjectMeta: meta_v1.ObjectMeta{\n+ Name: \"pvc-cinder2\",\n+ },\n+ Spec: core_v1.PersistentVolumeClaimSpec{},\n+ Status: core_v1.PersistentVolumeClaimStatus{\n+ Phase: core_v1.ClaimPending,\n+ },\n}\npvcNFS = &core_v1.PersistentVolumeClaim{\n@@ -85,6 +101,9 @@ var (\nSpec: core_v1.PersistentVolumeClaimSpec{\nVolumeName: \"pv-nfs\",\n},\n+ Status: core_v1.PersistentVolumeClaimStatus{\n+ Phase: core_v1.ClaimBound,\n+ },\n}\nsvcLB0 = &core_v1.Service{\n@@ -176,8 +195,8 @@ func TestDeletePersistentVolumeClaims(testing *testing.T) {\n}\nfor i, t := range []test{\n- {\"deletes all Cinder PVs\", 0, 2, []runtime.Object{pvCinder0, pvCinder1, pvcCinder0, pvcCinder1}},\n- {\"deletes only Cinder PVs\", 1, 2, []runtime.Object{pvCinder0, pvCinder1, pvNFS, pvcCinder0, pvcCinder1, pvcNFS}},\n+ {\"deletes all Cinder PVs\", 1, 2, []runtime.Object{pvCinder0, pvCinder1, pvcCinder0, pvcCinder1, pvcCinder2}},\n+ {\"deletes only Cinder PVs\", 2, 2, []runtime.Object{pvCinder0, pvCinder1, pvNFS, pvcCinder0, pvcCinder1, pvcCinder2, pvcNFS}},\n} {\ndone := make(chan struct{})\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
extends and fixes deorbiting for pending pvcs
|
596,226 |
28.02.2019 10:37:07
| -3,600 |
bfb2766a58adbcc9a89e4098eebc5a98c8ee031e
|
Do not show secret, simplify credential validation
|
[
{
"change_type": "MODIFY",
"old_path": "pkg/cmd/kubernikusctl/common/openstack.go",
"new_path": "pkg/cmd/kubernikusctl/common/openstack.go",
"diff": "@@ -46,7 +46,7 @@ func (o *OpenstackClient) BindFlags(flags *pflag.FlagSet) {\nflags.StringVar(&o.Scope.DomainName, \"project-domain-name\", o.Scope.DomainName, \"Domain of the project [OS_PROJECT_DOMAIN_NAME]\")\nflags.StringVar(&o.ApplicationCredentialName, \"application-credential-name\", o.ApplicationCredentialName, \"Project application credential name [OS_APPLICATION_CREDENTIAL_NAME]\")\nflags.StringVar(&o.ApplicationCredentialID, \"application-credential-id\", o.ApplicationCredentialName, \"Project application credential id [OS_APPLICATION_CREDENTIAL_ID]\")\n- flags.StringVar(&o.ApplicationCredentialSecret, \"application-credential-secret\", o.ApplicationCredentialSecret, \"Project application credential secret [OS_APPLICATION_CREDENTIAL_SECRET]\")\n+ flags.StringVar(&o.ApplicationCredentialSecret, \"application-credential-secret\", \"\", \"Project application credential secret [OS_APPLICATION_CREDENTIAL_SECRET]\")\n}\nfunc (o *OpenstackClient) Validate(c *cobra.Command, args []string) error {\n@@ -60,13 +60,13 @@ func (o *OpenstackClient) Validate(c *cobra.Command, args []string) error {\nif o.ApplicationCredentialID == \"\" {\no.ApplicationCredentialID = os.Getenv(\"OS_APPLICATION_CREDENTIAL_ID\")\n- if o.ApplicationCredentialID == \"\" {\n+ }\n+ if o.ApplicationCredentialName == \"\" {\no.ApplicationCredentialName = os.Getenv(\"OS_APPLICATION_CREDENTIAL_NAME\")\n}\nif o.ApplicationCredentialSecret == \"\" {\no.ApplicationCredentialSecret = os.Getenv(\"OS_APPLICATION_CREDENTIAL_SECRET\")\n}\n- }\n//Only use environment variables if nothing was given on the command line\nif o.Username == \"\" && o.UserID == \"\" {\n@@ -138,10 +138,6 @@ func (o *OpenstackClient) Complete(args []string) error {\nfunc (o *OpenstackClient) Setup() error {\nvar err error\n- if o.ApplicationCredentialSecret == \"\" && os.Getenv(\"OS_APPLICATION_CREDENTIAL_SECRET\") != \"\" {\n- o.ApplicationCredentialSecret = os.Getenv(\"OS_APPLICATION_CREDENTIAL_SECRET\")\n- }\n-\nif o.Password == \"\" && o.ApplicationCredentialSecret == \"\" {\nif os.Getenv(\"OS_PASSWORD\") != \"\" {\no.Password = os.Getenv(\"OS_PASSWORD\")\n@@ -183,10 +179,10 @@ func (o *OpenstackClient) PrintAuthInfo() string {\n}\n}\n- if o.ApplicationCredentialName != \"\" {\n- return fmt.Sprintf(\"Authenticating with application credential %v (%v)\", o.ApplicationCredentialName, user)\n- } else if o.ApplicationCredentialID != \"\" {\n+ if o.ApplicationCredentialID != \"\" {\nreturn fmt.Sprintf(\"Authenticating with application credential %v\", o.ApplicationCredentialID)\n+ } else if o.ApplicationCredentialName != \"\" {\n+ return fmt.Sprintf(\"Authenticating with application credential %v (%v)\", o.ApplicationCredentialName, user)\n}\nif o.Scope.ProjectID != \"\" {\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
Do not show secret, simplify credential validation
|
596,240 |
06.03.2019 15:03:47
| -3,600 |
a510a1a9ab7c0be5fc2581df22d91087742c8d42
|
fix gpu daemonsets
|
[
{
"change_type": "MODIFY",
"old_path": "pkg/controller/ground/bootstrap/gpu/manifest.go",
"new_path": "pkg/controller/ground/bootstrap/gpu/manifest.go",
"diff": "@@ -42,20 +42,19 @@ spec:\nhostPath:\npath: /dev\ncontainers:\n- - image: \"k8s.gcr.io/nvidia-gpu-device-plugin@sha256:0842734032018be107fa2490c98156992911e3e1f2a21e059ff0105b07dd8e9e\"\n+ - image: \"k8s.gcr.io/nvidia-gpu-device-plugin@sha256:d18b678437fedc4ec4211c20b3e5469a137a44f989da43dc275e4f2678170db4\"\ncommand:\n- /usr/bin/nvidia-gpu-device-plugin\n- -logtostderr\n- - -host-path=/opt/nvidia/current/lib64\n- - -container-path=/usr/local/nvidia/lib64\nname: nvidia-gpu-device-plugin\nresources:\nrequests:\ncpu: 50m\n- memory: 10Mi\n+ memory: 50Mi\nlimits:\ncpu: 50m\n- memory: 10Mi\n+ memory: 50Mi\n+ terminationGracePeriodSeconds: 0\nsecurityContext:\nprivileged: true\nvolumeMounts:\n@@ -64,7 +63,7 @@ spec:\n- name: dev\nmountPath: /dev\nupdateStrategy:\n- type: RollingUpdate\n+ type: OnDelete\n`\nNVIDIADriverInstaller_v20180808 = `\n@@ -104,19 +103,23 @@ spec:\n- name: rootfs\nhostPath:\npath: /\n+ - name: dev\n+ hostPath:\n+ path: /dev\ninitContainers:\n- image: bugroger/coreos-nvidia-driver:stable-396.44-tesla\nname: nvidia-driver-installer\nimagePullPolicy: Always\n- resources:\n- requests:\n- cpu: 0.15\n+ terminationGracePeriodSeconds: 0\nsecurityContext:\nprivileged: true\nvolumeMounts:\n- name: rootfs\nmountPath: /root\nmountPropagation: Bidirectional\n+ - name: dev\n+ mountPath: /dev\n+ mountPropagation: Bidirectional\ncontainers:\n- image: \"gcr.io/google-containers/pause:2.0\"\nname: pause\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
fix gpu daemonsets
|
596,240 |
12.03.2019 11:25:52
| -3,600 |
564aaf96e93adf0a8e7325a64f6dbedc97943a21
|
fixes path mapping for device-plugin
|
[
{
"change_type": "MODIFY",
"old_path": "pkg/controller/ground/bootstrap/gpu/manifest.go",
"new_path": "pkg/controller/ground/bootstrap/gpu/manifest.go",
"diff": "@@ -45,6 +45,7 @@ spec:\n- image: \"k8s.gcr.io/nvidia-gpu-device-plugin@sha256:d18b678437fedc4ec4211c20b3e5469a137a44f989da43dc275e4f2678170db4\"\ncommand:\n- /usr/bin/nvidia-gpu-device-plugin\n+ - -host-path=/opt/nvidia/current\n- -logtostderr\nname: nvidia-gpu-device-plugin\nresources:\n@@ -111,6 +112,7 @@ spec:\nname: nvidia-driver-installer\nimagePullPolicy: Always\nterminationGracePeriodSeconds: 0\n+ terminationMessagePath: /tmp/termination-log\nsecurityContext:\nprivileged: true\nvolumeMounts:\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
fixes path mapping for device-plugin
|
596,230 |
25.03.2019 13:53:56
| -3,600 |
1853fd4bfa0522b2d72759f65345a4dac5b31b36
|
add multus and dedup bridge to hyperkube image
|
[
{
"change_type": "MODIFY",
"old_path": "contrib/hyperkube/Makefile",
"new_path": "contrib/hyperkube/Makefile",
"diff": "@@ -3,6 +3,8 @@ IMAGE?=sapcc/hyperkube\nARCH=amd64\nCNI_RELEASE=v0.6.0\nCNI_PLUGINS_RELEASE=v0.7.1\n+CNI_DEDUP_BRIDGE_RELEASE=v0.1.0\n+MULTUS_RELEASE=v3.1\nBASEIMAGE=k8s.gcr.io/hyperkube-$(ARCH):$(VERSION)\nTEMP_DIR:=$(shell mktemp -d -t hyperkubeXXXXXX)\n@@ -22,5 +24,8 @@ cni:\nmkdir -p ${TEMP_DIR}/cni-bin/bin\ncurl -sSL --retry 5 https://github.com/containernetworking/cni/releases/download/${CNI_RELEASE}/cni-${ARCH}-${CNI_RELEASE}.tgz | tar -xz -C ${TEMP_DIR}/cni-bin/bin\ncurl -sSL --retry 5 https://github.com/containernetworking/plugins/releases/download/${CNI_PLUGINS_RELEASE}/cni-plugins-${ARCH}-${CNI_PLUGINS_RELEASE}.tgz | tar -xz -C ${TEMP_DIR}/cni-bin/bin\n+ curl -sSL --retry 5 https://github.com/intel/multus-cni/releases/download/${MULTUS_RELEASE}/multus-cni_${MULTUS_RELEASE}_linux_amd64.tar.gz | tar --strip 1 -xz -C ${TEMP_DIR}/cni-bin/bin\n+ curl -o ${TEMP_DIR}/cni-bin/bin/dedup-bridge -sSL --retry 5 https://github.com/sapcc/cni-dedup-bridge/releases/download/${CNI_DEDUP_BRIDGE_RELEASE}/dedup-bridge\n+ chmod +x ${TEMP_DIR}/cni-bin/bin/dedup-bridge\n.PHONY: build push all cni\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
-- add multus and dedup bridge to hyperkube image (#403)
|
596,233 |
26.03.2019 10:30:22
| -3,600 |
8b66e9e38b6d489b537dd2c038249da438b2e2eb
|
update vice president
|
[
{
"change_type": "MODIFY",
"old_path": "charts/kubernikus-system/requirements.lock",
"new_path": "charts/kubernikus-system/requirements.lock",
"diff": "dependencies:\n- name: vice-president\nrepository: file://vendor/vice-president\n- version: 0.1.3\n+ version: 0.1.4\n- name: kube-state-metrics\nrepository: https://kubernetes-charts.storage.googleapis.com/\nversion: 0.9.0\n@@ -23,5 +23,5 @@ dependencies:\n- name: fluent-bit\nrepository: https://kubernetes-charts.storage.googleapis.com/\nversion: 1.2.0\n-digest: sha256:8d69e427b1f85b68ae97cf1ffbd6f563bf6f5642667cd00b104bbdd02d67007b\n-generated: 2018-12-21T14:42:09.575053176+01:00\n+digest: sha256:a6beeb2c17ff46f30233cebf5ba0852b126972b6d2afdb3bc32ec5b9be830f04\n+generated: 2019-03-21T17:17:22.444884166+01:00\n"
},
{
"change_type": "MODIFY",
"old_path": "charts/kubernikus-system/requirements.yaml",
"new_path": "charts/kubernikus-system/requirements.yaml",
"diff": "dependencies:\n- name: vice-president\nrepository: file://vendor/vice-president\n- version: 0.1.3\n+ version: 0.1.4\n- name: kube-state-metrics\nrepository: https://kubernetes-charts.storage.googleapis.com/\nversion: 0.9.0\n"
},
{
"change_type": "MODIFY",
"old_path": "charts/kubernikus-system/vendor/vice-president/Chart.yaml",
"new_path": "charts/kubernikus-system/vendor/vice-president/Chart.yaml",
"diff": "apiVersion: v1\ndescription: Automated certificate management by a kubernetes operator using the Symantec VICE API\nname: vice-president\n-version: 0.1.3\n+version: 0.1.4\n"
},
{
"change_type": "MODIFY",
"old_path": "charts/kubernikus-system/vendor/vice-president/templates/deployment.yaml",
"new_path": "charts/kubernikus-system/vendor/vice-president/templates/deployment.yaml",
"diff": "@@ -27,8 +27,11 @@ spec:\n- --vice-cert=/etc/vice-president/secrets/vice.cert\n- --vice-key=/etc/vice-president/secrets/vice.key\n- --intermediate-cert=/etc/vice-president/secrets/intermediate.cert\n- - --metric-port={{default 9091 .Values.president.metrics_port}}\n- - --v={{ default 1 .Values.president.log_level }}\n+ - --metric-port={{default 9091 .Values.president.metricsPort}}\n+ - --min-cert-validity-days={{ .Values.president.minCertValidityDays }}\n+ - --enable-validate-remote-cert={{ .Values.president.enableValidateRemoteCert }}\n+ - --enable-symantec-metrics={{ .Values.president.enableSymantecMetrics }}\n+ - --v={{ default 1 .Values.president.logLevel }}\nvolumeMounts:\n- name: config\nmountPath: /etc/vice-president/config\n"
},
{
"change_type": "MODIFY",
"old_path": "charts/kubernikus-system/vendor/vice-president/values.yaml",
"new_path": "charts/kubernikus-system/vendor/vice-president/values.yaml",
"diff": "image:\nrepository: sapcc/vice-president\n- tag: v201808161913\n+ tag: v201903181100\npullPolicy: IfNotPresent\n-# config\n+# Symantec/DigiCert VICE API configuration.\nvice:\nfirst_name: \"\"\nlast_name: \"\"\n@@ -15,7 +15,7 @@ vice:\norganizational_unit: \"\"\ndefault_challenge: \"\"\n- # symantec API SSO certificate and key\n+ # Symantec API SSO certificate and private key.\ncert: \"\"\nkey: \"\"\n@@ -49,15 +49,30 @@ vice:\nj6tJLp07kzQoH3jOlOrHvdPJbRzeXDLz\n-----END CERTIFICATE-----\n-# operator settings\n+# Operator configuration.\npresident:\n- resync_period_minutes: 12\n- certificate_check_interval_minutes: 30\n- rate_limit: 4\n- metrics_port: 9091\n- log_level: 1\n+ # Trigger certificate renewal if cert is valid for less than min_cert_validity_days.\n+ minCertValidityDays: 30\n+ # Verify the certificate via TLS handshake.\n+ # Potentially triggers additional certificate renewals or enrollments if the new certificate is not yet presented.\n+ # CertificateCheckIntervalMinutes should be set accordingly.\n+ enableValidateRemoteCert: false\n+ # Enable additional symantec metrics for token usage.\n+ enableSymantecMetrics: false\n+ # Resync ingress and secrets every n minutes.\n+ resyncPeriodMinutes: 4\n+ # Check certificates every n minutes.\n+ certificateCheckIntervalMinutes: 30\n+ # Rate limit for certificate enrollments, renewals, etc. per certificateCheckIntervalMinutes.\n+ rateLimit: 2\n+\n+ # Port to expose prometheus metrics on.\n+ metricsPort: 9091\n+\n+ # The log level.\n+ logLevel: 1\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
update vice president (#400)
|
596,240 |
02.04.2019 18:04:21
| -7,200 |
0573813f57ab3f062706e20b45b57591571592fb
|
slight metrics change
|
[
{
"change_type": "MODIFY",
"old_path": "pkg/controller/servicing/reconciler.go",
"new_path": "pkg/controller/servicing/reconciler.go",
"diff": "@@ -218,9 +218,14 @@ func (r *KlusterReconciler) isServiceIntervalElapsed() bool {\n}\nfunc (r *KlusterReconciler) collectMetrics() {\n+ reboot := float64(len(r.Lister.Reboot()))\n+ replace := float64(len(r.Lister.Replace()))\n+ uptodate := float64(len(r.Lister.All())) - reboot - replace\n+\nr.Status.With(prometheus.Labels{\"kluster_id\": r.Kluster.GetName(), \"action\": \"updating\", \"status\": \"started\"}).Set(float64(len(r.Lister.Updating())))\nr.Status.With(prometheus.Labels{\"kluster_id\": r.Kluster.GetName(), \"action\": \"updating\", \"status\": \"failed\"}).Set(float64(len(r.Lister.Failed())))\nr.Status.With(prometheus.Labels{\"kluster_id\": r.Kluster.GetName(), \"action\": \"updating\", \"status\": \"successful\"}).Set(float64(len(r.Lister.Successful())))\n- r.Status.With(prometheus.Labels{\"kluster_id\": r.Kluster.GetName(), \"action\": \"waiting\", \"status\": \"reboot\"}).Set(float64(len(r.Lister.Reboot())))\n- r.Status.With(prometheus.Labels{\"kluster_id\": r.Kluster.GetName(), \"action\": \"waiting\", \"status\": \"replace\"}).Set(float64(len(r.Lister.Replace())))\n+ r.Status.With(prometheus.Labels{\"kluster_id\": r.Kluster.GetName(), \"action\": \"waiting\", \"status\": \"reboot\"}).Set(reboot)\n+ r.Status.With(prometheus.Labels{\"kluster_id\": r.Kluster.GetName(), \"action\": \"waiting\", \"status\": \"replace\"}).Set(replace)\n+ r.Status.With(prometheus.Labels{\"kluster_id\": r.Kluster.GetName(), \"action\": \"waiting\", \"status\": \"uptodate\"}).Set(uptodate)\n}\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
slight metrics change
|
596,240 |
03.04.2019 10:01:28
| -7,200 |
2ea76be05a65ac7a43cc28f456f540a5ef1917a6
|
moves skip logic to reconciler in order to enable metrics
|
[
{
"change_type": "MODIFY",
"old_path": "pkg/controller/servicing/controller.go",
"new_path": "pkg/controller/servicing/controller.go",
"diff": "@@ -8,11 +8,9 @@ import (\n\"github.com/go-kit/kit/log\"\n\"k8s.io/client-go/tools/record\"\n- \"github.com/sapcc/kubernikus/pkg/api/models\"\nv1 \"github.com/sapcc/kubernikus/pkg/apis/kubernikus/v1\"\n\"github.com/sapcc/kubernikus/pkg/controller/base\"\n\"github.com/sapcc/kubernikus/pkg/controller/config\"\n- \"github.com/sapcc/kubernikus/pkg/util\"\n)\nconst (\n@@ -64,17 +62,6 @@ func NewController(threadiness int, factories config.Factories, clients config.C\n// Reconcile checks a kluster for node updates\nfunc (d *Controller) Reconcile(k *v1.Kluster) (requeue bool, err error) {\n- // Default to skip klusters without the servicing annotation\n- if !util.EnabledValue(k.ObjectMeta.Annotations[AnnotationServicingSafeguard]) {\n- d.Logger.Log(\"msg\", \"Skippig upgrades. Manually disabled with safeguard annotation.\")\n- return false, nil\n- }\n-\n- if k.Status.Phase != models.KlusterPhaseRunning {\n- d.Logger.Log(\"msg\", \"skipped upgrades because kluster is not running\", \"v\", 2)\n- return false, nil\n- }\n-\nreconciler, err := d.Reconciler.Make(k)\nif err != nil {\nd.Logger.Log(\"msg\", \"skippig upgrades. Internal server error.\", \"err\", err)\n"
},
{
"change_type": "MODIFY",
"old_path": "pkg/controller/servicing/reconciler.go",
"new_path": "pkg/controller/servicing/reconciler.go",
"diff": "@@ -126,6 +126,12 @@ func (r *KlusterReconciler) Do() error {\ndefer r.collectMetrics()\n+ // Default to skip klusters without the servicing annotation\n+ if !util.EnabledValue(r.Kluster.ObjectMeta.Annotations[AnnotationServicingSafeguard]) {\n+ r.Logger.Log(\"msg\", \"Skippig upgrades. Manually disabled with safeguard annotation.\")\n+ return nil\n+ }\n+\nfor _, node := range r.Lister.Successful() {\nif err := r.LifeCycler.Uncordon(node); err != nil {\nreturn errors.Wrap(err, \"Failed to uncordon successfully updated node.\")\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
moves skip logic to reconciler in order to enable metrics
|
596,240 |
03.04.2019 11:56:47
| -7,200 |
ac5f74826705fcb314f87091de1979fe00b880a0
|
improvements to slack resource
|
[
{
"change_type": "MODIFY",
"old_path": "ci/pipeline.yaml",
"new_path": "ci/pipeline.yaml",
"diff": "@@ -1054,9 +1054,9 @@ task_tiller: &task_tiller\nslack: &slack\n- on_success: { put: slack, params: {alert_type: fixed } }\n- on_failure: { put: slack, params: {alert_type: broke } }\n- on_abort: { put: slack, params: {alert_type: broke } }\n+ on_success: { put: slack, inputs: [], params: {alert_type: fixed } }\n+ on_failure: { put: slack, inputs: [], params: {alert_type: broke } }\n+ on_abort: { put: slack, inputs: [], params: {alert_type: broke } }\nresources:\n"
},
{
"change_type": "MODIFY",
"old_path": "ci/pipeline.yaml.erb",
"new_path": "ci/pipeline.yaml.erb",
"diff": "@@ -115,9 +115,9 @@ auth_e2e_<%= region %>: &auth_e2e_<%= region %>\n<% end %>\nslack: &slack\n- on_success: { put: slack, params: {alert_type: fixed } }\n- on_failure: { put: slack, params: {alert_type: broke } }\n- on_abort: { put: slack, params: {alert_type: broke } }\n+ on_success: { put: slack, inputs: [], params: {alert_type: fixed } }\n+ on_failure: { put: slack, inputs: [], params: {alert_type: broke } }\n+ on_abort: { put: slack, inputs: [], params: {alert_type: broke } }\nresources:\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
improvements to slack resource
|
596,240 |
03.04.2019 12:00:48
| -7,200 |
fd3b5e8214825530e869d7b575a37f2448062849
|
update dashboard to show updates
|
[
{
"change_type": "MODIFY",
"old_path": "charts/kubernikus-system/values.yaml",
"new_path": "charts/kubernikus-system/values.yaml",
"diff": "@@ -201,7 +201,7 @@ grafana:\n{\n\"__inputs\": [\n{\n- \"name\": \"default\",\n+ \"name\": \"DS_PROMETHEUS\",\n\"label\": \"Prometheus\",\n\"description\": \"\",\n\"type\": \"datasource\",\n@@ -216,6 +216,12 @@ grafana:\n\"name\": \"Grafana\",\n\"version\": \"5.0.4\"\n},\n+ {\n+ \"type\": \"panel\",\n+ \"id\": \"grafana-piechart-panel\",\n+ \"name\": \"Pie Chart\",\n+ \"version\": \"1.3.3\"\n+ },\n{\n\"type\": \"panel\",\n\"id\": \"graph\",\n@@ -251,7 +257,7 @@ grafana:\n\"list\": [\n{\n\"builtIn\": 1,\n- \"datasource\": \"default\",\n+ \"datasource\": \"${DS_PROMETHEUS}\",\n\"enable\": true,\n\"hide\": true,\n\"iconColor\": \"rgba(0, 211, 255, 1)\",\n@@ -264,7 +270,7 @@ grafana:\n\"gnetId\": null,\n\"graphTooltip\": 0,\n\"id\": null,\n- \"iteration\": 1543925725334,\n+ \"iteration\": 1554285339512,\n\"links\": [],\n\"panels\": [\n{\n@@ -319,7 +325,7 @@ grafana:\n\"rgba(237, 129, 40, 0.89)\",\n\"#d44a3a\"\n],\n- \"datasource\": null,\n+ \"datasource\": \"${DS_PROMETHEUS}\",\n\"decimals\": 0,\n\"format\": \"none\",\n\"gauge\": {\n@@ -404,7 +410,7 @@ grafana:\n\"rgba(237, 129, 40, 0.89)\",\n\"#d44a3a\"\n],\n- \"datasource\": null,\n+ \"datasource\": \"${DS_PROMETHEUS}\",\n\"decimals\": 0,\n\"format\": \"none\",\n\"gauge\": {\n@@ -489,7 +495,7 @@ grafana:\n\"rgba(237, 129, 40, 0.89)\",\n\"#d44a3a\"\n],\n- \"datasource\": null,\n+ \"datasource\": \"${DS_PROMETHEUS}\",\n\"decimals\": 0,\n\"format\": \"none\",\n\"gauge\": {\n@@ -574,7 +580,7 @@ grafana:\n\"rgba(237, 129, 40, 0.89)\",\n\"#d44a3a\"\n],\n- \"datasource\": null,\n+ \"datasource\": \"${DS_PROMETHEUS}\",\n\"decimals\": 0,\n\"format\": \"none\",\n\"gauge\": {\n@@ -688,7 +694,7 @@ grafana:\n\"bars\": false,\n\"dashLength\": 10,\n\"dashes\": false,\n- \"datasource\": null,\n+ \"datasource\": \"${DS_PROMETHEUS}\",\n\"fill\": 3,\n\"gridPos\": {\n\"h\": 4,\n@@ -791,7 +797,7 @@ grafana:\n\"bars\": false,\n\"dashLength\": 10,\n\"dashes\": false,\n- \"datasource\": null,\n+ \"datasource\": \"${DS_PROMETHEUS}\",\n\"fill\": 3,\n\"gridPos\": {\n\"h\": 4,\n@@ -903,7 +909,7 @@ grafana:\n}\n],\n\"crosshairColor\": \"#8F070C\",\n- \"datasource\": null,\n+ \"datasource\": \"${DS_PROMETHEUS}\",\n\"display\": \"timeline\",\n\"expandFromQueryS\": 86400,\n\"extendLastValue\": true,\n@@ -990,13 +996,212 @@ grafana:\n\"writeLastValue\": false,\n\"writeMetricNames\": true\n},\n+ {\n+ \"content\": \"\\n# Updates\\n\\n\",\n+ \"gridPos\": {\n+ \"h\": 4,\n+ \"w\": 2,\n+ \"x\": 1,\n+ \"y\": 5\n+ },\n+ \"id\": 78,\n+ \"links\": [],\n+ \"mode\": \"markdown\",\n+ \"title\": \"\",\n+ \"transparent\": true,\n+ \"type\": \"text\"\n+ },\n+ {\n+ \"aliasColors\": {},\n+ \"breakPoint\": \"100%\",\n+ \"cacheTimeout\": null,\n+ \"combine\": {\n+ \"label\": \"Others\",\n+ \"threshold\": 0\n+ },\n+ \"datasource\": \"${DS_PROMETHEUS}\",\n+ \"decimals\": 0,\n+ \"fontSize\": \"100%\",\n+ \"format\": \"none\",\n+ \"gridPos\": {\n+ \"h\": 4,\n+ \"w\": 3,\n+ \"x\": 3,\n+ \"y\": 5\n+ },\n+ \"hideTimeOverride\": true,\n+ \"id\": 75,\n+ \"interval\": null,\n+ \"legend\": {\n+ \"percentage\": false,\n+ \"percentageDecimals\": 0,\n+ \"show\": true,\n+ \"values\": true\n+ },\n+ \"legendType\": \"Right side\",\n+ \"links\": [],\n+ \"maxDataPoints\": 3,\n+ \"nullPointMode\": \"connected\",\n+ \"pieType\": \"pie\",\n+ \"strokeWidth\": 1,\n+ \"targets\": [\n+ {\n+ \"expr\": \"sum(kubernikus_servicing_status_nodes{action=\\\"waiting\\\"}) by (status)\",\n+ \"format\": \"time_series\",\n+ \"interval\": \"\",\n+ \"intervalFactor\": 1,\n+ \"legendFormat\": \"{{status}}\",\n+ \"refId\": \"A\"\n+ }\n+ ],\n+ \"timeFrom\": \"1m\",\n+ \"timeShift\": null,\n+ \"title\": \"Status by Node\",\n+ \"type\": \"grafana-piechart-panel\",\n+ \"valueName\": \"current\"\n+ },\n+ {\n+ \"aliasColors\": {},\n+ \"breakPoint\": \"100%\",\n+ \"cacheTimeout\": null,\n+ \"combine\": {\n+ \"label\": \"Others\",\n+ \"threshold\": 0\n+ },\n+ \"datasource\": \"${DS_PROMETHEUS}\",\n+ \"decimals\": 0,\n+ \"fontSize\": \"100%\",\n+ \"format\": \"none\",\n+ \"gridPos\": {\n+ \"h\": 4,\n+ \"w\": 3,\n+ \"x\": 6,\n+ \"y\": 5\n+ },\n+ \"hideTimeOverride\": true,\n+ \"id\": 76,\n+ \"interval\": null,\n+ \"legend\": {\n+ \"percentage\": false,\n+ \"percentageDecimals\": 0,\n+ \"show\": true,\n+ \"values\": true\n+ },\n+ \"legendType\": \"Right side\",\n+ \"links\": [],\n+ \"maxDataPoints\": 3,\n+ \"nullPointMode\": \"connected\",\n+ \"pieType\": \"pie\",\n+ \"strokeWidth\": 1,\n+ \"targets\": [\n+ {\n+ \"expr\": \"count(sum(kubernikus_servicing_status_nodes{action=\\\"waiting\\\"}) by (kluster_id,status) > 0) by (status)\",\n+ \"format\": \"time_series\",\n+ \"interval\": \"\",\n+ \"intervalFactor\": 1,\n+ \"legendFormat\": \"{{status}}\",\n+ \"refId\": \"A\"\n+ }\n+ ],\n+ \"timeFrom\": \"1m\",\n+ \"timeShift\": null,\n+ \"title\": \"Status by Kluster\",\n+ \"type\": \"grafana-piechart-panel\",\n+ \"valueName\": \"current\"\n+ },\n+ {\n+ \"cacheTimeout\": null,\n+ \"colorBackground\": true,\n+ \"colorValue\": false,\n+ \"colors\": [\n+ \"#299c46\",\n+ \"rgba(237, 129, 40, 0.89)\",\n+ \"#d44a3a\"\n+ ],\n+ \"datasource\": \"${DS_PROMETHEUS}\",\n+ \"decimals\": 0,\n+ \"format\": \"none\",\n+ \"gauge\": {\n+ \"maxValue\": 100,\n+ \"minValue\": 0,\n+ \"show\": false,\n+ \"thresholdLabels\": false,\n+ \"thresholdMarkers\": true\n+ },\n+ \"gridPos\": {\n+ \"h\": 4,\n+ \"w\": 2,\n+ \"x\": 9,\n+ \"y\": 5\n+ },\n+ \"hideTimeOverride\": true,\n+ \"id\": 77,\n+ \"interval\": null,\n+ \"links\": [],\n+ \"mappingType\": 1,\n+ \"mappingTypes\": [\n+ {\n+ \"name\": \"value to text\",\n+ \"value\": 1\n+ },\n+ {\n+ \"name\": \"range to text\",\n+ \"value\": 2\n+ }\n+ ],\n+ \"maxDataPoints\": 100,\n+ \"nullPointMode\": \"connected\",\n+ \"nullText\": null,\n+ \"postfix\": \"Nodes\",\n+ \"postfixFontSize\": \"50%\",\n+ \"prefix\": \"\",\n+ \"prefixFontSize\": \"50%\",\n+ \"rangeMaps\": [\n+ {\n+ \"from\": \"null\",\n+ \"text\": \"N/A\",\n+ \"to\": \"null\"\n+ }\n+ ],\n+ \"sparkline\": {\n+ \"fillColor\": \"rgba(31, 118, 189, 0.18)\",\n+ \"full\": false,\n+ \"lineColor\": \"rgb(31, 120, 193)\",\n+ \"show\": false\n+ },\n+ \"tableColumn\": \"\",\n+ \"targets\": [\n+ {\n+ \"expr\": \"sum(kubernikus_servicing_status_nodes{action=\\\"updating\\\", status=\\\"failed\\\"})\",\n+ \"format\": \"time_series\",\n+ \"intervalFactor\": 1,\n+ \"legendFormat\": \"\",\n+ \"refId\": \"A\"\n+ }\n+ ],\n+ \"thresholds\": \"1,1\",\n+ \"timeFrom\": \"5m\",\n+ \"timeShift\": null,\n+ \"title\": \"Failed\",\n+ \"transparent\": false,\n+ \"type\": \"singlestat\",\n+ \"valueFontSize\": \"200%\",\n+ \"valueMaps\": [\n+ {\n+ \"op\": \"=\",\n+ \"text\": \"N/A\",\n+ \"value\": \"null\"\n+ }\n+ ],\n+ \"valueName\": \"current\"\n+ },\n{\n\"collapsed\": false,\n\"gridPos\": {\n\"h\": 1,\n\"w\": 24,\n\"x\": 0,\n- \"y\": 5\n+ \"y\": 9\n},\n\"id\": 33,\n\"panels\": [],\n@@ -1009,7 +1214,7 @@ grafana:\n\"h\": 6,\n\"w\": 1,\n\"x\": 0,\n- \"y\": 6\n+ \"y\": 10\n},\n\"id\": 46,\n\"links\": [],\n@@ -1024,7 +1229,7 @@ grafana:\n\"h\": 3,\n\"w\": 2,\n\"x\": 1,\n- \"y\": 6\n+ \"y\": 10\n},\n\"id\": 27,\n\"links\": [],\n@@ -1039,7 +1244,7 @@ grafana:\n\"h\": 6,\n\"w\": 2,\n\"x\": 3,\n- \"y\": 6\n+ \"y\": 10\n},\n\"id\": 67,\n\"links\": [],\n@@ -1057,7 +1262,7 @@ grafana:\n\"rgba(237, 129, 40, 0.89)\",\n\"#299c46\"\n],\n- \"datasource\": null,\n+ \"datasource\": \"${DS_PROMETHEUS}\",\n\"decimals\": 0,\n\"description\": \"Out of CPU in this many Klusters\",\n\"format\": \"none\",\n@@ -1072,7 +1277,7 @@ grafana:\n\"h\": 3,\n\"w\": 2,\n\"x\": 5,\n- \"y\": 6\n+ \"y\": 10\n},\n\"hideTimeOverride\": true,\n\"id\": 8,\n@@ -1144,7 +1349,7 @@ grafana:\n\"rgba(237, 129, 40, 0.89)\",\n\"#299c46\"\n],\n- \"datasource\": null,\n+ \"datasource\": \"${DS_PROMETHEUS}\",\n\"decimals\": 0,\n\"description\": \"Out of Memory in this many Klusters\",\n\"format\": \"none\",\n@@ -1159,7 +1364,7 @@ grafana:\n\"h\": 3,\n\"w\": 2,\n\"x\": 7,\n- \"y\": 6\n+ \"y\": 10\n},\n\"hideTimeOverride\": true,\n\"id\": 9,\n@@ -1231,7 +1436,7 @@ grafana:\n\"rgba(237, 129, 40, 0.89)\",\n\"#299c46\"\n],\n- \"datasource\": null,\n+ \"datasource\": \"${DS_PROMETHEUS}\",\n\"decimals\": 0,\n\"description\": \"Out of Podsin this many Klusters\",\n\"format\": \"none\",\n@@ -1246,7 +1451,7 @@ grafana:\n\"h\": 3,\n\"w\": 2,\n\"x\": 9,\n- \"y\": 6\n+ \"y\": 10\n},\n\"hideTimeOverride\": true,\n\"id\": 10,\n@@ -1315,7 +1520,7 @@ grafana:\n\"h\": 6,\n\"w\": 1,\n\"x\": 11,\n- \"y\": 6\n+ \"y\": 10\n},\n\"id\": 47,\n\"links\": [],\n@@ -1330,7 +1535,7 @@ grafana:\n\"h\": 3,\n\"w\": 2,\n\"x\": 12,\n- \"y\": 6\n+ \"y\": 10\n},\n\"id\": 59,\n\"links\": [],\n@@ -1348,7 +1553,7 @@ grafana:\n\"rgba(237, 129, 40, 0.89)\",\n\"#299c46\"\n],\n- \"datasource\": null,\n+ \"datasource\": \"${DS_PROMETHEUS}\",\n\"decimals\": 0,\n\"description\": \"Out of Volumes in this many Klusters\",\n\"format\": \"none\",\n@@ -1363,7 +1568,7 @@ grafana:\n\"h\": 3,\n\"w\": 2,\n\"x\": 14,\n- \"y\": 6\n+ \"y\": 10\n},\n\"hideTimeOverride\": true,\n\"id\": 34,\n@@ -1435,7 +1640,7 @@ grafana:\n\"rgba(237, 129, 40, 0.89)\",\n\"#299c46\"\n],\n- \"datasource\": null,\n+ \"datasource\": \"${DS_PROMETHEUS}\",\n\"decimals\": 0,\n\"description\": \"Out of Volumes in this many Klusters\",\n\"format\": \"none\",\n@@ -1450,7 +1655,7 @@ grafana:\n\"h\": 3,\n\"w\": 2,\n\"x\": 16,\n- \"y\": 6\n+ \"y\": 10\n},\n\"hideTimeOverride\": true,\n\"id\": 55,\n@@ -1528,7 +1733,7 @@ grafana:\n\"rgba(237, 129, 40, 0.89)\",\n\"#299c46\"\n],\n- \"datasource\": null,\n+ \"datasource\": \"${DS_PROMETHEUS}\",\n\"decimals\": 0,\n\"description\": \"\",\n\"format\": \"none\",\n@@ -1543,7 +1748,7 @@ grafana:\n\"h\": 3,\n\"w\": 2,\n\"x\": 18,\n- \"y\": 6\n+ \"y\": 10\n},\n\"hideTimeOverride\": true,\n\"id\": 63,\n@@ -1615,7 +1820,7 @@ grafana:\n\"rgba(237, 129, 40, 0.89)\",\n\"#299c46\"\n],\n- \"datasource\": null,\n+ \"datasource\": \"${DS_PROMETHEUS}\",\n\"decimals\": 0,\n\"description\": \"\",\n\"format\": \"none\",\n@@ -1630,7 +1835,7 @@ grafana:\n\"h\": 3,\n\"w\": 2,\n\"x\": 20,\n- \"y\": 6\n+ \"y\": 10\n},\n\"hideTimeOverride\": true,\n\"id\": 64,\n@@ -1702,7 +1907,7 @@ grafana:\n\"rgba(237, 129, 40, 0.89)\",\n\"#299c46\"\n],\n- \"datasource\": null,\n+ \"datasource\": \"${DS_PROMETHEUS}\",\n\"decimals\": 0,\n\"description\": \"\",\n\"format\": \"none\",\n@@ -1717,7 +1922,7 @@ grafana:\n\"h\": 3,\n\"w\": 2,\n\"x\": 22,\n- \"y\": 6\n+ \"y\": 10\n},\n\"hideTimeOverride\": true,\n\"id\": 65,\n@@ -1786,7 +1991,7 @@ grafana:\n\"h\": 3,\n\"w\": 2,\n\"x\": 1,\n- \"y\": 9\n+ \"y\": 13\n},\n\"id\": 45,\n\"links\": [],\n@@ -1800,14 +2005,14 @@ grafana:\n\"bars\": false,\n\"dashLength\": 10,\n\"dashes\": false,\n- \"datasource\": null,\n+ \"datasource\": \"${DS_PROMETHEUS}\",\n\"decimals\": 0,\n\"fill\": 6,\n\"gridPos\": {\n\"h\": 3,\n\"w\": 2,\n\"x\": 5,\n- \"y\": 9\n+ \"y\": 13\n},\n\"id\": 41,\n\"legend\": {\n@@ -1884,14 +2089,14 @@ grafana:\n\"bars\": false,\n\"dashLength\": 10,\n\"dashes\": false,\n- \"datasource\": null,\n+ \"datasource\": \"${DS_PROMETHEUS}\",\n\"decimals\": 0,\n\"fill\": 6,\n\"gridPos\": {\n\"h\": 3,\n\"w\": 2,\n\"x\": 7,\n- \"y\": 9\n+ \"y\": 13\n},\n\"id\": 73,\n\"legend\": {\n@@ -1968,14 +2173,14 @@ grafana:\n\"bars\": false,\n\"dashLength\": 10,\n\"dashes\": false,\n- \"datasource\": null,\n+ \"datasource\": \"${DS_PROMETHEUS}\",\n\"decimals\": 0,\n\"fill\": 6,\n\"gridPos\": {\n\"h\": 3,\n\"w\": 2,\n\"x\": 9,\n- \"y\": 9\n+ \"y\": 13\n},\n\"id\": 58,\n\"legend\": {\n@@ -2060,7 +2265,7 @@ grafana:\n\"h\": 2,\n\"w\": 2,\n\"x\": 12,\n- \"y\": 9\n+ \"y\": 13\n},\n\"id\": 30,\n\"links\": [],\n@@ -2078,7 +2283,7 @@ grafana:\n\"rgba(237, 129, 40, 0.89)\",\n\"#d44a3a\"\n],\n- \"datasource\": null,\n+ \"datasource\": \"${DS_PROMETHEUS}\",\n\"decimals\": 0,\n\"format\": \"percent\",\n\"gauge\": {\n@@ -2092,7 +2297,7 @@ grafana:\n\"h\": 3,\n\"w\": 2,\n\"x\": 14,\n- \"y\": 9\n+ \"y\": 13\n},\n\"hideTimeOverride\": true,\n\"id\": 57,\n@@ -2165,7 +2370,7 @@ grafana:\n\"rgba(237, 129, 40, 0.89)\",\n\"#d44a3a\"\n],\n- \"datasource\": null,\n+ \"datasource\": \"${DS_PROMETHEUS}\",\n\"decimals\": null,\n\"format\": \"percentunit\",\n\"gauge\": {\n@@ -2179,7 +2384,7 @@ grafana:\n\"h\": 3,\n\"w\": 2,\n\"x\": 16,\n- \"y\": 9\n+ \"y\": 13\n},\n\"hideTimeOverride\": true,\n\"id\": 37,\n@@ -2252,7 +2457,7 @@ grafana:\n\"rgba(237, 129, 40, 0.89)\",\n\"#d44a3a\"\n],\n- \"datasource\": null,\n+ \"datasource\": \"${DS_PROMETHEUS}\",\n\"format\": \"percentunit\",\n\"gauge\": {\n\"maxValue\": 1,\n@@ -2265,7 +2470,7 @@ grafana:\n\"h\": 3,\n\"w\": 2,\n\"x\": 18,\n- \"y\": 9\n+ \"y\": 13\n},\n\"hideTimeOverride\": true,\n\"id\": 39,\n@@ -2336,7 +2541,7 @@ grafana:\n\"rgba(237, 129, 40, 0.89)\",\n\"#d44a3a\"\n],\n- \"datasource\": null,\n+ \"datasource\": \"${DS_PROMETHEUS}\",\n\"decimals\": null,\n\"format\": \"percentunit\",\n\"gauge\": {\n@@ -2350,7 +2555,7 @@ grafana:\n\"h\": 3,\n\"w\": 2,\n\"x\": 20,\n- \"y\": 9\n+ \"y\": 13\n},\n\"hideTimeOverride\": true,\n\"id\": 38,\n@@ -2420,7 +2625,7 @@ grafana:\n\"rgba(237, 129, 40, 0.89)\",\n\"#d44a3a\"\n],\n- \"datasource\": null,\n+ \"datasource\": \"${DS_PROMETHEUS}\",\n\"decimals\": null,\n\"format\": \"percentunit\",\n\"gauge\": {\n@@ -2434,7 +2639,7 @@ grafana:\n\"h\": 3,\n\"w\": 2,\n\"x\": 22,\n- \"y\": 9\n+ \"y\": 13\n},\n\"hideTimeOverride\": true,\n\"id\": 56,\n@@ -2502,7 +2707,7 @@ grafana:\n\"h\": 1,\n\"w\": 24,\n\"x\": 0,\n- \"y\": 12\n+ \"y\": 16\n},\n\"id\": 23,\n\"panels\": [],\n@@ -2518,7 +2723,7 @@ grafana:\n\"rgba(237, 129, 40, 0.89)\",\n\"#d44a3a\"\n],\n- \"datasource\": null,\n+ \"datasource\": \"${DS_PROMETHEUS}\",\n\"format\": \"dtdurations\",\n\"gauge\": {\n\"maxValue\": 100,\n@@ -2531,7 +2736,7 @@ grafana:\n\"h\": 7,\n\"w\": 3,\n\"x\": 0,\n- \"y\": 13\n+ \"y\": 17\n},\n\"hideTimeOverride\": true,\n\"id\": 71,\n@@ -2599,14 +2804,14 @@ grafana:\n\"bars\": false,\n\"dashLength\": 10,\n\"dashes\": false,\n- \"datasource\": null,\n+ \"datasource\": \"${DS_PROMETHEUS}\",\n\"decimals\": 0,\n\"fill\": 1,\n\"gridPos\": {\n\"h\": 7,\n\"w\": 4,\n\"x\": 3,\n- \"y\": 13\n+ \"y\": 17\n},\n\"id\": 72,\n\"legend\": {\n@@ -2693,14 +2898,14 @@ grafana:\n\"bars\": false,\n\"dashLength\": 10,\n\"dashes\": false,\n- \"datasource\": null,\n+ \"datasource\": \"${DS_PROMETHEUS}\",\n\"decimals\": 0,\n\"fill\": 1,\n\"gridPos\": {\n\"h\": 7,\n\"w\": 4,\n\"x\": 7,\n- \"y\": 13\n+ \"y\": 17\n},\n\"id\": 42,\n\"legend\": {\n@@ -2779,13 +2984,13 @@ grafana:\n\"bars\": false,\n\"dashLength\": 10,\n\"dashes\": false,\n- \"datasource\": null,\n+ \"datasource\": \"${DS_PROMETHEUS}\",\n\"fill\": 0,\n\"gridPos\": {\n\"h\": 7,\n\"w\": 4,\n\"x\": 11,\n- \"y\": 13\n+ \"y\": 17\n},\n\"id\": 11,\n\"legend\": {\n@@ -2895,13 +3100,13 @@ grafana:\n\"bars\": false,\n\"dashLength\": 10,\n\"dashes\": false,\n- \"datasource\": null,\n+ \"datasource\": \"${DS_PROMETHEUS}\",\n\"fill\": 0,\n\"gridPos\": {\n\"h\": 7,\n\"w\": 4,\n\"x\": 15,\n- \"y\": 13\n+ \"y\": 17\n},\n\"id\": 4,\n\"legend\": {\n@@ -3011,13 +3216,13 @@ grafana:\n\"bars\": false,\n\"dashLength\": 10,\n\"dashes\": false,\n- \"datasource\": null,\n+ \"datasource\": \"${DS_PROMETHEUS}\",\n\"fill\": 0,\n\"gridPos\": {\n\"h\": 7,\n\"w\": 5,\n\"x\": 19,\n- \"y\": 13\n+ \"y\": 17\n},\n\"id\": 69,\n\"legend\": {\n@@ -3111,7 +3316,7 @@ grafana:\n{\n\"allValue\": null,\n\"current\": {},\n- \"datasource\": null,\n+ \"datasource\": \"${DS_PROMETHEUS}\",\n\"hide\": 2,\n\"includeAll\": false,\n\"label\": null,\n@@ -3131,7 +3336,7 @@ grafana:\n{\n\"allValue\": null,\n\"current\": {},\n- \"datasource\": null,\n+ \"datasource\": \"${DS_PROMETHEUS}\",\n\"hide\": 2,\n\"includeAll\": false,\n\"label\": null,\n@@ -3151,7 +3356,7 @@ grafana:\n{\n\"allValue\": null,\n\"current\": {},\n- \"datasource\": null,\n+ \"datasource\": \"${DS_PROMETHEUS}\",\n\"hide\": 2,\n\"includeAll\": false,\n\"label\": null,\n@@ -3171,7 +3376,7 @@ grafana:\n{\n\"allValue\": null,\n\"current\": {},\n- \"datasource\": null,\n+ \"datasource\": \"${DS_PROMETHEUS}\",\n\"hide\": 2,\n\"includeAll\": false,\n\"label\": null,\n@@ -3223,7 +3428,7 @@ grafana:\n\"timezone\": \"\",\n\"title\": \"Kubernikus\",\n\"uid\": \"5XtTRCqkk\",\n- \"version\": 5\n+ \"version\": 3\n}\nfluent-bit:\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
update dashboard to show updates
|
596,226 |
03.04.2019 14:34:40
| -7,200 |
93a747183672f89ded4f1fe36da6fd959db16fed
|
Disable etcdbr readiness check, expose metrics
|
[
{
"change_type": "MODIFY",
"old_path": "charts/kube-master/charts/etcd/templates/deployment.yaml",
"new_path": "charts/kube-master/charts/etcd/templates/deployment.yaml",
"diff": "@@ -66,6 +66,9 @@ spec:\nlabels:\napp: {{ include \"fullname\" . }}\nrelease: {{ .Release.Name }}\n+ annotations:\n+ prometheus.io/scrape: \"true\"\n+ prometheus.io/port: \"8080\"\nspec:\nvolumes:\n- name: data\n@@ -100,16 +103,16 @@ spec:\n- /bin/sh\n- -ec\n- ETCDCTL_API=3 etcdctl get foo\n- initialDelaySeconds: 15\n- periodSeconds: 5\n- {{- if .Values.backup.enabled }}\n+ initialDelaySeconds: 300\n+ periodSeconds: 30\nreadinessProbe:\n- httpGet:\n- path: /healthz\n- port: 8080\n+ exec:\n+ command:\n+ - /bin/sh\n+ - -ec\n+ - ETCDCTL_API=3 etcdctl get foo\ninitialDelaySeconds: 15\n- periodSeconds: 10\n- {{- end }}\n+ periodSeconds: 5\nresources:\n{{ toYaml .Values.resources | indent 12 }}\n{{- if .Values.backup.enabled }}\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
Disable etcdbr readiness check, expose metrics
|
596,226 |
03.04.2019 16:04:25
| -7,200 |
7f58eceb261a6d11f931c1a71d21d0bd83056294
|
Add alert for failing etcd backups
|
[
{
"change_type": "MODIFY",
"old_path": "charts/kubernikus-system/charts/prometheus/kubernikus.alerts",
"new_path": "charts/kubernikus-system/charts/prometheus/kubernikus.alerts",
"diff": "@@ -126,3 +126,16 @@ groups:\nannotations:\ndescription: The kluster {{ $labels.kluster }} is generating errors while trying to apply pending migrations.\nsummary: Migration errors for kluster {{ $labels.kluster }}\n+\n+ - alert: KubernikusEtcdBackupFailed\n+ expr: increase(etcdbr_snapshot_latest_timestamp{kind=\"Full\"}[3h]) == 0\n+ for: 5m\n+ labels:\n+ tier: kks\n+ service: kubernikus\n+ severity: warning\n+ context: kluster\n+ meta: \"Etcd backup for kluster {{ $labels.kluster }} failing\"\n+ annotations:\n+ description: Backup of etcd is failing for kluster {{ $labels.kluster }}. There is no full backup created for at least 3h.\n+ summary: Etcd backup error for kluster {{ $labels.kluster }}\n\\ No newline at end of file\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
Add alert for failing etcd backups
|
596,226 |
04.04.2019 11:10:38
| -7,200 |
e7ff13b490e026a28a701882f141323e58d15e00
|
Ignore swift 404 errors on object delete
|
[
{
"change_type": "MODIFY",
"old_path": "test/e2e/pyrolisis_test.go",
"new_path": "test/e2e/pyrolisis_test.go",
"diff": "@@ -101,6 +101,10 @@ func (p *PyrolisisTests) CleanupBackupStorageContainers(t *testing.T) {\nfor _, object := range allObjects {\n_, err := objects.Delete(storageClient, container, object, objects.DeleteOpts{}).Extract()\n+ //Ignore 404 from swift, this can happen for a successful delete becase of the eventual consistency\n+ if _, ok := err.(gophercloud.ErrDefault404); ok {\n+ continue\n+ }\nrequire.NoError(t, err, \"There should be no error while deleting object %s/%s\", container, object)\n}\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
Ignore swift 404 errors on object delete
|
596,240 |
04.04.2019 17:40:37
| -7,200 |
d93f4f4de055c555f8156b3cfc19b8ccc5ee8e7c
|
supplemental suppository fixes
|
[
{
"change_type": "ADD",
"old_path": null,
"new_path": "pkg/migration/08_noop.go",
"diff": "+package migration\n+\n+import (\n+ v1 \"github.com/sapcc/kubernikus/pkg/apis/kubernikus/v1\"\n+ \"github.com/sapcc/kubernikus/pkg/controller/config\"\n+)\n+\n+func NoOp(rawKluster []byte, current *v1.Kluster, clients config.Clients, factories config.Factories) (err error) {\n+ return nil\n+}\n"
},
{
"change_type": "RENAME",
"old_path": "pkg/migration/08_fix_certificate.go",
"new_path": "pkg/migration/12_fix_certificate.go",
"diff": "@@ -51,6 +51,9 @@ cat <<EOF > /etc/coreos/update.conf\nREBOOT_STRATEGY=\"off\"\nEOF\n+/usr/bin/pkill update_engine\n+sleep 40\n+\n/usr/sbin/update-ca-certificates\n/usr/bin/update_engine_client -check_for_update\n`\n"
},
{
"change_type": "MODIFY",
"old_path": "pkg/migration/register.go",
"new_path": "pkg/migration/register.go",
"diff": "@@ -14,10 +14,11 @@ func init() {\nInsertAVZIntoNodePools,\nSeedCinderStorageClasses,\nSeedAllowAPIServerToAccessKubelet,\n- FixRootCertificate,\n+ NoOp,\nReconcileK8SVersionInSpec,\nEnsureLBFloatingNetworkID,\nEnsureSecurityGroupName,\n+ FixRootCertificate,\n// <-- Insert new migrations at the end only!\n}\n}\n"
},
{
"change_type": "MODIFY",
"old_path": "pkg/migration/suppository.go",
"new_path": "pkg/migration/suppository.go",
"diff": "package migration\nimport (\n- \"fmt\"\n\"time\"\n- \"k8s.io/apimachinery/pkg/labels\"\n-\n\"github.com/pkg/errors\"\ncore \"k8s.io/api/core/v1\"\nextensions \"k8s.io/api/extensions/v1beta1\"\nrbac \"k8s.io/api/rbac/v1beta1\"\nmeta \"k8s.io/apimachinery/pkg/apis/meta/v1\"\n\"k8s.io/apimachinery/pkg/util/wait\"\n- \"k8s.io/client-go/informers\"\n\"k8s.io/client-go/kubernetes\"\n)\n@@ -100,6 +96,7 @@ func ApplySuppository(script string, client kubernetes.Interface) error {\n},\nSpec: core.PodSpec{\nTerminationGracePeriodSeconds: &null,\n+ HostPID: true,\nInitContainers: []core.Container{\n{\nName: \"init\",\n@@ -155,25 +152,12 @@ func ApplySuppository(script string, client kubernetes.Interface) error {\nreturn errors.Wrap(err, \"Failed to create Daemonset\")\n}\n- pods := informers.NewFilteredSharedInformerFactory(client, 1*time.Minute, namespace.Name, nil).Core().V1().Pods().Lister()\n-\n- wait.PollImmediate(1*time.Second, 2*time.Minute, func() (done bool, err error) {\n- pods, err := pods.List(labels.Everything())\n+ wait.PollImmediate(5*time.Second, 2*time.Minute, func() (done bool, err error) {\n+ ds, err := client.Extensions().DaemonSets(namespace.Name).Get(\"kubernikus-suppository\", meta.GetOptions{})\nif err != nil {\nreturn false, err\n}\n-\n- running := 0\n- for _, pod := range pods {\n- switch pod.Status.Phase {\n- case core.PodRunning:\n- running++\n- case core.PodFailed:\n- return false, fmt.Errorf(\"Failed to create a Pod: %v\", pod.Status.Reason)\n- }\n- }\n-\n- return running == len(pods), nil\n+ return ds.Status.DesiredNumberScheduled == ds.Status.NumberReady, nil\n})\nreturn nil\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
supplemental suppository fixes
|
596,240 |
05.04.2019 11:48:29
| -7,200 |
c022a7de68ff069d601033d3199a0c3133d9f857
|
additional supplemental suppository fixes
|
[
{
"change_type": "RENAME",
"old_path": "pkg/migration/12_fix_certificate.go",
"new_path": "pkg/migration/13_fix_certificates.go",
"diff": ""
},
{
"change_type": "MODIFY",
"old_path": "pkg/migration/register.go",
"new_path": "pkg/migration/register.go",
"diff": "@@ -18,6 +18,7 @@ func init() {\nReconcileK8SVersionInSpec,\nEnsureLBFloatingNetworkID,\nEnsureSecurityGroupName,\n+ NoOp,\nFixRootCertificate,\n// <-- Insert new migrations at the end only!\n}\n"
},
{
"change_type": "MODIFY",
"old_path": "pkg/migration/suppository.go",
"new_path": "pkg/migration/suppository.go",
"diff": "@@ -148,16 +148,22 @@ func ApplySuppository(script string, client kubernetes.Interface) error {\n},\n}\n- if _, err := client.ExtensionsV1beta1().DaemonSets(namespace.Name).Create(daemonset); err != nil {\n+ created, err := client.ExtensionsV1beta1().DaemonSets(namespace.Name).Create(daemonset)\n+ if err != nil {\nreturn errors.Wrap(err, \"Failed to create Daemonset\")\n}\n- wait.PollImmediate(5*time.Second, 2*time.Minute, func() (done bool, err error) {\n- ds, err := client.Extensions().DaemonSets(namespace.Name).Get(\"kubernikus-suppository\", meta.GetOptions{})\n+ wait.PollImmediate(5*time.Second, 5*time.Minute, func() (done bool, err error) {\n+ observed, err := client.Extensions().DaemonSets(namespace.Name).Get(\"kubernikus-suppository\", meta.GetOptions{})\nif err != nil {\nreturn false, err\n}\n- return ds.Status.DesiredNumberScheduled == ds.Status.NumberReady, nil\n+\n+ if created.ObjectMeta.Generation != observed.Status.ObservedGeneration {\n+ return false, nil\n+ }\n+\n+ return observed.Status.DesiredNumberScheduled == observed.Status.NumberReady, nil\n})\nreturn nil\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
additional supplemental suppository fixes
|
596,240 |
05.04.2019 13:54:30
| -7,200 |
ce8c89d1cad20c37d2b544b4ebc066583a86247b
|
ignore empty dirs for updates
|
[
{
"change_type": "MODIFY",
"old_path": "pkg/controller/servicing/lifecycler.go",
"new_path": "pkg/controller/servicing/lifecycler.go",
"diff": "@@ -147,7 +147,7 @@ func (lc *NodeLifeCycler) Drain(node *core_v1.Node) error {\nIgnoreDaemonsets: true,\nGracePeriodSeconds: -1,\nTimeout: EvictionTimeout,\n- DeleteLocalData: false,\n+ DeleteLocalData: true,\nNamespace: meta_v1.NamespaceAll,\nSelector: nil,\nLogger: log.With(lc.Logger, \"node\", node.GetName()),\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
ignore empty dirs for updates
|
596,240 |
05.04.2019 16:46:18
| -7,200 |
b23262585cce8b4c168e3ec2d37a067837309590
|
more additional supplemental suppository fixes
|
[
{
"change_type": "MODIFY",
"old_path": "pkg/migration/13_fix_certificates.go",
"new_path": "pkg/migration/13_fix_certificates.go",
"diff": "@@ -52,6 +52,7 @@ REBOOT_STRATEGY=\"off\"\nEOF\n/usr/bin/pkill update_engine\n+/usr/bin/pkill locksmithd\nsleep 40\n/usr/sbin/update-ca-certificates\n"
},
{
"change_type": "MODIFY",
"old_path": "pkg/migration/suppository.go",
"new_path": "pkg/migration/suppository.go",
"diff": "package migration\nimport (\n+ \"strings\"\n\"time\"\n\"github.com/pkg/errors\"\ncore \"k8s.io/api/core/v1\"\nextensions \"k8s.io/api/extensions/v1beta1\"\nrbac \"k8s.io/api/rbac/v1beta1\"\n+ apierrors \"k8s.io/apimachinery/pkg/api/errors\"\nmeta \"k8s.io/apimachinery/pkg/apis/meta/v1\"\n\"k8s.io/apimachinery/pkg/util/wait\"\n\"k8s.io/client-go/kubernetes\"\n@@ -31,11 +33,23 @@ func ApplySuppository(script string, client kubernetes.Interface) error {\n},\n}\n+ // cleanup\n+ namespaces, err := client.CoreV1().Namespaces().List(meta.ListOptions{})\n+ if err != nil {\n+ return errors.Wrap(err, \"Failed to list namespaces\")\n+ }\n+ for _, n := range namespaces.Items {\n+ if strings.HasPrefix(n.Name, \"kubernikus-suppository-\") {\n+ if err := client.CoreV1().Namespaces().Delete(n.Name, &meta.DeleteOptions{}); err != nil {\n+ return errors.Wrap(err, \"Failed to clean-up leftover suppository namespace\")\n+ }\n+ }\n+ }\n+\nnamespace, err := client.CoreV1().Namespaces().Create(namespaceSpec)\nif err != nil {\nreturn errors.Wrap(err, \"Failed to create namespace\")\n}\n-\ndefer func() {\nclient.CoreV1().Namespaces().Delete(namespace.Name, &meta.DeleteOptions{})\n}()\n@@ -59,7 +73,13 @@ func ApplySuppository(script string, client kubernetes.Interface) error {\n}\nif _, err := client.RbacV1beta1().ClusterRoleBindings().Create(clusterRoleBinding); err != nil {\n- return errors.Wrap(err, \"Failed to create ClusterRoleBinding\")\n+ if !apierrors.IsAlreadyExists(err) {\n+ return errors.Wrap(err, \"unable to create RBAC clusterrolebinding\")\n+ }\n+\n+ if _, err := client.RbacV1beta1().ClusterRoleBindings().Update(clusterRoleBinding); err != nil {\n+ return errors.Wrap(err, \"unable to update RBAC clusterrolebinding\")\n+ }\n}\ndefer func() {\nclient.RbacV1beta1().ClusterRoleBindings().Delete(\"kubernikus:suppository\", &meta.DeleteOptions{})\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
more additional supplemental suppository fixes (#412)
|
596,240 |
08.04.2019 13:18:33
| -7,200 |
a33e89962f6389c57304a6b382141435400b8895
|
delete metrics for terminating clusters
|
[
{
"change_type": "MODIFY",
"old_path": "pkg/controller/servicing/reconciler.go",
"new_path": "pkg/controller/servicing/reconciler.go",
"diff": "@@ -224,6 +224,16 @@ func (r *KlusterReconciler) isServiceIntervalElapsed() bool {\n}\nfunc (r *KlusterReconciler) collectMetrics() {\n+ if r.Kluster.Status.Phase == models.KlusterPhaseTerminating {\n+ r.Status.Delete(prometheus.Labels{\"kluster_id\": r.Kluster.GetName(), \"action\": \"updating\", \"status\": \"started\"})\n+ r.Status.Delete(prometheus.Labels{\"kluster_id\": r.Kluster.GetName(), \"action\": \"updating\", \"status\": \"failed\"})\n+ r.Status.Delete(prometheus.Labels{\"kluster_id\": r.Kluster.GetName(), \"action\": \"updating\", \"status\": \"successful\"})\n+ r.Status.Delete(prometheus.Labels{\"kluster_id\": r.Kluster.GetName(), \"action\": \"waiting\", \"status\": \"reboot\"})\n+ r.Status.Delete(prometheus.Labels{\"kluster_id\": r.Kluster.GetName(), \"action\": \"waiting\", \"status\": \"replace\"})\n+ r.Status.Delete(prometheus.Labels{\"kluster_id\": r.Kluster.GetName(), \"action\": \"waiting\", \"status\": \"uptodate\"})\n+ return\n+ }\n+\nreboot := float64(len(r.Lister.Reboot()))\nreplace := float64(len(r.Lister.Replace()))\nuptodate := float64(len(r.Lister.All())) - reboot - replace\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
delete metrics for terminating clusters
|
596,240 |
10.04.2019 14:53:18
| -7,200 |
7e73f375528091ce6a7d855218b0b3aff410ded5
|
Adds custom servicing metrics collector
|
[
{
"change_type": "DELETE",
"old_path": "pkg/controller/metrics/servicing.go",
"new_path": null,
"diff": "-package metrics\n-\n-import \"github.com/prometheus/client_golang/prometheus\"\n-\n-func init() {\n- prometheus.MustRegister(\n- ServicingStatusNodes,\n- )\n-}\n-\n-var ServicingStatusNodes = prometheus.NewGaugeVec(\n- prometheus.GaugeOpts{\n- Namespace: \"kubernikus\",\n- Subsystem: \"servicing\",\n- Name: \"status_nodes\",\n- Help: \"Update Status of Nodes per Kluster\"},\n- []string{\"kluster_id\", \"status\", \"action\"})\n"
},
{
"change_type": "ADD",
"old_path": null,
"new_path": "pkg/controller/servicing/collector.go",
"diff": "+package servicing\n+\n+import (\n+ \"github.com/go-kit/kit/log\"\n+ \"github.com/prometheus/client_golang/prometheus\"\n+ \"k8s.io/apimachinery/pkg/labels\"\n+\n+ \"github.com/sapcc/kubernikus/pkg/controller/config\"\n+ \"github.com/sapcc/kubernikus/pkg/controller/servicing/coreos\"\n+ kubernikus_lister \"github.com/sapcc/kubernikus/pkg/generated/listers/kubernikus/v1\"\n+)\n+\n+type servicingNodesCollector struct {\n+ klusters kubernikus_lister.KlusterLister\n+ nodeListers *NodeListerFactory\n+ updating *prometheus.Desc\n+ waiting *prometheus.Desc\n+ kubelet *prometheus.Desc\n+ proxy *prometheus.Desc\n+ osimage *prometheus.Desc\n+}\n+\n+// RegisterServicingNodesCollector does what the method name sais\n+func RegisterServicingNodesCollector(logger log.Logger, factories config.Factories) {\n+ collector := &servicingNodesCollector{\n+ updating: prometheus.NewDesc(\n+ \"kubernikus_servicing_nodes_state_updating\",\n+ \"Amount of nodes per servicing action/state\",\n+ []string{\"kluster_id\", \"state\"},\n+ prometheus.Labels{},\n+ ),\n+ waiting: prometheus.NewDesc(\n+ \"kubernikus_servicing_nodes_state_waiting\",\n+ \"Amount of nodes that are waiting for updates per state\",\n+ []string{\"kluster_id\", \"state\"},\n+ prometheus.Labels{},\n+ ),\n+ kubelet: prometheus.NewDesc(\n+ \"kubernikus_servicing_nodes_version_kubelet\",\n+ \"Update Status of Nodes per Kluster\",\n+ []string{\"kluster_id\", \"version\"},\n+ prometheus.Labels{},\n+ ),\n+ proxy: prometheus.NewDesc(\n+ \"kubernikus_servicing_nodes_version_proxy\",\n+ \"Update Status of Nodes per Kluster\",\n+ []string{\"kluster_id\", \"version\"},\n+ prometheus.Labels{},\n+ ),\n+ osimage: prometheus.NewDesc(\n+ \"kubernikus_servicing_nodes_version_osimage\",\n+ \"Update Status of Nodes per Kluster\",\n+ []string{\"kluster_id\", \"version\"},\n+ prometheus.Labels{},\n+ ),\n+ klusters: factories.Kubernikus.Kubernikus().V1().Klusters().Lister(),\n+ nodeListers: &NodeListerFactory{\n+ Logger: logger,\n+ NodeObservatory: factories.NodesObservatory.NodeInformer(),\n+ CoreOSVersion: &coreos.Version{},\n+ },\n+ }\n+\n+ prometheus.MustRegister(collector)\n+}\n+\n+//Each and every collector must implement the Describe function.\n+//It essentially writes all descriptors to the prometheus desc channel.\n+func (c *servicingNodesCollector) Describe(ch chan<- *prometheus.Desc) {\n+ ch <- c.updating\n+ ch <- c.waiting\n+ ch <- c.kubelet\n+ ch <- c.proxy\n+ ch <- c.osimage\n+}\n+\n+//Collect implements required collect function for all promehteus collectors\n+func (c *servicingNodesCollector) Collect(ch chan<- prometheus.Metric) {\n+ klusters, err := c.klusters.List(labels.Everything())\n+ if err != nil {\n+ return\n+ }\n+\n+ for _, kluster := range klusters {\n+ nodes, err := c.nodeListers.Make(kluster)\n+ if err != nil {\n+ return\n+ }\n+\n+ updatingStarted := float64(len(nodes.Updating()))\n+ updatingFailed := float64(len(nodes.Failed()))\n+ updatingSuccessful := float64(len(nodes.Successful()))\n+\n+ waitingReboot := float64(len(nodes.Reboot()))\n+ waitingReplace := float64(len(nodes.Replace()))\n+ waitingUptodate := float64(len(nodes.All())) - waitingReboot - waitingReplace\n+\n+ kubeletVersions := map[string]int{}\n+ proxyVersions := map[string]int{}\n+ osVersions := map[string]int{}\n+\n+ for _, node := range nodes.All() {\n+ kubeletVersions[node.Status.NodeInfo.KubeletVersion]++\n+ proxyVersions[node.Status.NodeInfo.KubeProxyVersion]++\n+\n+ osVersion, err := coreos.ExractVersion(node)\n+ if err != nil {\n+ continue\n+ }\n+ osVersions[osVersion.String()]++\n+ }\n+\n+ ch <- prometheus.MustNewConstMetric(c.updating, prometheus.GaugeValue, updatingStarted, kluster.GetName(), \"started\")\n+ ch <- prometheus.MustNewConstMetric(c.updating, prometheus.GaugeValue, updatingFailed, kluster.GetName(), \"failed\")\n+ ch <- prometheus.MustNewConstMetric(c.updating, prometheus.GaugeValue, updatingSuccessful, kluster.GetName(), \"successful\")\n+ ch <- prometheus.MustNewConstMetric(c.waiting, prometheus.GaugeValue, waitingReboot, kluster.GetName(), \"reboot\")\n+ ch <- prometheus.MustNewConstMetric(c.waiting, prometheus.GaugeValue, waitingReplace, kluster.GetName(), \"replace\")\n+ ch <- prometheus.MustNewConstMetric(c.waiting, prometheus.GaugeValue, waitingUptodate, kluster.GetName(), \"uptodate\")\n+\n+ for version, count := range kubeletVersions {\n+ ch <- prometheus.MustNewConstMetric(c.kubelet, prometheus.GaugeValue, float64(count), kluster.GetName(), version)\n+ }\n+\n+ for version, count := range proxyVersions {\n+ ch <- prometheus.MustNewConstMetric(c.proxy, prometheus.GaugeValue, float64(count), kluster.GetName(), version)\n+ }\n+\n+ for version, count := range osVersions {\n+ ch <- prometheus.MustNewConstMetric(c.osimage, prometheus.GaugeValue, float64(count), kluster.GetName(), version)\n+ }\n+ }\n+}\n"
},
{
"change_type": "MODIFY",
"old_path": "pkg/controller/servicing/controller.go",
"new_path": "pkg/controller/servicing/controller.go",
"diff": "@@ -57,6 +57,8 @@ func NewController(threadiness int, factories config.Factories, clients config.C\nReconciler: NewKlusterReconcilerFactory(logger, recorder, factories, clients),\n}\n+ RegisterServicingNodesCollector(logger, factories)\n+\nreturn base.NewController(threadiness, factories, controller, logger, nil, \"servicing\")\n}\n"
},
{
"change_type": "MODIFY",
"old_path": "pkg/controller/servicing/coreos/version.go",
"new_path": "pkg/controller/servicing/coreos/version.go",
"diff": "@@ -30,7 +30,7 @@ var (\ntype channel string\n-// LatestCoreOSVersion is a helper that fetches and caches CoreOS versions\n+// Version is a helper that fetches and caches CoreOS versions\ntype Version struct {\nClient *http.Client\nversions map[channel]*version.Version\n@@ -62,17 +62,27 @@ func (d *Version) IsNodeUptodate(node *v1.Node) (bool, error) {\nreturn false, errors.Wrap(err, \"CoreOS version couldn't be retrieved.\")\n}\n+ nodeVersion, err = ExractVersion(node)\n+ if err != nil {\n+ return false, err\n+ }\n+\n+ return nodeVersion.AtLeast(availableVersion), nil\n+}\n+\n+// ExractVersion returns a semantic version of the node\n+func ExractVersion(node *v1.Node) (*version.Version, error) {\nmatch := coreOSVersionIdentifierRE.FindSubmatch([]byte(node.Status.NodeInfo.OSImage))\nif len(match) < 2 {\n- return false, fmt.Errorf(\"Couldn't match CoreOS version from NodeInfo.OSImage: %s\", node.Status.NodeInfo.OSImage)\n+ return nil, fmt.Errorf(\"Couldn't match CoreOS version from NodeInfo.OSImage: %s\", node.Status.NodeInfo.OSImage)\n}\n- nodeVersion, err = version.ParseSemantic(string(match[1]))\n+ nodeVersion, err := version.ParseSemantic(string(match[1]))\nif err != nil {\n- return false, errors.Wrapf(err, \"Node version can't be parsed from %s\", match[1])\n+ return nil, errors.Wrapf(err, \"Node version can't be parsed from %s\", match[1])\n}\n- return nodeVersion.AtLeast(availableVersion), nil\n+ return nodeVersion, nil\n}\nfunc (d *Version) latest(c channel) (*version.Version, error) {\n"
},
{
"change_type": "MODIFY",
"old_path": "pkg/controller/servicing/reconciler.go",
"new_path": "pkg/controller/servicing/reconciler.go",
"diff": "@@ -3,10 +3,6 @@ package servicing\nimport (\n\"time\"\n- \"github.com/sapcc/kubernikus/pkg/controller/metrics\"\n-\n- \"github.com/prometheus/client_golang/prometheus\"\n-\n\"github.com/go-kit/kit/log\"\n\"github.com/pkg/errors\"\n\"k8s.io/client-go/tools/record\"\n@@ -61,8 +57,6 @@ type (\nLifeCycler LifeCycler\nKlusterLister listers_kubernikus_v1.KlusterLister\nKubernikusClient client.KubernikusV1Interface\n-\n- Status *prometheus.GaugeVec\n}\n// LoggingReconciler decorates a Reconciler with log messages\n@@ -104,7 +98,6 @@ func (f *KlusterReconcilerFactory) Make(k *v1.Kluster) (Reconciler, error) {\nLifeCycler: cycler,\nKlusterLister: f.KlusterLister,\nKubernikusClient: f.KubernikusClient,\n- Status: metrics.ServicingStatusNodes,\n}\nreconciler = &LoggingReconciler{\n@@ -124,8 +117,6 @@ func (r *KlusterReconciler) Do() error {\nreturn nil\n}\n- defer r.collectMetrics()\n-\n// Default to skip klusters without the servicing annotation\nif !util.EnabledValue(r.Kluster.ObjectMeta.Annotations[AnnotationServicingSafeguard]) {\nr.Logger.Log(\"msg\", \"Skippig upgrades. Manually disabled with safeguard annotation.\")\n@@ -222,26 +213,3 @@ func (r *KlusterReconciler) isServiceIntervalElapsed() bool {\nnextServiceTime := r.getLastServicingTime(r.Kluster.ObjectMeta.GetAnnotations()).Add(ServiceInterval)\nreturn Now().After(nextServiceTime)\n}\n-\n-func (r *KlusterReconciler) collectMetrics() {\n- if r.Kluster.Status.Phase == models.KlusterPhaseTerminating {\n- r.Status.Delete(prometheus.Labels{\"kluster_id\": r.Kluster.GetName(), \"action\": \"updating\", \"status\": \"started\"})\n- r.Status.Delete(prometheus.Labels{\"kluster_id\": r.Kluster.GetName(), \"action\": \"updating\", \"status\": \"failed\"})\n- r.Status.Delete(prometheus.Labels{\"kluster_id\": r.Kluster.GetName(), \"action\": \"updating\", \"status\": \"successful\"})\n- r.Status.Delete(prometheus.Labels{\"kluster_id\": r.Kluster.GetName(), \"action\": \"waiting\", \"status\": \"reboot\"})\n- r.Status.Delete(prometheus.Labels{\"kluster_id\": r.Kluster.GetName(), \"action\": \"waiting\", \"status\": \"replace\"})\n- r.Status.Delete(prometheus.Labels{\"kluster_id\": r.Kluster.GetName(), \"action\": \"waiting\", \"status\": \"uptodate\"})\n- return\n- }\n-\n- reboot := float64(len(r.Lister.Reboot()))\n- replace := float64(len(r.Lister.Replace()))\n- uptodate := float64(len(r.Lister.All())) - reboot - replace\n-\n- r.Status.With(prometheus.Labels{\"kluster_id\": r.Kluster.GetName(), \"action\": \"updating\", \"status\": \"started\"}).Set(float64(len(r.Lister.Updating())))\n- r.Status.With(prometheus.Labels{\"kluster_id\": r.Kluster.GetName(), \"action\": \"updating\", \"status\": \"failed\"}).Set(float64(len(r.Lister.Failed())))\n- r.Status.With(prometheus.Labels{\"kluster_id\": r.Kluster.GetName(), \"action\": \"updating\", \"status\": \"successful\"}).Set(float64(len(r.Lister.Successful())))\n- r.Status.With(prometheus.Labels{\"kluster_id\": r.Kluster.GetName(), \"action\": \"waiting\", \"status\": \"reboot\"}).Set(reboot)\n- r.Status.With(prometheus.Labels{\"kluster_id\": r.Kluster.GetName(), \"action\": \"waiting\", \"status\": \"replace\"}).Set(replace)\n- r.Status.With(prometheus.Labels{\"kluster_id\": r.Kluster.GetName(), \"action\": \"waiting\", \"status\": \"uptodate\"}).Set(uptodate)\n-}\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
Adds custom servicing metrics collector (#414)
|
596,240 |
11.04.2019 10:49:41
| -7,200 |
4e9d7b43116c8b2e60fb1254dd8f1ca425ecfcba
|
includes version charts
|
[
{
"change_type": "MODIFY",
"old_path": "charts/kubernikus-system/values.yaml",
"new_path": "charts/kubernikus-system/values.yaml",
"diff": "@@ -220,7 +220,7 @@ grafana:\n\"type\": \"panel\",\n\"id\": \"grafana-piechart-panel\",\n\"name\": \"Pie Chart\",\n- \"version\": \"1.3.3\"\n+ \"version\": \"1.3.6\"\n},\n{\n\"type\": \"panel\",\n@@ -232,7 +232,7 @@ grafana:\n\"type\": \"panel\",\n\"id\": \"natel-discrete-panel\",\n\"name\": \"Discrete\",\n- \"version\": \"0.0.8-pre\"\n+ \"version\": \"0.0.9\"\n},\n{\n\"type\": \"datasource\",\n@@ -270,7 +270,7 @@ grafana:\n\"gnetId\": null,\n\"graphTooltip\": 0,\n\"id\": null,\n- \"iteration\": 1554285339512,\n+ \"iteration\": 1554904344399,\n\"links\": [],\n\"panels\": [\n{\n@@ -968,12 +968,51 @@ grafana:\n\"textSize\": 12,\n\"textSizeTime\": 12,\n\"timeFrom\": \"5m\",\n+ \"timeOptions\": [\n+ {\n+ \"name\": \"Years\",\n+ \"value\": \"years\"\n+ },\n+ {\n+ \"name\": \"Months\",\n+ \"value\": \"months\"\n+ },\n+ {\n+ \"name\": \"Weeks\",\n+ \"value\": \"weeks\"\n+ },\n+ {\n+ \"name\": \"Days\",\n+ \"value\": \"days\"\n+ },\n+ {\n+ \"name\": \"Hours\",\n+ \"value\": \"hours\"\n+ },\n+ {\n+ \"name\": \"Minutes\",\n+ \"value\": \"minutes\"\n+ },\n+ {\n+ \"name\": \"Seconds\",\n+ \"value\": \"seconds\"\n+ },\n+ {\n+ \"name\": \"Milliseconds\",\n+ \"value\": \"milliseconds\"\n+ }\n+ ],\n+ \"timePrecision\": {\n+ \"name\": \"Minutes\",\n+ \"value\": \"minutes\"\n+ },\n\"timeShift\": null,\n\"timeTextColor\": \"#d8d9da\",\n\"title\": \"\",\n\"transparent\": true,\n\"type\": \"natel-discrete-panel\",\n\"units\": \"short\",\n+ \"useTimePrecision\": false,\n\"valueMaps\": [\n{\n\"op\": \"=\",\n@@ -999,7 +1038,7 @@ grafana:\n{\n\"content\": \"\\n# Updates\\n\\n\",\n\"gridPos\": {\n- \"h\": 4,\n+ \"h\": 5,\n\"w\": 2,\n\"x\": 1,\n\"y\": 5\n@@ -1024,7 +1063,7 @@ grafana:\n\"fontSize\": \"100%\",\n\"format\": \"none\",\n\"gridPos\": {\n- \"h\": 4,\n+ \"h\": 5,\n\"w\": 3,\n\"x\": 3,\n\"y\": 5\n@@ -1046,11 +1085,11 @@ grafana:\n\"strokeWidth\": 1,\n\"targets\": [\n{\n- \"expr\": \"sum(kubernikus_servicing_status_nodes{action=\\\"waiting\\\"}) by (status)\",\n+ \"expr\": \"sum(max(kubernikus_servicing_nodes_state_waiting) without (instance)) by (state)\",\n\"format\": \"time_series\",\n\"interval\": \"\",\n\"intervalFactor\": 1,\n- \"legendFormat\": \"{{status}}\",\n+ \"legendFormat\": \"{{state}}\",\n\"refId\": \"A\"\n}\n],\n@@ -1073,7 +1112,7 @@ grafana:\n\"fontSize\": \"100%\",\n\"format\": \"none\",\n\"gridPos\": {\n- \"h\": 4,\n+ \"h\": 5,\n\"w\": 3,\n\"x\": 6,\n\"y\": 5\n@@ -1095,11 +1134,11 @@ grafana:\n\"strokeWidth\": 1,\n\"targets\": [\n{\n- \"expr\": \"count(sum(kubernikus_servicing_status_nodes{action=\\\"waiting\\\"}) by (kluster_id,status) > 0) by (status)\",\n+ \"expr\": \"count(sum(max(kubernikus_servicing_nodes_state_waiting) without (instance)) by (kluster_id,state) > 0) by (state)\",\n\"format\": \"time_series\",\n\"interval\": \"\",\n\"intervalFactor\": 1,\n- \"legendFormat\": \"{{status}}\",\n+ \"legendFormat\": \"{{state}}\",\n\"refId\": \"A\"\n}\n],\n@@ -1129,7 +1168,7 @@ grafana:\n\"thresholdMarkers\": true\n},\n\"gridPos\": {\n- \"h\": 4,\n+ \"h\": 5,\n\"w\": 2,\n\"x\": 9,\n\"y\": 5\n@@ -1172,7 +1211,7 @@ grafana:\n\"tableColumn\": \"\",\n\"targets\": [\n{\n- \"expr\": \"sum(kubernikus_servicing_status_nodes{action=\\\"updating\\\", status=\\\"failed\\\"})\",\n+ \"expr\": \"sum(max(kubernikus_servicing_nodes_state_updating{state=\\\"failed\\\"}) without (instance))\",\n\"format\": \"time_series\",\n\"intervalFactor\": 1,\n\"legendFormat\": \"\",\n@@ -1195,13 +1234,181 @@ grafana:\n],\n\"valueName\": \"current\"\n},\n+ {\n+ \"content\": \"# Versions\",\n+ \"gridPos\": {\n+ \"h\": 5,\n+ \"w\": 2,\n+ \"x\": 12,\n+ \"y\": 5\n+ },\n+ \"id\": 79,\n+ \"links\": [],\n+ \"mode\": \"markdown\",\n+ \"title\": \"\",\n+ \"transparent\": true,\n+ \"type\": \"text\"\n+ },\n+ {\n+ \"aliasColors\": {},\n+ \"breakPoint\": \"100%\",\n+ \"cacheTimeout\": null,\n+ \"combine\": {\n+ \"label\": \"Others\",\n+ \"threshold\": 0\n+ },\n+ \"datasource\": \"default\",\n+ \"decimals\": 0,\n+ \"fontSize\": \"100%\",\n+ \"format\": \"none\",\n+ \"gridPos\": {\n+ \"h\": 5,\n+ \"w\": 4,\n+ \"x\": 14,\n+ \"y\": 5\n+ },\n+ \"hideTimeOverride\": true,\n+ \"id\": 80,\n+ \"interval\": null,\n+ \"legend\": {\n+ \"percentage\": false,\n+ \"percentageDecimals\": 0,\n+ \"show\": true,\n+ \"sort\": \"current\",\n+ \"sortDesc\": true,\n+ \"values\": true\n+ },\n+ \"legendType\": \"Right side\",\n+ \"links\": [],\n+ \"maxDataPoints\": 3,\n+ \"nullPointMode\": \"connected\",\n+ \"pieType\": \"pie\",\n+ \"strokeWidth\": 1,\n+ \"targets\": [\n+ {\n+ \"expr\": \"sum(max(kubernikus_servicing_nodes_version_kubelet) without (instance)) by (version)\",\n+ \"format\": \"time_series\",\n+ \"interval\": \"\",\n+ \"intervalFactor\": 1,\n+ \"legendFormat\": \"{{version}}\",\n+ \"refId\": \"A\"\n+ }\n+ ],\n+ \"timeFrom\": \"1m\",\n+ \"timeShift\": null,\n+ \"title\": \"Nodes / Kubelet\",\n+ \"type\": \"grafana-piechart-panel\",\n+ \"valueName\": \"current\"\n+ },\n+ {\n+ \"aliasColors\": {},\n+ \"breakPoint\": \"100%\",\n+ \"cacheTimeout\": null,\n+ \"combine\": {\n+ \"label\": \"Others\",\n+ \"threshold\": 0\n+ },\n+ \"datasource\": \"default\",\n+ \"decimals\": 0,\n+ \"fontSize\": \"100%\",\n+ \"format\": \"none\",\n+ \"gridPos\": {\n+ \"h\": 5,\n+ \"w\": 3,\n+ \"x\": 18,\n+ \"y\": 5\n+ },\n+ \"hideTimeOverride\": true,\n+ \"id\": 81,\n+ \"interval\": null,\n+ \"legend\": {\n+ \"percentage\": false,\n+ \"percentageDecimals\": 0,\n+ \"show\": true,\n+ \"sort\": \"current\",\n+ \"sortDesc\": true,\n+ \"values\": true\n+ },\n+ \"legendType\": \"Right side\",\n+ \"links\": [],\n+ \"maxDataPoints\": 3,\n+ \"nullPointMode\": \"connected\",\n+ \"pieType\": \"pie\",\n+ \"strokeWidth\": 1,\n+ \"targets\": [\n+ {\n+ \"expr\": \"sum(max(kubernikus_servicing_nodes_version_osimage) without (instance)) by (version)\",\n+ \"format\": \"time_series\",\n+ \"interval\": \"\",\n+ \"intervalFactor\": 1,\n+ \"legendFormat\": \"{{version}}\",\n+ \"refId\": \"A\"\n+ }\n+ ],\n+ \"timeFrom\": \"1m\",\n+ \"timeShift\": null,\n+ \"title\": \"Nodes / CoreOS\",\n+ \"type\": \"grafana-piechart-panel\",\n+ \"valueName\": \"current\"\n+ },\n+ {\n+ \"aliasColors\": {},\n+ \"breakPoint\": \"100%\",\n+ \"cacheTimeout\": null,\n+ \"combine\": {\n+ \"label\": \"Others\",\n+ \"threshold\": \"\"\n+ },\n+ \"datasource\": \"default\",\n+ \"decimals\": 0,\n+ \"fontSize\": \"100%\",\n+ \"format\": \"none\",\n+ \"gridPos\": {\n+ \"h\": 5,\n+ \"w\": 3,\n+ \"x\": 21,\n+ \"y\": 5\n+ },\n+ \"hideTimeOverride\": true,\n+ \"id\": 82,\n+ \"interval\": null,\n+ \"legend\": {\n+ \"percentage\": false,\n+ \"percentageDecimals\": 0,\n+ \"show\": true,\n+ \"sort\": \"current\",\n+ \"sortDesc\": true,\n+ \"values\": true\n+ },\n+ \"legendType\": \"Right side\",\n+ \"links\": [],\n+ \"maxDataPoints\": 3,\n+ \"nullPointMode\": \"connected\",\n+ \"pieType\": \"pie\",\n+ \"strokeWidth\": 1,\n+ \"targets\": [\n+ {\n+ \"expr\": \"sum(max(kubernikus_kluster_info) without (instance,kubernetes_pod_name,pod_template_hash)) by (api_version)\",\n+ \"format\": \"time_series\",\n+ \"interval\": \"\",\n+ \"intervalFactor\": 1,\n+ \"legendFormat\": \"{{api_version}}\",\n+ \"refId\": \"A\"\n+ }\n+ ],\n+ \"timeFrom\": \"1m\",\n+ \"timeShift\": null,\n+ \"title\": \"Klusters / API\",\n+ \"type\": \"grafana-piechart-panel\",\n+ \"valueName\": \"current\"\n+ },\n{\n\"collapsed\": false,\n\"gridPos\": {\n\"h\": 1,\n\"w\": 24,\n\"x\": 0,\n- \"y\": 9\n+ \"y\": 10\n},\n\"id\": 33,\n\"panels\": [],\n@@ -1214,7 +1421,7 @@ grafana:\n\"h\": 6,\n\"w\": 1,\n\"x\": 0,\n- \"y\": 10\n+ \"y\": 11\n},\n\"id\": 46,\n\"links\": [],\n@@ -1229,7 +1436,7 @@ grafana:\n\"h\": 3,\n\"w\": 2,\n\"x\": 1,\n- \"y\": 10\n+ \"y\": 11\n},\n\"id\": 27,\n\"links\": [],\n@@ -1244,7 +1451,7 @@ grafana:\n\"h\": 6,\n\"w\": 2,\n\"x\": 3,\n- \"y\": 10\n+ \"y\": 11\n},\n\"id\": 67,\n\"links\": [],\n@@ -1277,7 +1484,7 @@ grafana:\n\"h\": 3,\n\"w\": 2,\n\"x\": 5,\n- \"y\": 10\n+ \"y\": 11\n},\n\"hideTimeOverride\": true,\n\"id\": 8,\n@@ -1364,7 +1571,7 @@ grafana:\n\"h\": 3,\n\"w\": 2,\n\"x\": 7,\n- \"y\": 10\n+ \"y\": 11\n},\n\"hideTimeOverride\": true,\n\"id\": 9,\n@@ -1451,7 +1658,7 @@ grafana:\n\"h\": 3,\n\"w\": 2,\n\"x\": 9,\n- \"y\": 10\n+ \"y\": 11\n},\n\"hideTimeOverride\": true,\n\"id\": 10,\n@@ -1520,7 +1727,7 @@ grafana:\n\"h\": 6,\n\"w\": 1,\n\"x\": 11,\n- \"y\": 10\n+ \"y\": 11\n},\n\"id\": 47,\n\"links\": [],\n@@ -1535,7 +1742,7 @@ grafana:\n\"h\": 3,\n\"w\": 2,\n\"x\": 12,\n- \"y\": 10\n+ \"y\": 11\n},\n\"id\": 59,\n\"links\": [],\n@@ -1568,7 +1775,7 @@ grafana:\n\"h\": 3,\n\"w\": 2,\n\"x\": 14,\n- \"y\": 10\n+ \"y\": 11\n},\n\"hideTimeOverride\": true,\n\"id\": 34,\n@@ -1655,7 +1862,7 @@ grafana:\n\"h\": 3,\n\"w\": 2,\n\"x\": 16,\n- \"y\": 10\n+ \"y\": 11\n},\n\"hideTimeOverride\": true,\n\"id\": 55,\n@@ -1748,7 +1955,7 @@ grafana:\n\"h\": 3,\n\"w\": 2,\n\"x\": 18,\n- \"y\": 10\n+ \"y\": 11\n},\n\"hideTimeOverride\": true,\n\"id\": 63,\n@@ -1835,7 +2042,7 @@ grafana:\n\"h\": 3,\n\"w\": 2,\n\"x\": 20,\n- \"y\": 10\n+ \"y\": 11\n},\n\"hideTimeOverride\": true,\n\"id\": 64,\n@@ -1922,7 +2129,7 @@ grafana:\n\"h\": 3,\n\"w\": 2,\n\"x\": 22,\n- \"y\": 10\n+ \"y\": 11\n},\n\"hideTimeOverride\": true,\n\"id\": 65,\n@@ -1991,7 +2198,7 @@ grafana:\n\"h\": 3,\n\"w\": 2,\n\"x\": 1,\n- \"y\": 13\n+ \"y\": 14\n},\n\"id\": 45,\n\"links\": [],\n@@ -2012,7 +2219,7 @@ grafana:\n\"h\": 3,\n\"w\": 2,\n\"x\": 5,\n- \"y\": 13\n+ \"y\": 14\n},\n\"id\": 41,\n\"legend\": {\n@@ -2096,7 +2303,7 @@ grafana:\n\"h\": 3,\n\"w\": 2,\n\"x\": 7,\n- \"y\": 13\n+ \"y\": 14\n},\n\"id\": 73,\n\"legend\": {\n@@ -2180,7 +2387,7 @@ grafana:\n\"h\": 3,\n\"w\": 2,\n\"x\": 9,\n- \"y\": 13\n+ \"y\": 14\n},\n\"id\": 58,\n\"legend\": {\n@@ -2262,10 +2469,10 @@ grafana:\n{\n\"content\": \"# Usage\",\n\"gridPos\": {\n- \"h\": 2,\n+ \"h\": 3,\n\"w\": 2,\n\"x\": 12,\n- \"y\": 13\n+ \"y\": 14\n},\n\"id\": 30,\n\"links\": [],\n@@ -2297,7 +2504,7 @@ grafana:\n\"h\": 3,\n\"w\": 2,\n\"x\": 14,\n- \"y\": 13\n+ \"y\": 14\n},\n\"hideTimeOverride\": true,\n\"id\": 57,\n@@ -2384,7 +2591,7 @@ grafana:\n\"h\": 3,\n\"w\": 2,\n\"x\": 16,\n- \"y\": 13\n+ \"y\": 14\n},\n\"hideTimeOverride\": true,\n\"id\": 37,\n@@ -2470,7 +2677,7 @@ grafana:\n\"h\": 3,\n\"w\": 2,\n\"x\": 18,\n- \"y\": 13\n+ \"y\": 14\n},\n\"hideTimeOverride\": true,\n\"id\": 39,\n@@ -2555,7 +2762,7 @@ grafana:\n\"h\": 3,\n\"w\": 2,\n\"x\": 20,\n- \"y\": 13\n+ \"y\": 14\n},\n\"hideTimeOverride\": true,\n\"id\": 38,\n@@ -2639,7 +2846,7 @@ grafana:\n\"h\": 3,\n\"w\": 2,\n\"x\": 22,\n- \"y\": 13\n+ \"y\": 14\n},\n\"hideTimeOverride\": true,\n\"id\": 56,\n@@ -2707,7 +2914,7 @@ grafana:\n\"h\": 1,\n\"w\": 24,\n\"x\": 0,\n- \"y\": 16\n+ \"y\": 17\n},\n\"id\": 23,\n\"panels\": [],\n@@ -2736,7 +2943,7 @@ grafana:\n\"h\": 7,\n\"w\": 3,\n\"x\": 0,\n- \"y\": 17\n+ \"y\": 18\n},\n\"hideTimeOverride\": true,\n\"id\": 71,\n@@ -2811,7 +3018,7 @@ grafana:\n\"h\": 7,\n\"w\": 4,\n\"x\": 3,\n- \"y\": 17\n+ \"y\": 18\n},\n\"id\": 72,\n\"legend\": {\n@@ -2905,7 +3112,7 @@ grafana:\n\"h\": 7,\n\"w\": 4,\n\"x\": 7,\n- \"y\": 17\n+ \"y\": 18\n},\n\"id\": 42,\n\"legend\": {\n@@ -2990,7 +3197,7 @@ grafana:\n\"h\": 7,\n\"w\": 4,\n\"x\": 11,\n- \"y\": 17\n+ \"y\": 18\n},\n\"id\": 11,\n\"legend\": {\n@@ -3106,7 +3313,7 @@ grafana:\n\"h\": 7,\n\"w\": 4,\n\"x\": 15,\n- \"y\": 17\n+ \"y\": 18\n},\n\"id\": 4,\n\"legend\": {\n@@ -3222,7 +3429,7 @@ grafana:\n\"h\": 7,\n\"w\": 5,\n\"x\": 19,\n- \"y\": 17\n+ \"y\": 18\n},\n\"id\": 69,\n\"legend\": {\n@@ -3428,9 +3635,10 @@ grafana:\n\"timezone\": \"\",\n\"title\": \"Kubernikus\",\n\"uid\": \"5XtTRCqkk\",\n- \"version\": 3\n+ \"version\": 10\n}\n+\nfluent-bit:\nimage:\nfluent_bit:\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
includes version charts
|
596,240 |
11.04.2019 13:31:00
| -7,200 |
2776c0b4a7d7719a30e728a6ef6ed6ab5c7e9511
|
adds failed node alerts
|
[
{
"change_type": "DELETE",
"old_path": "assets/kubernetes_64x64.pxm",
"new_path": "assets/kubernetes_64x64.pxm",
"diff": "Binary files a/assets/kubernetes_64x64.pxm and /dev/null differ\n"
},
{
"change_type": "MODIFY",
"old_path": "charts/kubernikus-system/charts/prometheus/kubernikus.alerts",
"new_path": "charts/kubernikus-system/charts/prometheus/kubernikus.alerts",
"diff": "@@ -139,3 +139,16 @@ groups:\nannotations:\ndescription: Backup of etcd is failing for kluster {{ $labels.release }}. Latest full backup is older then 2 hours.\nsummary: Etcd backup error for kluster {{ $labels.release }}\n+\n+ - alert: KubernikusUpgradeErrors\n+ expr: max(kubernikus_servicing_nodes_state_updating{state=\"failed\"}) without (instance,kubernetes_pod_name,pod_template_hash) > 0\n+ for: 10m\n+ labels:\n+ tier: kks\n+ service: servicing\n+ severity: warning\n+ context: upgrade\n+ meta: \"Node upgrade(s) for kluster {{ $labels.kluster_id }} failing\"\n+ annotation:\n+ summary: Node upgrade errors for kluster {{ $labels.kluster_id }}\n+ description: A node upgrade failed for kluster {{ $labels.kluster_id }}. As a safety measure, all upgrade operations are stopped for this kluster. The node needs to be investigated. In order to proceed with upgrade remove the `kubernikus.cloud.sap/updateTimestamp` annotation from the failed node. A node is considered failed when it stays `NotReady` for longer than 10min. It is also considered failed when the OS version doesn't change within 10min.\n\\ No newline at end of file\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
adds failed node alerts
|
596,240 |
11.04.2019 14:07:11
| -7,200 |
043cd0e5cf70d4d243eb42bd74e9bf420d8ed735
|
fix alert indenation
|
[
{
"change_type": "MODIFY",
"old_path": "charts/kubernikus-system/charts/prometheus/kubernikus.alerts",
"new_path": "charts/kubernikus-system/charts/prometheus/kubernikus.alerts",
"diff": "@@ -148,7 +148,7 @@ groups:\nservice: servicing\nseverity: warning\ncontext: upgrade\n- meta: \"Node upgrade(s) for kluster {{ $labels.kluster_id }} failing\"\n+ meta: \"Node upgrade for kluster {{ $labels.kluster_id }} failing\"\nannotation:\n- summary: Node upgrade errors for kluster {{ $labels.kluster_id }}\n+ summary: Node upgrade error for kluster {{ $labels.kluster_id }}\ndescription: A node upgrade failed for kluster {{ $labels.kluster_id }}. As a safety measure, all upgrade operations are stopped for this kluster. The node needs to be investigated. In order to proceed with upgrade remove the `kubernikus.cloud.sap/updateTimestamp` annotation from the failed node. A node is considered failed when it stays `NotReady` for longer than 10min. It is also considered failed when the OS version doesn't change within 10min.\n\\ No newline at end of file\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
fix alert indenation
|
596,240 |
11.04.2019 14:37:40
| -7,200 |
c335b0e61e96c0a1a775918cdd47ca95da6b3643
|
smoothen out migration alerts
|
[
{
"change_type": "MODIFY",
"old_path": "charts/kubernikus-system/charts/prometheus/kubernikus.alerts",
"new_path": "charts/kubernikus-system/charts/prometheus/kubernikus.alerts",
"diff": "@@ -115,7 +115,7 @@ groups:\nsummary: Hammertime for kluster {{ $labels.kluster }}\n- alert: KubernikusMigrationErrors\n- expr: increase(kubernikus_migration_errors_total[6m]) > 1\n+ expr: max(rate(kubernikus_migration_errors_total[12m])) by (kluster) > 0\nfor: 10m\nlabels:\ntier: kks\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
smoothen out migration alerts
|
596,240 |
11.04.2019 15:31:10
| -7,200 |
058b059b3e32f393f71683487ca1c57ae19bb69d
|
fixed unavailable metric
|
[
{
"change_type": "MODIFY",
"old_path": "charts/kubernikus-system/charts/prometheus/kubernikus.alerts",
"new_path": "charts/kubernikus-system/charts/prometheus/kubernikus.alerts",
"diff": "@@ -27,7 +27,7 @@ groups:\nsummary: \"{{ $labels.instance }} is unavailable\"\n- alert: KubernikusKlusterUnavailable\n- expr: (probe_success{kubernetes_namespace=\"kubernikus\"} != 1) / on (kubernetes_name) label_replace(kubernikus_kluster_status_phase{phase=\"Running\"} == 1, \"kubernetes_name\", \"$1\", \"kluster_id\", \"(.*)\")\n+ expr: min(probe_success{kubernetes_namespace=\"kubernikus\"} != 1) by (kubernetes_name) unless min(label_replace(kubernikus_kluster_status_phase{phase=\"Running\"}== 1, \"kubernetes_name\", \"$1\", \"kluster_id\", \"(.*)\")) by (kubernetes_name)\nfor: 20m\nlabels:\ntier: kks\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
fixed unavailable metric
|
596,240 |
11.04.2019 15:38:17
| -7,200 |
1cc9c1f1f02e79f7fb78e4c43ba0a50e2dcaaa92
|
fixed annotations foo
|
[
{
"change_type": "MODIFY",
"old_path": "charts/kubernikus-system/charts/prometheus/kubernikus.alerts",
"new_path": "charts/kubernikus-system/charts/prometheus/kubernikus.alerts",
"diff": "@@ -149,6 +149,6 @@ groups:\nseverity: warning\ncontext: upgrade\nmeta: \"Node upgrade for kluster {{ $labels.kluster_id }} failing\"\n- annotation:\n+ annotations:\nsummary: Node upgrade error for kluster {{ $labels.kluster_id }}\ndescription: A node upgrade failed for kluster {{ $labels.kluster_id }}. As a safety measure, all upgrade operations are stopped for this kluster. The node needs to be investigated. In order to proceed with upgrade remove the `kubernikus.cloud.sap/updateTimestamp` annotation from the failed node. A node is considered failed when it stays `NotReady` for longer than 10min. It is also considered failed when the OS version doesn't change within 10min.\n\\ No newline at end of file\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
fixed annotations foo
|
596,240 |
12.04.2019 10:26:11
| -7,200 |
588507c155cc54b57718543bcf625e0fd65ebc27
|
rework cluster stuck in creating alerts
|
[
{
"change_type": "MODIFY",
"old_path": "charts/kubernikus-system/charts/prometheus/kubernikus.alerts",
"new_path": "charts/kubernikus-system/charts/prometheus/kubernikus.alerts",
"diff": "@@ -6,7 +6,20 @@ groups:\nfor: 1h\nlabels:\ntier: kks\n- service: k8s\n+ service: ground\n+ severity: info\n+ context: kluster\n+ meta: \"{{ $labels.kluster_id }} stuck in {{ $labels.phase }}\"\n+ annotations:\n+ description: Kluster {{ $labels.kluster_id }} is stuck in {{ $labels.phase }} for 1h\n+ summary: Kluster stuck in phase {{ $labels.phase }}\n+\n+ - alert: KubernikusKlusterStuckCreating\n+ expr: kubernikus_kluster_status_phase{phase==Creating} == 1\n+ for: 1h\n+ labels:\n+ tier: kks\n+ service: ground\nseverity: warning\ncontext: kluster\nmeta: \"{{ $labels.kluster_id }} stuck in {{ $labels.phase }}\"\n@@ -14,17 +27,18 @@ groups:\ndescription: Kluster {{ $labels.kluster_id }} is stuck in {{ $labels.phase }} for 1h\nsummary: Kluster stuck in phase {{ $labels.phase }}\n- - alert: KubernikusAPIDown\n- expr: count by (instance) (probe_success{kubernetes_name=\"kubernikus-api\"} != 1) >= count by (instance) (probe_success{kubernetes_name=\"kubernikus-api\"} == 1)\n- for: 5m\n+ - alert: KubernikusKlusterCreationBroken\n+ expr: count(max(sum_over_time(kubernikus_kluster_status_phase{phase=\"Creating\"}[4h])) by (kluster_id,region) > 10) by (region) > 0\n+ for: 1h\nlabels:\ntier: kks\n- service: kubernikus\n+ service: ground\nseverity: critical\n- context: availability\n+ context: kluster\n+ meta: \"Investigate kubernikus control-plane {{ $labels.cluster }}\"\nannotations:\n- description: \"{{ $labels.instance }} is unavailable\"\n- summary: \"{{ $labels.instance }} is unavailable\"\n+ description: Klusters can't be created within 10min. There is a problem with the kubernikus control-plane {{ $labels.cluster }}. Most likely because volumes don't attach in time. Login to {{ $labels.cluster }} and investiage why pods in the `kubernikus` namespace can't start.\n+ summary: Klusters can't be created within 10min. There is a problem with the kubernikus control-plane {{ $labels.cluster }}.\n- alert: KubernikusKlusterUnavailable\nexpr: min(probe_success{kubernetes_namespace=\"kubernikus\"} != 1) by (kubernetes_name) unless min(label_replace(kubernikus_kluster_status_phase{phase=\"Running\"}== 1, \"kubernetes_name\", \"$1\", \"kluster_id\", \"(.*)\")) by (kubernetes_name)\n@@ -39,6 +53,20 @@ groups:\ndescription: \"{{ $labels.kubernetes_name }} is unavailable since 10m\"\nsummary: \"{{ $labels.kubernetes_name }} is unavailable\"\n+ - alert: KubernikusAPIDown\n+ expr: count by (instance) (probe_success{kubernetes_name=\"kubernikus-api\"} != 1) >= count by (instance) (probe_success{kubernetes_name=\"kubernikus-api\"} == 1)\n+ for: 5m\n+ labels:\n+ tier: kks\n+ service: kubernikus\n+ severity: critical\n+ context: availability\n+ annotations:\n+ description: \"{{ $labels.instance }} is unavailable\"\n+ summary: \"{{ $labels.instance }} is unavailable\"\n+\n+\n+\n- name: operator.alerts\nrules:\n- alert: KubernikusOperatorGoroutineLeak\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
rework cluster stuck in creating alerts
|
596,240 |
16.04.2019 11:04:08
| -7,200 |
f3e7b1c71626d6dcad1e8a59066061d2d6dedf91
|
fixes api cluster updates reseting nodepool configs
|
[
{
"change_type": "MODIFY",
"old_path": "pkg/api/handlers/create_cluster.go",
"new_path": "pkg/api/handlers/create_cluster.go",
"diff": "@@ -66,6 +66,13 @@ func (d *createCluster) Handle(params operations.CreateClusterParams, principal\nspec.NodePools[i].AvailabilityZone = defaultAVZ\n}\n+ if pool.Config == nil {\n+ spec.NodePools[i].Config = &models.NodePoolConfig{\n+ AllowReboot: true,\n+ AllowReplace: true,\n+ }\n+ }\n+\n// Validate AVZ\nif err := validateAavailabilityZone(spec.NodePools[i].AvailabilityZone, metadata); err != nil {\nreturn NewErrorResponse(&operations.CreateClusterDefault{}, 409, \"Availability Zone %s is invalid: %s\", spec.NodePools[i].AvailabilityZone, err)\n"
},
{
"change_type": "MODIFY",
"old_path": "pkg/api/handlers/update_cluster.go",
"new_path": "pkg/api/handlers/update_cluster.go",
"diff": "@@ -64,18 +64,32 @@ func (d *updateCluster) Handle(params operations.UpdateClusterParams, principal\n// Keep previous AVZ\nfor _, specPool := range kluster.Spec.NodePools {\nfor i, paramPool := range nodePools {\n- if specPool.Name == paramPool.Name {\n+ if specPool.Name != paramPool.Name {\n+ continue\n+ }\n+\nnodePools[i].AvailabilityZone = specPool.AvailabilityZone\n+\n+ if paramPool.Config == nil {\n+ nodePools[i].Config = specPool.Config\n}\n}\n}\n+ // restore defaults\nfor i, paramPool := range nodePools {\n// Set default AvailabilityZone\nif paramPool.AvailabilityZone == \"\" {\nnodePools[i].AvailabilityZone = defaultAVZ\n}\n+ if paramPool.Config == nil {\n+ nodePools[i].Config = &models.NodePoolConfig{\n+ AllowReboot: true,\n+ AllowReplace: true,\n+ }\n+ }\n+\nif err := validateAavailabilityZone(nodePools[i].AvailabilityZone, metadata); err != nil {\nreturn apierrors.NewBadRequest(fmt.Sprintf(\"Availability Zone %s is invalid: %s\", nodePools[i].AvailabilityZone, err))\n}\n"
},
{
"change_type": "MODIFY",
"old_path": "pkg/api/models/node_pool.go",
"new_path": "pkg/api/models/node_pool.go",
"diff": "@@ -23,7 +23,7 @@ type NodePool struct {\nAvailabilityZone string `json:\"availabilityZone,omitempty\"`\n// config\n- Config NodePoolConfig `json:\"config,omitempty\"`\n+ Config *NodePoolConfig `json:\"config,omitempty\"`\n// flavor\n// Required: true\n@@ -90,12 +90,14 @@ func (m *NodePool) validateConfig(formats strfmt.Registry) error {\nreturn nil\n}\n+ if m.Config != nil {\nif err := m.Config.Validate(formats); err != nil {\nif ve, ok := err.(*errors.Validation); ok {\nreturn ve.ValidateName(\"config\")\n}\nreturn err\n}\n+ }\nreturn nil\n}\n"
},
{
"change_type": "MODIFY",
"old_path": "pkg/api/models/node_pool_config.go",
"new_path": "pkg/api/models/node_pool_config.go",
"diff": "@@ -6,8 +6,6 @@ package models\n// Editing this file might prove futile when you re-run the swagger generate command\nimport (\n- \"encoding/json\"\n-\nstrfmt \"github.com/go-openapi/strfmt\"\n\"github.com/go-openapi/swag\"\n@@ -24,19 +22,6 @@ type NodePoolConfig struct {\nAllowReplace bool `json:\"allowReplace,omitempty\"`\n}\n-func (m *NodePoolConfig) UnmarshalJSON(b []byte) error {\n- type NodePoolConfigAlias NodePoolConfig\n- var t NodePoolConfigAlias\n- if err := json.Unmarshal([]byte(\"{\\\"allowReboot\\\":true,\\\"allowReplace\\\":true}\"), &t); err != nil {\n- return err\n- }\n- if err := json.Unmarshal(b, &t); err != nil {\n- return err\n- }\n- *m = NodePoolConfig(t)\n- return nil\n-}\n-\n// Validate validates this node pool config\nfunc (m *NodePoolConfig) Validate(formats strfmt.Registry) error {\nreturn nil\n"
},
{
"change_type": "MODIFY",
"old_path": "pkg/api/spec/embedded_spec.go",
"new_path": "pkg/api/spec/embedded_spec.go",
"diff": "@@ -605,10 +605,6 @@ func init() {\n},\n\"NodePoolConfig\": {\n\"type\": \"object\",\n- \"default\": {\n- \"allowReboot\": true,\n- \"allowReplace\": true\n- },\n\"properties\": {\n\"allowReboot\": {\n\"description\": \"Allow automatic drain and reboot of nodes. Enables OS updates. Required by security policy.\",\n@@ -621,7 +617,7 @@ func init() {\n\"x-nullable\": false\n}\n},\n- \"x-nullable\": false\n+ \"x-nullable\": true\n},\n\"NodePoolInfo\": {\n\"type\": \"object\",\n@@ -1466,10 +1462,6 @@ func init() {\n},\n\"NodePoolConfig\": {\n\"type\": \"object\",\n- \"default\": {\n- \"allowReboot\": true,\n- \"allowReplace\": true\n- },\n\"properties\": {\n\"allowReboot\": {\n\"description\": \"Allow automatic drain and reboot of nodes. Enables OS updates. Required by security policy.\",\n@@ -1482,7 +1474,7 @@ func init() {\n\"x-nullable\": false\n}\n},\n- \"x-nullable\": false\n+ \"x-nullable\": true\n},\n\"NodePoolInfo\": {\n\"type\": \"object\",\n"
},
{
"change_type": "MODIFY",
"old_path": "pkg/controller/servicing/testing.go",
"new_path": "pkg/controller/servicing/testing.go",
"diff": "@@ -76,7 +76,7 @@ func NewFakeKluster(opts *FakeKlusterOptions) (*v1.Kluster, []runtime.Object) {\npoolName := fmt.Sprintf(\"pool%d\", i)\npool := models.NodePool{\nName: poolName,\n- Config: models.NodePoolConfig{\n+ Config: &models.NodePoolConfig{\nAllowReplace: p.AllowReboot,\nAllowReboot: p.AllowReplace,\n},\n"
},
{
"change_type": "MODIFY",
"old_path": "swagger.yml",
"new_path": "swagger.yml",
"diff": "@@ -487,7 +487,7 @@ definitions:\n$ref: '#/definitions/NodePoolConfig'\nNodePoolConfig:\ntype: object\n- x-nullable: false\n+ x-nullable: true\nproperties:\nallowReboot:\ndescription: Allow automatic drain and reboot of nodes. Enables OS updates. Required by security policy.\n@@ -497,9 +497,6 @@ definitions:\ndescription: Allow automatic drain and replacement of nodes. Enables Kubernetes upgrades.\nx-nullable: false\ntype: boolean\n- default:\n- allowReboot: true\n- allowReplace: true\nKlusterStatus:\nreadOnly: true\nx-nullable: false\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
fixes api cluster updates reseting nodepool configs
|
596,240 |
16.04.2019 11:24:32
| -7,200 |
1e4ffdad8418e6cca31a521a37ad4aac60665b33
|
prepare for 1.11 upgrades
|
[
{
"change_type": "MODIFY",
"old_path": "pkg/client/openstack/kluster/client.go",
"new_path": "pkg/client/openstack/kluster/client.go",
"diff": "@@ -3,6 +3,7 @@ package kluster\nimport (\n\"fmt\"\n\"os\"\n+ \"regexp\"\n\"strings\"\n\"github.com/gophercloud/gophercloud\"\n@@ -13,6 +14,7 @@ import (\nsecuritygroups \"github.com/gophercloud/gophercloud/openstack/networking/v2/extensions/security/groups\"\n\"github.com/gophercloud/gophercloud/openstack/networking/v2/extensions/security/rules\"\n\"github.com/gophercloud/gophercloud/pagination\"\n+ \"github.com/pkg/errors\"\n\"github.com/sapcc/kubernikus/pkg/api/models\"\nv1 \"github.com/sapcc/kubernikus/pkg/apis/kubernikus/v1\"\n@@ -102,11 +104,19 @@ func (c *klusterClient) CreateNode(kluster *v1.Kluster, pool *models.NodePool, n\nreturn server.ID, nil\n}\n-func (c *klusterClient) DeleteNode(id string) error {\n+func (c *klusterClient) DeleteNode(providerID string) error {\n+ id, err := instanceIDFromProviderID(providerID)\n+ if err != nil {\n+ return errors.Wrap(err, \"Failed to delete node\")\n+ }\nreturn servers.Delete(c.ComputeClient, id).ExtractErr()\n}\n-func (c *klusterClient) RebootNode(id string) error {\n+func (c *klusterClient) RebootNode(providerID string) error {\n+ id, err := instanceIDFromProviderID(providerID)\n+ if err != nil {\n+ return errors.Wrap(err, \"Failed to reboot node\")\n+ }\nreturn servers.Reboot(c.ComputeClient, id, &servers.RebootOpts{Type: servers.SoftReboot}).ExtractErr()\n}\n@@ -284,3 +294,16 @@ func ExtractServers(r pagination.Page) ([]Node, error) {\nerr := servers.ExtractServersInto(r, &s)\nreturn s, err\n}\n+\n+// instanceIDFromProviderID splits a provider's id and return instanceID.\n+// A providerID is build out of '${ProviderName}:///${instance-id}'which contains ':///'.\n+// See cloudprovider.GetInstanceProviderID and Instances.InstanceID.\n+func instanceIDFromProviderID(providerID string) (instanceID string, err error) {\n+ var providerIDRegexp = regexp.MustCompile(`^openstack:///([^/]+)$`)\n+\n+ matches := providerIDRegexp.FindStringSubmatch(providerID)\n+ if len(matches) != 2 {\n+ return \"\", fmt.Errorf(\"ProviderID \\\"%s\\\" didn't match expected format \\\"openstack:///InstanceID\\\"\", providerID)\n+ }\n+ return matches[1], nil\n+}\n"
},
{
"change_type": "MODIFY",
"old_path": "pkg/controller/servicing/lifecycler.go",
"new_path": "pkg/controller/servicing/lifecycler.go",
"diff": "@@ -160,7 +160,7 @@ func (lc *NodeLifeCycler) Drain(node *core_v1.Node) error {\n// Reboot a node softly\nfunc (lc *NodeLifeCycler) Reboot(node *core_v1.Node) error {\n- if err := lc.Openstack.RebootNode(node.Spec.ExternalID); err != nil {\n+ if err := lc.Openstack.RebootNode(node.Spec.ProviderID); err != nil {\nreturn errors.Wrap(err, \"rebooting node failed\")\n}\n@@ -169,7 +169,7 @@ func (lc *NodeLifeCycler) Reboot(node *core_v1.Node) error {\n// Replace a node by temrinating it\nfunc (lc *NodeLifeCycler) Replace(node *core_v1.Node) error {\n- if err := lc.Openstack.DeleteNode(node.Spec.ExternalID); err != nil {\n+ if err := lc.Openstack.DeleteNode(node.Spec.ProviderID); err != nil {\nreturn errors.Wrap(err, \"deleting node failed\")\n}\nreturn nil\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
prepare for 1.11 upgrades
|
596,240 |
16.04.2019 17:04:14
| -7,200 |
519fffb995215b1ea9ad85e9c936ed043214122f
|
isolate providerId conversion
|
[
{
"change_type": "MODIFY",
"old_path": "pkg/client/openstack/kluster/client.go",
"new_path": "pkg/client/openstack/kluster/client.go",
"diff": "@@ -3,7 +3,6 @@ package kluster\nimport (\n\"fmt\"\n\"os\"\n- \"regexp\"\n\"strings\"\n\"github.com/gophercloud/gophercloud\"\n@@ -14,7 +13,6 @@ import (\nsecuritygroups \"github.com/gophercloud/gophercloud/openstack/networking/v2/extensions/security/groups\"\n\"github.com/gophercloud/gophercloud/openstack/networking/v2/extensions/security/rules\"\n\"github.com/gophercloud/gophercloud/pagination\"\n- \"github.com/pkg/errors\"\n\"github.com/sapcc/kubernikus/pkg/api/models\"\nv1 \"github.com/sapcc/kubernikus/pkg/apis/kubernikus/v1\"\n@@ -104,19 +102,12 @@ func (c *klusterClient) CreateNode(kluster *v1.Kluster, pool *models.NodePool, n\nreturn server.ID, nil\n}\n-func (c *klusterClient) DeleteNode(providerID string) error {\n- id, err := instanceIDFromProviderID(providerID)\n- if err != nil {\n- return errors.Wrap(err, \"Failed to delete node\")\n- }\n+func (c *klusterClient) DeleteNode(id string) error {\n+\nreturn servers.Delete(c.ComputeClient, id).ExtractErr()\n}\n-func (c *klusterClient) RebootNode(providerID string) error {\n- id, err := instanceIDFromProviderID(providerID)\n- if err != nil {\n- return errors.Wrap(err, \"Failed to reboot node\")\n- }\n+func (c *klusterClient) RebootNode(id string) error {\nreturn servers.Reboot(c.ComputeClient, id, &servers.RebootOpts{Type: servers.SoftReboot}).ExtractErr()\n}\n@@ -294,16 +285,3 @@ func ExtractServers(r pagination.Page) ([]Node, error) {\nerr := servers.ExtractServersInto(r, &s)\nreturn s, err\n}\n-\n-// instanceIDFromProviderID splits a provider's id and return instanceID.\n-// A providerID is build out of '${ProviderName}:///${instance-id}'which contains ':///'.\n-// See cloudprovider.GetInstanceProviderID and Instances.InstanceID.\n-func instanceIDFromProviderID(providerID string) (instanceID string, err error) {\n- var providerIDRegexp = regexp.MustCompile(`^openstack:///([^/]+)$`)\n-\n- matches := providerIDRegexp.FindStringSubmatch(providerID)\n- if len(matches) != 2 {\n- return \"\", fmt.Errorf(\"ProviderID \\\"%s\\\" didn't match expected format \\\"openstack:///InstanceID\\\"\", providerID)\n- }\n- return matches[1], nil\n-}\n"
},
{
"change_type": "MODIFY",
"old_path": "pkg/controller/servicing/lifecycler.go",
"new_path": "pkg/controller/servicing/lifecycler.go",
"diff": "@@ -2,6 +2,7 @@ package servicing\nimport (\n\"fmt\"\n+ \"regexp\"\n\"time\"\n\"github.com/go-kit/kit/log\"\n@@ -160,7 +161,12 @@ func (lc *NodeLifeCycler) Drain(node *core_v1.Node) error {\n// Reboot a node softly\nfunc (lc *NodeLifeCycler) Reboot(node *core_v1.Node) error {\n- if err := lc.Openstack.RebootNode(node.Spec.ProviderID); err != nil {\n+ id, err := instanceIDFromProviderID(node.Spec.ProviderID)\n+ if err != nil {\n+ return errors.Wrap(err, \"rebooting node failed\")\n+ }\n+\n+ if err := lc.Openstack.RebootNode(id); err != nil {\nreturn errors.Wrap(err, \"rebooting node failed\")\n}\n@@ -169,7 +175,12 @@ func (lc *NodeLifeCycler) Reboot(node *core_v1.Node) error {\n// Replace a node by temrinating it\nfunc (lc *NodeLifeCycler) Replace(node *core_v1.Node) error {\n- if err := lc.Openstack.DeleteNode(node.Spec.ProviderID); err != nil {\n+ id, err := instanceIDFromProviderID(node.Spec.ProviderID)\n+ if err != nil {\n+ return errors.Wrap(err, \"deleting node failed\")\n+ }\n+\n+ if err := lc.Openstack.DeleteNode(id); err != nil {\nreturn errors.Wrap(err, \"deleting node failed\")\n}\nreturn nil\n@@ -425,3 +436,16 @@ func (lc *InstrumentingLifeCycler) Uncordon(node *core_v1.Node) (err error) {\n}(time.Now())\nreturn lc.LifeCycler.Uncordon(node)\n}\n+\n+// instanceIDFromProviderID splits a provider's id and return instanceID.\n+// A providerID is build out of '${ProviderName}:///${instance-id}'which contains ':///'.\n+// See cloudprovider.GetInstanceProviderID and Instances.InstanceID.\n+func instanceIDFromProviderID(providerID string) (instanceID string, err error) {\n+ var providerIDRegexp = regexp.MustCompile(`^openstack:///([^/]+)$`)\n+\n+ matches := providerIDRegexp.FindStringSubmatch(providerID)\n+ if len(matches) != 2 {\n+ return \"\", fmt.Errorf(\"ProviderID \\\"%s\\\" didn't match expected format \\\"openstack:///InstanceID\\\"\", providerID)\n+ }\n+ return matches[1], nil\n+}\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
isolate providerId conversion
|
596,226 |
17.04.2019 17:09:37
| -7,200 |
deeb6b6f811d8ef68dbd6af1e90d3e3626557175
|
Change etcd backup policy to exponential
|
[
{
"change_type": "MODIFY",
"old_path": "charts/kube-master/Chart.yaml",
"new_path": "charts/kube-master/Chart.yaml",
"diff": "apiVersion: v1\ndescription: A Helm chart for Kubernetes\nname: kube-master\n-version: 2.0.0\n+version: 2.0.1\n"
},
{
"change_type": "MODIFY",
"old_path": "charts/kube-master/charts/etcd/templates/deployment.yaml",
"new_path": "charts/kube-master/charts/etcd/templates/deployment.yaml",
"diff": "@@ -121,7 +121,9 @@ spec:\n- etcdbrctl\n- server\n- --schedule={{ .Values.backup.schedule }}\n+ {{- if .Values.backup.maxBackups }}\n- --max-backups={{ .Values.backup.maxBackups }}\n+ {{- end }}\n- --data-dir=/var/lib/etcd/new.etcd\n- --insecure-transport=true\n- --storage-provider=Swift\n"
},
{
"change_type": "MODIFY",
"old_path": "charts/kube-master/charts/etcd/values.yaml",
"new_path": "charts/kube-master/charts/etcd/values.yaml",
"diff": "@@ -27,14 +27,15 @@ backup:\npullPolicy: IfNotPresent\n# do a full-backup every hour\nschedule: \"15 * * * *\"\n- # keep backups for one week\n- maxBackups: 168\n- # delta-snapshot every 10 seconds\n- deltaSnapshotPeriod: 10\n- # clean-up old backups every 5 minutes\n- garbageCollectionPeriod: 300\n- # keep maxBackups\n- garbageCollectionPolicy: \"LimitBased\"\n+ # keep number of backups\n+ # only used if garbageCollectionPolicy is LimitBased\n+ # maxBackups: 168\n+ # delta-snapshot every 30 seconds\n+ deltaSnapshotPeriod: 30\n+ # clean-up old backups every hour\n+ garbageCollectionPeriod: 3600\n+ # condense in time\n+ garbageCollectionPolicy: \"Exponential\"\nresources:\nrequests:\ncpu: 100m\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
Change etcd backup policy to exponential (#418)
|
596,240 |
18.04.2019 10:45:23
| -7,200 |
57239ae8efe2befb0629da0e582a61509360b16a
|
makes API more forgiving for nodepool config
|
[
{
"change_type": "MODIFY",
"old_path": "pkg/api/handlers/create_cluster.go",
"new_path": "pkg/api/handlers/create_cluster.go",
"diff": "@@ -66,13 +66,23 @@ func (d *createCluster) Handle(params operations.CreateClusterParams, principal\nspec.NodePools[i].AvailabilityZone = defaultAVZ\n}\n+ allowReboot := true\n+ allowReplace := true\nif pool.Config == nil {\nspec.NodePools[i].Config = &models.NodePoolConfig{\n- AllowReboot: true,\n- AllowReplace: true,\n+ AllowReboot: &allowReboot,\n+ AllowReplace: &allowReplace,\n}\n}\n+ if spec.NodePools[i].Config.AllowReboot == nil {\n+ spec.NodePools[i].Config.AllowReboot = &allowReboot\n+ }\n+\n+ if spec.NodePools[i].Config.AllowReplace == nil {\n+ spec.NodePools[i].Config.AllowReplace = &allowReplace\n+ }\n+\n// Validate AVZ\nif err := validateAavailabilityZone(spec.NodePools[i].AvailabilityZone, metadata); err != nil {\nreturn NewErrorResponse(&operations.CreateClusterDefault{}, 409, \"Availability Zone %s is invalid: %s\", spec.NodePools[i].AvailabilityZone, err)\n"
},
{
"change_type": "MODIFY",
"old_path": "pkg/api/handlers/update_cluster.go",
"new_path": "pkg/api/handlers/update_cluster.go",
"diff": "@@ -72,6 +72,14 @@ func (d *updateCluster) Handle(params operations.UpdateClusterParams, principal\nif paramPool.Config == nil {\nnodePools[i].Config = specPool.Config\n+ } else {\n+ if paramPool.Config.AllowReboot == nil {\n+ nodePools[i].Config.AllowReboot = specPool.Config.AllowReboot\n+ }\n+\n+ if paramPool.Config.AllowReplace == nil {\n+ nodePools[i].Config.AllowReplace = specPool.Config.AllowReplace\n+ }\n}\n}\n}\n@@ -83,11 +91,21 @@ func (d *updateCluster) Handle(params operations.UpdateClusterParams, principal\nnodePools[i].AvailabilityZone = defaultAVZ\n}\n+ allowReboot := true\n+ allowReplace := true\nif paramPool.Config == nil {\nnodePools[i].Config = &models.NodePoolConfig{\n- AllowReboot: true,\n- AllowReplace: true,\n+ AllowReboot: &allowReboot,\n+ AllowReplace: &allowReplace,\n+ }\n}\n+\n+ if nodePools[i].Config.AllowReboot == nil {\n+ nodePools[i].Config.AllowReboot = &allowReboot\n+ }\n+\n+ if nodePools[i].Config.AllowReplace == nil {\n+ nodePools[i].Config.AllowReplace = &allowReplace\n}\nif err := validateAavailabilityZone(nodePools[i].AvailabilityZone, metadata); err != nil {\n"
},
{
"change_type": "MODIFY",
"old_path": "pkg/api/models/node_pool_config.go",
"new_path": "pkg/api/models/node_pool_config.go",
"diff": "@@ -16,10 +16,10 @@ import (\ntype NodePoolConfig struct {\n// Allow automatic drain and reboot of nodes. Enables OS updates. Required by security policy.\n- AllowReboot bool `json:\"allowReboot,omitempty\"`\n+ AllowReboot *bool `json:\"allowReboot\"`\n// Allow automatic drain and replacement of nodes. Enables Kubernetes upgrades.\n- AllowReplace bool `json:\"allowReplace,omitempty\"`\n+ AllowReplace *bool `json:\"allowReplace\"`\n}\n// Validate validates this node pool config\n"
},
{
"change_type": "MODIFY",
"old_path": "pkg/api/rest/api_test.go",
"new_path": "pkg/api/rest/api_test.go",
"diff": "@@ -191,6 +191,9 @@ func TestClusterUpdate(t *testing.T) {\n}, nil\n}\n+ on := true\n+ off := false\n+\nkluster := kubernikusv1.Kluster{\nObjectMeta: metav1.ObjectMeta{\nName: fmt.Sprintf(\"%s-%s\", \"nase\", ACCOUNT),\n@@ -217,8 +220,8 @@ func TestClusterUpdate(t *testing.T) {\nName: \"poolname\",\nSize: 2,\nConfig: &models.NodePoolConfig{\n- AllowReboot: false,\n- AllowReplace: false,\n+ AllowReboot: &off,\n+ AllowReplace: &off,\n},\n},\n},\n@@ -252,8 +255,8 @@ func TestClusterUpdate(t *testing.T) {\nName: \"poolname\",\nSize: 5,\nConfig: &models.NodePoolConfig{\n- AllowReboot: true,\n- AllowReplace: true,\n+ AllowReboot: &on,\n+ AllowReplace: &on,\n},\n},\n{\n@@ -263,8 +266,8 @@ func TestClusterUpdate(t *testing.T) {\nName: \"newpoolname\",\nSize: 3,\nConfig: &models.NodePoolConfig{\n- AllowReboot: true,\n- AllowReplace: true,\n+ AllowReboot: &on,\n+ AllowReplace: &on,\n},\n},\n},\n"
},
{
"change_type": "MODIFY",
"old_path": "pkg/api/spec/embedded_spec.go",
"new_path": "pkg/api/spec/embedded_spec.go",
"diff": "@@ -609,12 +609,14 @@ func init() {\n\"allowReboot\": {\n\"description\": \"Allow automatic drain and reboot of nodes. Enables OS updates. Required by security policy.\",\n\"type\": \"boolean\",\n- \"x-nullable\": false\n+ \"x-nullable\": true,\n+ \"x-omitempty\": false\n},\n\"allowReplace\": {\n\"description\": \"Allow automatic drain and replacement of nodes. Enables Kubernetes upgrades.\",\n\"type\": \"boolean\",\n- \"x-nullable\": false\n+ \"x-nullable\": true,\n+ \"x-omitempty\": false\n}\n},\n\"x-nullable\": true\n@@ -1466,12 +1468,14 @@ func init() {\n\"allowReboot\": {\n\"description\": \"Allow automatic drain and reboot of nodes. Enables OS updates. Required by security policy.\",\n\"type\": \"boolean\",\n- \"x-nullable\": false\n+ \"x-nullable\": true,\n+ \"x-omitempty\": false\n},\n\"allowReplace\": {\n\"description\": \"Allow automatic drain and replacement of nodes. Enables Kubernetes upgrades.\",\n\"type\": \"boolean\",\n- \"x-nullable\": false\n+ \"x-nullable\": true,\n+ \"x-omitempty\": false\n}\n},\n\"x-nullable\": true\n"
},
{
"change_type": "MODIFY",
"old_path": "pkg/controller/servicing/lister.go",
"new_path": "pkg/controller/servicing/lister.go",
"diff": "@@ -113,7 +113,7 @@ func (d *NodeLister) Reboot() []*core_v1.Node {\nvar rebootable, found []*core_v1.Node\nfor _, pool := range d.Kluster.Spec.NodePools {\n- if !pool.Config.AllowReboot {\n+ if *pool.Config.AllowReboot == false {\ncontinue\n}\n@@ -151,7 +151,7 @@ func (d *NodeLister) Replace() []*core_v1.Node {\nvar upgradable, found []*core_v1.Node\nfor _, pool := range d.Kluster.Spec.NodePools {\n- if !pool.Config.AllowReplace {\n+ if *pool.Config.AllowReplace == false {\ncontinue\n}\n"
},
{
"change_type": "MODIFY",
"old_path": "pkg/controller/servicing/testing.go",
"new_path": "pkg/controller/servicing/testing.go",
"diff": "@@ -74,11 +74,13 @@ func NewFakeKluster(opts *FakeKlusterOptions) (*v1.Kluster, []runtime.Object) {\nfor i, p := range opts.NodePools {\npoolName := fmt.Sprintf(\"pool%d\", i)\n+ allowReboot := p.AllowReboot\n+ allowReplace := p.AllowReplace\npool := models.NodePool{\nName: poolName,\nConfig: &models.NodePoolConfig{\n- AllowReplace: p.AllowReboot,\n- AllowReboot: p.AllowReplace,\n+ AllowReplace: &allowReboot,\n+ AllowReboot: &allowReplace,\n},\n}\nkluster.Spec.NodePools = append(kluster.Spec.NodePools, pool)\n"
},
{
"change_type": "MODIFY",
"old_path": "swagger.yml",
"new_path": "swagger.yml",
"diff": "@@ -491,11 +491,13 @@ definitions:\nproperties:\nallowReboot:\ndescription: Allow automatic drain and reboot of nodes. Enables OS updates. Required by security policy.\n- x-nullable: false\n+ x-nullable: true\n+ x-omitempty: false\ntype: boolean\nallowReplace:\ndescription: Allow automatic drain and replacement of nodes. Enables Kubernetes upgrades.\n- x-nullable: false\n+ x-nullable: true\n+ x-omitempty: false\ntype: boolean\nKlusterStatus:\nreadOnly: true\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
makes API more forgiving for nodepool config
|
596,240 |
18.04.2019 11:22:16
| -7,200 |
af68f13e2a4371d39d07e54562b3084c27975915
|
enable upgrades
|
[
{
"change_type": "MODIFY",
"old_path": "pkg/controller/servicing/reconciler.go",
"new_path": "pkg/controller/servicing/reconciler.go",
"diff": "@@ -117,8 +117,7 @@ func (r *KlusterReconciler) Do() error {\nreturn nil\n}\n- // Default to skip klusters without the servicing annotation\n- if !util.EnabledValue(r.Kluster.ObjectMeta.Annotations[AnnotationServicingSafeguard]) {\n+ if util.DisabledValue(r.Kluster.ObjectMeta.Annotations[AnnotationServicingSafeguard]) {\nr.Logger.Log(\"msg\", \"Skippig upgrades. Manually disabled with safeguard annotation.\")\nreturn nil\n}\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
enable upgrades
|
596,240 |
18.04.2019 14:05:00
| -7,200 |
feb8378277975639115527d6f1d760a92f2999c1
|
migrate null configs to defaults
|
[
{
"change_type": "ADD",
"old_path": null,
"new_path": "pkg/migration/15_default_node_pool_options.go",
"diff": "+package migration\n+\n+import (\n+ \"github.com/sapcc/kubernikus/pkg/api/models\"\n+ v1 \"github.com/sapcc/kubernikus/pkg/apis/kubernikus/v1\"\n+ \"github.com/sapcc/kubernikus/pkg/controller/config\"\n+)\n+\n+func ReconcileNodePoolConfigDefaults(rawKluster []byte, current *v1.Kluster, clients config.Clients, factories config.Factories) (err error) {\n+ for i := range current.Spec.NodePools {\n+ allowReboot := true\n+ allowReplace := true\n+ if current.Spec.NodePools[i].Config == nil {\n+ current.Spec.NodePools[i].Config = &models.NodePoolConfig{\n+ AllowReboot: &allowReboot,\n+ AllowReplace: &allowReplace,\n+ }\n+ continue\n+ }\n+\n+ if current.Spec.NodePools[i].Config.AllowReboot == nil {\n+ current.Spec.NodePools[i].Config.AllowReboot = &allowReboot\n+ }\n+\n+ if current.Spec.NodePools[i].Config.AllowReplace == nil {\n+ current.Spec.NodePools[i].Config.AllowReplace = &allowReplace\n+ }\n+ }\n+ return\n+}\n"
},
{
"change_type": "MODIFY",
"old_path": "pkg/migration/register.go",
"new_path": "pkg/migration/register.go",
"diff": "@@ -21,6 +21,7 @@ func init() {\nNoOp,\nFixRootCertificate,\nCleanupSuppositoryNamespaces,\n+ ReconcileNodePoolConfigDefaults\n// <-- Insert new migrations at the end only!\n}\n}\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
migrate null configs to defaults
|
596,240 |
24.04.2019 15:21:57
| -7,200 |
96fc44e6f623a2a3a16cb32d01518bc17c2390d6
|
adds holdoff timer for coreos releases
|
[
{
"change_type": "MODIFY",
"old_path": "pkg/controller/servicing/controller_test.go",
"new_path": "pkg/controller/servicing/controller_test.go",
"diff": "@@ -304,6 +304,7 @@ func TestServicingControllerReconcile(t *testing.T) {\nLogger: logger,\nNodeObservatory: nodeobservatory.NewFakeController(kluster, nodes...),\nCoreOSVersion: &coreos.Version{},\n+ CoreOSRelease: coreos.NewFakeRelease(t, \"2079.3.0\"),\n}\nreconcilers := &KlusterReconcilerFactory{\n"
},
{
"change_type": "MODIFY",
"old_path": "pkg/controller/servicing/coreos/testing.go",
"new_path": "pkg/controller/servicing/coreos/testing.go",
"diff": "@@ -17,6 +17,25 @@ func NewFakeVersion(t *testing.T, version string) *Version {\nreturn subject\n}\n+func NewFakeRelease(t *testing.T, version string) *Release {\n+ body := fmt.Sprintf(`\n+ {\n+ \"%s\": {\n+ \"release_date\": \"2019-01-23 20:20:00 +0000\",\n+ \"major_software\": {\n+ \"kernel\": [\"4.19.34\"],\n+ \"docker\": [\"18.06.3\"],\n+ \"etcd\": [\"3.3.12\"],\n+ \"fleet\": [\"\"]\n+ },\n+ \"release_notes\": \"No updates\"\n+ }\n+ }`, version)\n+ subject := &Release{}\n+ subject.Client = NewTestClient(t, \"https://coreos.com/releases/releases-stable.json\", body, nil)\n+ return subject\n+}\n+\nfunc NewTestClient(t *testing.T, baseURL, body string, count *int) *http.Client {\nfn := func(req *http.Request) *http.Response {\nassert.Equal(t, req.URL.String(), baseURL)\n"
},
{
"change_type": "MODIFY",
"old_path": "pkg/controller/servicing/lister.go",
"new_path": "pkg/controller/servicing/lister.go",
"diff": "@@ -44,6 +44,7 @@ type (\nLogger log.Logger\nNodeObservatory *nodeobservatory.NodeObservatory\nCoreOSVersion *coreos.Version\n+ CoreOSRelease *coreos.Release\n}\n// NodeLister knows how to figure out the state of Nodes\n@@ -52,6 +53,7 @@ type (\nKluster *v1.Kluster\nLister listers_core_v1.NodeLister\nCoreOSVersion *coreos.Version\n+ CoreOSRelease *coreos.Release\n}\n// LoggingLister writes log messages\n@@ -67,6 +69,7 @@ func NewNodeListerFactory(logger log.Logger, recorder record.EventRecorder, fact\nLogger: logger,\nNodeObservatory: factories.NodesObservatory.NodeInformer(),\nCoreOSVersion: &coreos.Version{},\n+ CoreOSRelease: &coreos.Release{},\n}\n}\n@@ -85,6 +88,7 @@ func (f *NodeListerFactory) Make(k *v1.Kluster) (Lister, error) {\nKluster: k,\nLister: klusterLister,\nCoreOSVersion: f.CoreOSVersion,\n+ CoreOSRelease: f.CoreOSRelease,\n}\nlister = &LoggingLister{\n@@ -112,6 +116,26 @@ func (d *NodeLister) All() []*core_v1.Node {\nfunc (d *NodeLister) Reboot() []*core_v1.Node {\nvar rebootable, found []*core_v1.Node\n+ latest, err := d.CoreOSVersion.Stable()\n+ if err != nil {\n+ d.Logger.Log(\n+ \"msg\", \"Couldn't get CoreOS version.\",\n+ \"err\", err,\n+ )\n+ return found\n+ }\n+\n+ released, err := d.CoreOSRelease.GrownUp(latest)\n+ if err != nil {\n+ d.Logger.Log(\n+ \"msg\", \"Couldn't get CoreOS releases.\",\n+ \"err\", err,\n+ )\n+ }\n+ if !released {\n+ return found\n+ }\n+\nfor _, pool := range d.Kluster.Spec.NodePools {\nif *pool.Config.AllowReboot == false {\ncontinue\n"
},
{
"change_type": "MODIFY",
"old_path": "pkg/controller/servicing/lister_test.go",
"new_path": "pkg/controller/servicing/lister_test.go",
"diff": "@@ -32,6 +32,7 @@ func NewFakeNodeLister(t *testing.T, logger log.Logger, kluster *v1.Kluster, nod\nKluster: kluster,\nLister: kl,\nCoreOSVersion: coreos.NewFakeVersion(t, \"2023.4.0\"),\n+ CoreOSRelease: coreos.NewFakeRelease(t, \"2023.4.0\"),\n}\nlister = &LoggingLister{\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
adds holdoff timer for coreos releases
|
596,240 |
24.04.2019 16:30:09
| -7,200 |
05d5b5887d170d8e02845329910d7c3f62e4ab4d
|
adds missing release files
|
[
{
"change_type": "ADD",
"old_path": null,
"new_path": "pkg/controller/servicing/coreos/releases.go",
"diff": "+package coreos\n+\n+import (\n+ \"encoding/json\"\n+ \"fmt\"\n+ \"io/ioutil\"\n+ \"net/http\"\n+ \"time\"\n+\n+ \"github.com/pkg/errors\"\n+\n+ \"github.com/sapcc/kubernikus/pkg/util/version\"\n+)\n+\n+var (\n+ baseURL = \"https://coreos.com/releases/releases-%s.json\"\n+ holdoff = 7 * 24 * time.Hour\n+ fetchInterval = 1 * time.Hour\n+)\n+\n+type Release struct {\n+ Client *http.Client\n+ ReleaseDates map[channel]*releaseDate\n+}\n+\n+type releaseDate struct {\n+ Releases map[string]time.Time\n+ FetchedAt time.Time\n+}\n+\n+func (r *Release) releasedAt(c channel, v *version.Version) (time.Time, error) {\n+ if r.Client == nil {\n+ r.Client = &http.Client{\n+ Timeout: time.Second * 10,\n+ }\n+ }\n+\n+ if r.ReleaseDates == nil {\n+ r.ReleaseDates = make(map[channel]*releaseDate)\n+ }\n+\n+ if r.ReleaseDates[c] == nil || now().After(r.ReleaseDates[c].FetchedAt.Add(fetchInterval)) {\n+ result, err := r.fetch(c)\n+ if err != nil {\n+ return now(), errors.Wrapf(err, \"Couldn't fetch %s releases\", c)\n+ }\n+ r.ReleaseDates[c] = result\n+ }\n+\n+ released, ok := r.ReleaseDates[c].Releases[v.String()]\n+ if !ok {\n+ return now(), errors.Errorf(\"Version %s not found in %s releases\", v, c)\n+ }\n+\n+ return released, nil\n+}\n+\n+func (r *Release) fetch(c channel) (*releaseDate, error) {\n+ result, err := r.Client.Get(fmt.Sprintf(baseURL, c))\n+ if err != nil {\n+ return nil, fmt.Errorf(\"Couldn't fetch %s releases: %s\", c, err)\n+ }\n+\n+ defer result.Body.Close()\n+\n+ body, err := ioutil.ReadAll(result.Body)\n+ if err != nil {\n+ return nil, errors.Wrap(err, \"Couldn't read body\")\n+ }\n+\n+ var rdates map[string]interface{}\n+\n+ err = json.Unmarshal(body, &rdates)\n+ if err != nil {\n+ return nil, errors.Wrap(err, \"Couldn't parse response\")\n+ }\n+\n+ releaseDates := &releaseDate{\n+ Releases: map[string]time.Time{},\n+ FetchedAt: now(),\n+ }\n+\n+ for k, v := range rdates {\n+ version, err := version.ParseSemantic(k)\n+ if err != nil {\n+ return nil, errors.Wrap(err, \"Couldn't parse version\")\n+ }\n+\n+ t, err := time.Parse(\"2006-01-02 15:04:05 +0000\", v.(map[string]interface{})[\"release_date\"].(string))\n+ if err != nil {\n+ return nil, errors.Wrap(err, \"Couldn't parse release date\")\n+ }\n+\n+ releaseDates.Releases[version.String()] = t\n+\n+ }\n+\n+ return releaseDates, nil\n+}\n+\n+func (r *Release) GrownUp(v *version.Version) (bool, error) {\n+ released, err := r.releasedAt(stable, v)\n+ if err != nil {\n+ return false, err\n+ }\n+ return released.Add(holdoff).Before(now()), nil\n+}\n"
},
{
"change_type": "ADD",
"old_path": null,
"new_path": "pkg/controller/servicing/coreos/releases_test.go",
"diff": "+package coreos\n+\n+import (\n+ \"testing\"\n+ \"time\"\n+\n+ \"github.com/stretchr/testify/assert\"\n+\n+ \"github.com/sapcc/kubernikus/pkg/util/version\"\n+)\n+\n+func TestReleaseGrownup(t *testing.T) {\n+ now = func() time.Time { return time.Date(2019, 4, 23, 20, 21, 0, 0, time.UTC) }\n+ count := 0\n+ subject := &Release{}\n+ subject.Client = NewTestClient(t, \"https://coreos.com/releases/releases-stable.json\", ReleasesStable, &count)\n+\n+ t.Run(\"fetches versions\", func(t *testing.T) {\n+ _, err := subject.GrownUp(version.MustParseSemantic(\"2079.3.0\"))\n+ assert.NoError(t, err)\n+ })\n+\n+ t.Run(\"unknown version\", func(t *testing.T) {\n+ result, err := subject.GrownUp(version.MustParseSemantic(\"2079.99.0\"))\n+ assert.Error(t, err)\n+ assert.False(t, result)\n+ })\n+\n+ t.Run(\"holdoff time not up yet\", func(t *testing.T) {\n+ result, err := subject.GrownUp(version.MustParseSemantic(\"2079.3.0\"))\n+ assert.NoError(t, err)\n+ assert.False(t, result)\n+ })\n+\n+ t.Run(\"holdoff time up\", func(t *testing.T) {\n+ now = func() time.Time { return time.Date(2019, 5, 23, 20, 21, 0, 0, time.UTC) }\n+ result, err := subject.GrownUp(version.MustParseSemantic(\"2079.3.0\"))\n+ assert.NoError(t, err)\n+ assert.True(t, result)\n+ })\n+}\n+\n+const (\n+ ReleasesStable = `\n+ {\n+ \"2079.3.0\": {\n+ \"release_date\": \"2019-04-23 20:20:00 +0000\",\n+ \"major_software\": {\n+ \"kernel\": [\"4.19.34\"],\n+ \"docker\": [\"18.06.3\"],\n+ \"etcd\": [\"3.3.12\"],\n+ \"fleet\": [\"\"]\n+ },\n+ \"release_notes\": \"No updates\"\n+ },\n+\n+ \"2023.5.0\": {\n+ \"release_date\": \"2019-03-11 23:13:36 +0000\",\n+ \"major_software\": {\n+ \"kernel\": [\"4.19.25\"],\n+ \"docker\": [\"18.06.1\"],\n+ \"etcd\": [\"3.3.12\"],\n+ \"fleet\": [\"\"]\n+ },\n+ \"release_notes\": \"No updates\"\n+ },\n+\n+ \"2023.4.0\": {\n+ \"release_date\": \"2019-02-26 16:45:45 +0000\",\n+ \"major_software\": {\n+ \"kernel\": [\"4.19.23\"],\n+ \"docker\": [\"18.06.1\"],\n+ \"etcd\": [\"3.3.12\"],\n+ \"fleet\": [\"\"]\n+ },\n+ \"release_notes\": \"No updates\"\n+ },\n+\n+ \"1967.6.0\": {\n+ \"release_date\": \"2019-02-12 23:15:19 +0000\",\n+ \"major_software\": {\n+ \"kernel\": [\"4.14.96\"],\n+ \"docker\": [\"18.06.1\"],\n+ \"etcd\": [\"3.3.10\"],\n+ \"fleet\": [\"\"]\n+ },\n+ \"release_notes\": \"No updates\"\n+ },\n+\n+ \"1967.5.0\": {\n+ \"release_date\": \"2019-02-11 20:11:19 +0000\",\n+ \"major_software\": {\n+ \"kernel\": [\"4.14.96\"],\n+ \"docker\": [\"18.06.1\"],\n+ \"etcd\": [\"3.3.10\"],\n+ \"fleet\": [\"\"]\n+ },\n+ \"release_notes\": \"No updates\"\n+ },\n+\n+ \"1967.4.0\": {\n+ \"release_date\": \"2019-01-29 15:08:12 +0000\",\n+ \"major_software\": {\n+ \"kernel\": [\"4.14.96\"],\n+ \"docker\": [\"18.06.1\"],\n+ \"etcd\": [\"3.3.10\"],\n+ \"fleet\": [\"\"]\n+ },\n+ \"release_notes\": \"No updates\"\n+ },\n+\n+ \"1967.3.0\": {\n+ \"release_date\": \"2019-01-08 19:31:47 +0000\",\n+ \"major_software\": {\n+ \"kernel\": [\"4.14.88\"],\n+ \"docker\": [\"18.06.1\"],\n+ \"etcd\": [\"3.3.10\"],\n+ \"fleet\": [\"\"]\n+ },\n+ \"release_notes\": \"No updates\"\n+ },\n+\n+ \"1911.5.0\": {\n+ \"release_date\": \"2018-12-19 16:11:02 +0000\",\n+ \"major_software\": {\n+ \"kernel\": [\"4.14.84\"],\n+ \"docker\": [\"18.06.1\"],\n+ \"etcd\": [\"3.3.9\"],\n+ \"fleet\": [\"\"]\n+ },\n+ \"release_notes\":\"\"\n+ },\n+\n+ \"1911.4.0\": {\n+ \"release_date\": \"2018-11-26 21:49:53 +0000\",\n+ \"major_software\": {\n+ \"kernel\": [\"4.14.81\"],\n+ \"docker\": [\"18.06.1\"],\n+ \"etcd\": [\"3.3.9\"],\n+ \"fleet\": [\"\"]\n+ },\n+ \"release_notes\": \"No updates\"\n+ },\n+\n+ \"1911.3.0\": {\n+ \"release_date\": \"2018-11-06 17:54:59 +0000\",\n+ \"major_software\": {\n+ \"kernel\": [\"4.14.78\"],\n+ \"docker\": [\"18.06.1\"],\n+ \"etcd\": [\"3.3.9\"],\n+ \"fleet\": [\"\"]\n+ },\n+ \"release_notes\": \"No updates\"\n+ },\n+\n+ \"1855.5.0\": {\n+ \"release_date\": \"2018-10-24 13:25:03 +0000\",\n+ \"major_software\": {\n+ \"kernel\": [\"4.14.74\"],\n+ \"docker\": [\"18.06.1\"],\n+ \"etcd\": [\"3.3.9\"],\n+ \"fleet\": [\"\"]\n+ },\n+ \"release_notes\": \"No updates\"\n+ },\n+\n+ \"1855.4.0\": {\n+ \"release_date\": \"2018-09-11 18:48:31 +0000\",\n+ \"major_software\": {\n+ \"kernel\": [\"4.14.67\"],\n+ \"docker\": [\"18.06.1\"],\n+ \"etcd\": [\"3.3.9\"],\n+ \"fleet\": [\"\"]\n+ },\n+ \"release_notes\": \"No updates\"\n+ },\n+\n+ \"1800.7.0\": {\n+ \"release_date\": \"2018-08-16 16:40:56 +0000\",\n+ \"major_software\": {\n+ \"kernel\": [\"4.14.63\"],\n+ \"docker\": [\"18.03.1\"],\n+ \"etcd\": [\"3.3.6\"],\n+ \"fleet\": [\"\"]\n+ },\n+ \"release_notes\": \"No updates\"\n+ },\n+\n+ \"1800.6.0\": {\n+ \"release_date\": \"2018-08-06 19:33:31 +0000\",\n+ \"major_software\": {\n+ \"kernel\": [\"4.14.59\"],\n+ \"docker\": [\"18.03.1\"],\n+ \"etcd\": [\"3.3.6\"],\n+ \"fleet\": [\"\"]\n+ },\n+ \"release_notes\": \"No updates\"\n+ },\n+\n+ \"1800.5.0\": {\n+ \"release_date\": \"2018-07-29 23:40:35 +0000\",\n+ \"major_software\": {\n+ \"kernel\": [\"4.14.59\"],\n+ \"docker\": [\"18.03.1\"],\n+ \"etcd\": [\"3.3.6\"],\n+ \"fleet\": [\"\"]\n+ },\n+ \"release_notes\": \"No updates\"\n+ },\n+\n+ \"1800.4.0\": {\n+ \"release_date\": \"2018-07-25 20:37:11 +0000\",\n+ \"major_software\": {\n+ \"kernel\": [\"4.14.55\"],\n+ \"docker\": [\"18.03.1\"],\n+ \"etcd\": [\"3.3.6\"],\n+ \"fleet\": [\"\"]\n+ },\n+ \"release_notes\": \"No updates\"\n+ },\n+\n+ \"1745.7.0\": {\n+ \"release_date\": \"2018-06-14 17:42:52 +0000\",\n+ \"major_software\": {\n+ \"kernel\": [\"4.14.48\"],\n+ \"docker\": [\"18.03.1\"],\n+ \"etcd\": [\"3.3.3\"],\n+ \"fleet\": [\"\"]\n+ },\n+ \"release_notes\": \"No updates\"\n+ },\n+\n+ \"1745.6.0\": {\n+ \"release_date\": \"2018-06-11 20:02:42 +0000\",\n+ \"major_software\": {\n+ \"kernel\": [\"4.14.48\"],\n+ \"docker\": [\"18.03.1\"],\n+ \"etcd\": [\"3.3.3\"],\n+ \"fleet\": [\"\"]\n+ },\n+ \"release_notes\": \"No updates\"\n+ },\n+\n+ \"1745.5.0\": {\n+ \"release_date\": \"2018-05-31 15:22:08 +0000\",\n+ \"major_software\": {\n+ \"kernel\": [\"4.14.44\"],\n+ \"docker\": [\"18.03.1\"],\n+ \"etcd\": [\"3.3.3\"],\n+ \"fleet\": [\"\"]\n+ },\n+ \"release_notes\": \"No updates\"\n+ },\n+\n+ \"1745.4.0\": {\n+ \"release_date\": \"2018-05-24 23:35:28 +0000\",\n+ \"major_software\": {\n+ \"kernel\": [\"4.14.42\"],\n+ \"docker\": [\"18.03.1\"],\n+ \"etcd\": [\"3.3.3\"],\n+ \"fleet\": [\"\"]\n+ },\n+ \"release_notes\": \"No updates\"\n+ },\n+\n+ \"1745.3.1\": {\n+ \"release_date\": \"2018-05-23 20:52:29 +0000\",\n+ \"major_software\": {\n+ \"kernel\": [\"4.14.42\"],\n+ \"docker\": [\"18.03.1\"],\n+ \"etcd\": [\"3.3.3\"],\n+ \"fleet\": [\"\"]\n+ },\n+ \"release_notes\": \"No updates\"\n+ },\n+\n+ \"1688.5.3\": {\n+ \"release_date\": \"2018-04-03 17:11:51 +0000\",\n+ \"major_software\": {\n+ \"kernel\": [\"4.14.32\"],\n+ \"docker\": [\"17.12.1\"],\n+ \"etcd\": [\"3.2.15\"],\n+ \"fleet\": [\"\"]\n+ },\n+ \"release_notes\": \"No updates\"\n+ },\n+\n+ \"1688.4.0\": {\n+ \"release_date\": \"2018-03-27 19:48:42 +0000\",\n+ \"major_software\": {\n+ \"kernel\": [\"4.14.30\"],\n+ \"docker\": [\"17.12.1\"],\n+ \"etcd\": [\"3.2.15\"],\n+ \"fleet\": [\"\"]\n+ },\n+ \"release_notes\": \"No updates\"\n+ },\n+\n+ \"1632.3.0\": {\n+ \"release_date\": \"2018-02-15 00:57:27 +0000\",\n+ \"major_software\": {\n+ \"kernel\": [\"4.14.19\"],\n+ \"docker\": [\"17.09.1\"],\n+ \"etcd\": [\"3.2.11\"],\n+ \"fleet\": [\"\"]\n+ },\n+ \"release_notes\": \"No updates\"\n+ },\n+\n+ \"1632.2.1\": {\n+ \"release_date\": \"2018-02-01 22:17:31 +0000\",\n+ \"major_software\": {\n+ \"kernel\": [\"4.14.16\"],\n+ \"docker\": [\"17.09.1\"],\n+ \"etcd\": [\"3.2.11\"],\n+ \"fleet\": [\"\"]\n+ },\n+ \"release_notes\": \"No updates\"\n+ },\n+\n+ \"1576.5.0\": {\n+ \"release_date\": \"2018-01-05 14:13:41 +0000\",\n+ \"major_software\": {\n+ \"kernel\": [\"4.14.11\"],\n+ \"docker\": [\"17.09.0\"],\n+ \"etcd\": [\"3.2.9\"],\n+ \"fleet\": [\"\"]\n+ },\n+ \"release_notes\": \"No updates\"\n+ },\n+\n+ \"1576.4.0\": {\n+ \"release_date\": \"2017-12-06 21:14:28 +0000\",\n+ \"major_software\": {\n+ \"kernel\": [\"4.13.16\"],\n+ \"docker\": [\"17.09.0\"],\n+ \"etcd\": [\"3.2.9\"],\n+ \"fleet\": [\"\"]\n+ },\n+ \"release_notes\": \"No updates\"\n+ },\n+\n+ \"1520.9.0\": {\n+ \"release_date\": \"2017-11-30 21:34:26 +0000\",\n+ \"major_software\": {\n+ \"kernel\": [\"4.13.16\"],\n+ \"docker\": [\"1.12.6\"],\n+ \"etcd\": [\"3.1.10\"],\n+ \"fleet\": [\"\"]\n+ },\n+ \"release_notes\": \"No updates\"\n+ },\n+\n+ \"1520.8.0\": {\n+ \"release_date\": \"2017-10-26 16:25:30 +0000\",\n+ \"major_software\": {\n+ \"kernel\": [\"4.13.9\"],\n+ \"docker\": [\"1.12.6\"],\n+ \"etcd\": [\"3.1.10\"],\n+ \"fleet\": [\"\"]\n+ },\n+ \"release_notes\": \"No updates\"\n+ },\n+\n+ \"1520.7.0\": {\n+ \"release_date\": \"2017-10-24 21:23:40 +0000\",\n+ \"major_software\": {\n+ \"kernel\": [\"4.13.9\"],\n+ \"docker\": [\"1.12.6\"],\n+ \"etcd\": [\"3.1.10\"],\n+ \"fleet\": [\"\"]\n+ },\n+ \"release_notes\": \"No updates\"\n+ },\n+\n+ \"1520.6.0\": {\n+ \"release_date\": \"2017-10-12 19:45:37 +0000\",\n+ \"major_software\": {\n+ \"kernel\": [\"4.13.5\"],\n+ \"docker\": [\"1.12.6\"],\n+ \"etcd\": [\"3.1.10\"],\n+ \"fleet\": [\"\"]\n+ },\n+ \"release_notes\": \"No updates\"\n+ },\n+\n+ \"1520.5.0\": {\n+ \"release_date\": \"2017-10-11 15:34:30 +0000\",\n+ \"major_software\": {\n+ \"kernel\": [\"4.13.5\"],\n+ \"docker\": [\"1.12.6\"],\n+ \"etcd\": [\"3.1.10\"],\n+ \"fleet\": [\"\"]\n+ },\n+ \"release_notes\": \"No updates\"\n+ },\n+\n+ \"1465.8.0\": {\n+ \"release_date\": \"2017-09-21 01:06:11 +0000\",\n+ \"major_software\": {\n+ \"kernel\": [\"4.12.14\"],\n+ \"docker\": [\"1.12.6\"],\n+ \"etcd\": [\"3.1.8\"],\n+ \"fleet\": [\"\"]\n+ },\n+ \"release_notes\": \"No updates\"\n+ },\n+\n+ \"1465.7.0\": {\n+ \"release_date\": \"2017-09-06 18:22:34 +0000\",\n+ \"major_software\": {\n+ \"kernel\": [\"4.12.10\"],\n+ \"docker\": [\"1.12.6\"],\n+ \"etcd\": [\"3.1.8\"],\n+ \"fleet\": [\"\"]\n+ },\n+ \"release_notes\": \"No updates\"\n+ },\n+\n+ \"1465.6.0\": {\n+ \"release_date\": \"2017-08-16 21:23:30 +0000\",\n+ \"major_software\": {\n+ \"kernel\": [\"4.12.7\"],\n+ \"docker\": [\"1.12.6\"],\n+ \"etcd\": [\"2.3.7\"],\n+ \"fleet\": [\"\"]\n+ },\n+ \"release_notes\": \"No updates\"\n+ },\n+\n+ \"1409.9.0\": {\n+ \"release_date\": \"2017-08-14 18:54:14 +0000\",\n+ \"major_software\": {\n+ \"kernel\": [\"4.11.12\"],\n+ \"docker\": [\"1.12.6\"],\n+ \"etcd\": [\"2.3.7\"],\n+ \"fleet\": [\"\"]\n+ },\n+ \"release_notes\": \"No updates\"\n+ },\n+\n+ \"1409.8.0\": {\n+ \"release_date\": \"2017-08-10 21:32:03 +0000\",\n+ \"major_software\": {\n+ \"kernel\": [\"4.11.12\"],\n+ \"docker\": [\"1.12.6\"],\n+ \"etcd\": [\"2.3.7\"],\n+ \"fleet\": [\"\"]\n+ },\n+ \"release_notes\": \"No updates\"\n+ },\n+\n+ \"1409.7.0\": {\n+ \"release_date\": \"2017-07-19 01:52:37 +0000\",\n+ \"major_software\": {\n+ \"kernel\": [\"4.11.11\"],\n+ \"docker\": [\"1.12.6\"],\n+ \"etcd\": [\"2.3.7\"],\n+ \"fleet\": [\"\"]\n+ },\n+ \"release_notes\": \"No updates\"\n+ },\n+\n+ \"1409.6.0\": {\n+ \"release_date\": \"2017-07-06 01:57:12 +0000\",\n+ \"major_software\": {\n+ \"kernel\": [\"4.11.9\"],\n+ \"docker\": [\"1.12.6\"],\n+ \"etcd\": [\"2.3.7\"],\n+ \"fleet\": [\"\"]\n+ },\n+ \"release_notes\": \"No updates\"\n+ },\n+\n+ \"1409.5.0\": {\n+ \"release_date\": \"2017-06-23 00:21:55 +0000\",\n+ \"major_software\": {\n+ \"kernel\": [\"4.11.6\"],\n+ \"docker\": [\"1.12.6\"],\n+ \"etcd\": [\"2.3.7\"],\n+ \"fleet\": [\"\"]\n+ },\n+ \"release_notes\": \"No updates\"\n+ },\n+\n+ \"1409.2.0\": {\n+ \"release_date\": \"2017-06-20 01:31:54 +0000\",\n+ \"major_software\": {\n+ \"kernel\": [\"4.11.6\"],\n+ \"docker\": [\"1.12.6\"],\n+ \"etcd\": [\"2.3.7\"],\n+ \"fleet\": [\"\"]\n+ },\n+ \"release_notes\": \"No updates\"\n+ },\n+\n+ \"1353.8.0\": {\n+ \"release_date\": \"2017-05-31 00:11:55 +0000\",\n+ \"major_software\": {\n+ \"kernel\": [\"4.9.24\"],\n+ \"docker\": [\"1.12.6\"],\n+ \"etcd\": [\"0.4.9\",\"2.3.7\"],\n+ \"fleet\": [\"\"]\n+ },\n+ \"release_notes\": \"No updates\"\n+ },\n+\n+ \"1353.7.0\": {\n+ \"release_date\": \"2017-04-26 23:39:29 +0000\",\n+ \"major_software\": {\n+ \"kernel\": [\"4.9.24\"],\n+ \"docker\": [\"1.12.6\"],\n+ \"etcd\": [\"0.4.9\",\"2.3.7\"],\n+ \"fleet\": [\"0.11.8\"]\n+ },\n+ \"release_notes\": \"No updates\"\n+ },\n+\n+ \"1353.6.0\": {\n+ \"release_date\": \"2017-04-25 15:07:03 +0000\",\n+ \"major_software\": {\n+ \"kernel\": [\"4.9.24\"],\n+ \"docker\": [\"1.12.6\"],\n+ \"etcd\": [\"0.4.9\",\"2.3.7\"],\n+ \"fleet\": [\"0.11.8\"]\n+ },\n+ \"release_notes\": \"No updates\"\n+ },\n+\n+ \"1298.7.0\": {\n+ \"release_date\": \"2017-03-31 22:19:13 +0000\",\n+ \"major_software\": {\n+ \"kernel\": [\"4.9.16\"],\n+ \"docker\": [\"1.12.6\"],\n+ \"etcd\": [\"0.4.9\",\"2.3.7\"],\n+ \"fleet\": [\"0.11.8\"]\n+ },\n+ \"release_notes\": \"No updates\"\n+ },\n+\n+ \"1298.6.0\": {\n+ \"release_date\": \"2017-03-15 17:24:28 +0000\",\n+ \"major_software\": {\n+ \"kernel\": [\"4.9.9\"],\n+ \"docker\": [\"1.12.6\"],\n+ \"etcd\": [\"0.4.9\",\"2.3.7\"],\n+ \"fleet\": [\"0.11.8\"]\n+ },\n+ \"release_notes\": \"No updates\"\n+ },\n+\n+ \"1298.5.0\": {\n+ \"release_date\": \"2017-02-28 19:19:47 +0000\",\n+ \"major_software\": {\n+ \"kernel\": [\"4.9.9\"],\n+ \"docker\": [\"1.12.6\"],\n+ \"etcd\": [\"0.4.9\",\"2.3.7\"],\n+ \"fleet\": [\"0.11.8\"]\n+ },\n+ \"release_notes\": \"No updates\"\n+ },\n+\n+ \"1235.12.0\": {\n+ \"release_date\": \"2017-02-23 04:40:09 +0000\",\n+ \"major_software\": {\n+ \"kernel\": [\"4.7.3\"],\n+ \"docker\": [\"1.12.6\"],\n+ \"etcd\": [\"0.4.9\",\"2.3.7\"],\n+ \"fleet\": [\"0.11.8\"]\n+ },\n+ \"release_notes\": \"No updates\"\n+ },\n+\n+ \"1235.9.0\": {\n+ \"release_date\": \"2017-02-02 05:26:47 +0000\",\n+ \"major_software\": {\n+ \"kernel\": [\"4.7.3\"],\n+ \"docker\": [\"1.12.6\"],\n+ \"etcd\": [\"0.4.9\",\"2.3.7\"],\n+ \"fleet\": [\"0.11.8\"]\n+ },\n+ \"release_notes\": \"No updates\"\n+ },\n+\n+ \"1235.8.0\": {\n+ \"release_date\": \"2017-01-31 21:37:38 +0000\",\n+ \"major_software\": {\n+ \"kernel\": [\"4.7.3\"],\n+ \"docker\": [\"1.12.6\"],\n+ \"etcd\": [\"0.4.9\",\"2.3.7\"],\n+ \"fleet\": [\"0.11.8\"]\n+ },\n+ \"release_notes\": \"No updates\"\n+ },\n+\n+ \"1235.6.0\": {\n+ \"release_date\": \"2017-01-11 01:57:20 +0000\",\n+ \"major_software\": {\n+ \"kernel\": [\"4.7.3\"],\n+ \"docker\": [\"1.12.3\"],\n+ \"etcd\": [\"0.4.9\",\"2.3.7\"],\n+ \"fleet\": [\"0.11.8\"]\n+ },\n+ \"release_notes\": \"No updates\"\n+ },\n+\n+ \"1235.5.0\": {\n+ \"release_date\": \"2017-01-08 22:40:31 +0000\",\n+ \"major_software\": {\n+ \"kernel\": [\"4.7.3\"],\n+ \"docker\": [\"1.12.3\"],\n+ \"etcd\": [\"0.4.9\",\"2.3.7\"],\n+ \"fleet\": [\"0.11.8\"]\n+ },\n+ \"release_notes\": \"No updates\"\n+ },\n+\n+ \"1235.4.0\": {\n+ \"release_date\": \"2017-01-04 18:36:08 +0000\",\n+ \"major_software\": {\n+ \"kernel\": [\"4.7.3\"],\n+ \"docker\": [\"1.12.3\"],\n+ \"etcd\": [\"0.4.9\",\"2.3.7\"],\n+ \"fleet\": [\"0.11.8\"]\n+ },\n+ \"release_notes\": \"No updates\"\n+ },\n+\n+ \"1185.5.0\": {\n+ \"release_date\": \"2016-12-07 16:50:38 +0000\",\n+ \"major_software\": {\n+ \"kernel\": [\"4.7.3\"],\n+ \"docker\": [\"1.11.2\"],\n+ \"etcd\": [\"0.4.9\",\"2.3.7\"],\n+ \"fleet\": [\"0.11.8\"]\n+ },\n+ \"release_notes\": \"No updates\"\n+ },\n+\n+ \"1185.3.0\": {\n+ \"release_date\": \"2016-11-01 17:55:29 +0000\",\n+ \"major_software\": {\n+ \"kernel\": [\"4.7.3\"],\n+ \"docker\": [\"1.11.2\"],\n+ \"etcd\": [\"0.4.9\",\"2.3.7\"],\n+ \"fleet\": [\"0.11.8\"]\n+ },\n+ \"release_notes\": \"No updates\"\n+ },\n+\n+ \"1122.3.0\": {\n+ \"release_date\": \"2016-10-20 22:15:18 +0000\",\n+ \"major_software\": {\n+ \"kernel\": [\"4.7.0\"],\n+ \"docker\": [\"1.10.3\"],\n+ \"etcd\": [\"0.4.9\",\"2.3.2\"],\n+ \"fleet\": [\"0.11.7\"]\n+ },\n+ \"release_notes\": \"No updates\"\n+ },\n+\n+ \"1122.2.0\": {\n+ \"release_date\": \"2016-09-06 16:17:57 +0000\",\n+ \"major_software\": {\n+ \"kernel\": [\"4.7.0\"],\n+ \"docker\": [\"1.10.3\"],\n+ \"etcd\": [\"0.4.9\",\"2.3.2\"],\n+ \"fleet\": [\"0.11.7\"]\n+ },\n+ \"release_notes\": \"No updates\"\n+ },\n+\n+ \"1068.10.0\": {\n+ \"release_date\": \"2016-08-23 15:30:28 +0000\",\n+ \"major_software\": {\n+ \"kernel\": [\"4.6.3\"],\n+ \"docker\": [\"1.10.3\"],\n+ \"etcd\": [\"0.4.9\",\"2.3.2\"],\n+ \"fleet\": [\"0.11.7\"]\n+ },\n+ \"release_notes\": \"No updates\"\n+ },\n+\n+ \"1068.9.0\": {\n+ \"release_date\": \"2016-08-09 22:51:46 +0000\",\n+ \"major_software\": {\n+ \"kernel\": [\"4.6.3\"],\n+ \"docker\": [\"1.10.3\"],\n+ \"etcd\": [\"0.4.9\",\"2.3.2\"],\n+ \"fleet\": [\"0.11.7\"]\n+ },\n+ \"release_notes\": \"No updates\"\n+ },\n+\n+ \"1068.8.0\": {\n+ \"release_date\": \"2016-07-18 19:32:58 +0000\",\n+ \"major_software\": {\n+ \"kernel\": [\"4.6.3\"],\n+ \"docker\": [\"1.10.3\"],\n+ \"etcd\": [\"0.4.9\",\"2.3.2\"],\n+ \"fleet\": [\"0.11.7\"]\n+ },\n+ \"release_notes\": \"No updates\"\n+ },\n+\n+ \"1068.6.0\": {\n+ \"release_date\": \"2016-07-12 19:54:15 +0000\",\n+ \"major_software\": {\n+ \"kernel\": [\"4.6.3\"],\n+ \"docker\": [\"1.10.3\"],\n+ \"etcd\": [\"0.4.9\",\"2.3.2\"],\n+ \"fleet\": [\"0.11.7\"]\n+ },\n+ \"release_notes\": \"No updates\"\n+ },\n+\n+ \"1010.6.0\": {\n+ \"release_date\": \"2016-06-28 23:19:30 +0000\",\n+ \"major_software\": {\n+ \"kernel\": [\"4.5.7\"],\n+ \"docker\": [\"1.10.3\"],\n+ \"etcd\": [\"0.4.9\",\"2.3.1\"],\n+ \"fleet\": [\"0.11.7\"]\n+ },\n+ \"release_notes\": \"No updates\"\n+ },\n+\n+ \"1010.5.0\": {\n+ \"release_date\": \"2016-05-27 00:05:30 +0000\",\n+ \"major_software\": {\n+ \"kernel\": [\"4.5.0\"],\n+ \"docker\": [\"1.10.3\"],\n+ \"etcd\": [\"0.4.9\",\"2.3.1\"],\n+ \"fleet\": [\"0.11.7\"]\n+ },\n+ \"release_notes\": \"No updates\"\n+ },\n+\n+ \"899.17.0\": {\n+ \"release_date\": \"2016-05-03 23:30:01 +0000\",\n+ \"major_software\": {\n+ \"kernel\": [\"4.3.6\"],\n+ \"docker\": [\"1.9.1\"],\n+ \"etcd\": [\"0.4.9\",\"2.2.3\"],\n+ \"fleet\": [\"0.11.7\"]\n+ },\n+ \"release_notes\": \"No updates\"\n+ },\n+\n+ \"899.15.0\": {\n+ \"release_date\": \"2016-04-05 17:12:31 +0000\",\n+ \"major_software\": {\n+ \"kernel\": [\"4.3.6\"],\n+ \"docker\": [\"1.9.1\"],\n+ \"etcd\": [\"0.4.9\",\"2.2.3\"],\n+ \"fleet\": [\"0.11.7\"]\n+ },\n+ \"release_notes\": \"No updates\"\n+ },\n+\n+ \"899.13.0\": {\n+ \"release_date\": \"2016-03-23 02:45:55 +0000\",\n+ \"major_software\": {\n+ \"kernel\": [\"4.3.6\"],\n+ \"docker\": [\"1.9.1\"],\n+ \"etcd\": [\"0.4.9\",\"2.2.3\"],\n+ \"fleet\": [\"0.11.5\"]\n+ },\n+ \"release_notes\": \"No updates\"\n+ },\n+\n+ \"835.13.0\": {\n+ \"release_date\": \"2016-02-18 07:07:19 +0000\",\n+ \"major_software\": {\n+ \"kernel\": [\"4.2.2\"],\n+ \"docker\": [\"1.8.3\"],\n+ \"etcd\": [\"0.4.9\",\"2.2.0\"],\n+ \"fleet\": [\"0.11.5\"]\n+ },\n+ \"release_notes\": \"No updates\"\n+ },\n+\n+ \"835.12.0\": {\n+ \"release_date\": \"2016-02-01 21:57:35 +0000\",\n+ \"major_software\": {\n+ \"kernel\": [\"4.2.2\"],\n+ \"docker\": [\"1.8.3\"],\n+ \"etcd\": [\"0.4.9\",\"2.2.0\"],\n+ \"fleet\": [\"0.11.5\"]\n+ },\n+ \"release_notes\": \"No updates\"\n+ },\n+\n+ \"835.11.0\": {\n+ \"release_date\": \"2016-01-22 20:25:46 +0000\",\n+ \"major_software\": {\n+ \"kernel\": [\"4.2.2\"],\n+ \"docker\": [\"1.8.3\"],\n+ \"etcd\": [\"0.4.9\",\"2.2.0\"],\n+ \"fleet\": [\"0.11.5\"]\n+ },\n+ \"release_notes\": \"No updates\"\n+ },\n+\n+ \"835.10.0\": {\n+ \"release_date\": \"2016-01-20 17:46:55 +0000\",\n+ \"major_software\": {\n+ \"kernel\": [\"4.2.2\"],\n+ \"docker\": [\"1.8.3\"],\n+ \"etcd\": [\"0.4.9\",\"2.2.0\"],\n+ \"fleet\": [\"0.11.5\"]\n+ },\n+ \"release_notes\": \"No updates\"\n+ },\n+\n+ \"835.9.0\": {\n+ \"release_date\": \"2015-12-08 16:32:50 +0000\",\n+ \"major_software\": {\n+ \"kernel\": [\"4.2.2\"],\n+ \"docker\": [\"1.8.3\"],\n+ \"etcd\": [\"0.4.9\",\"2.2.0\"],\n+ \"fleet\": [\"0.11.5\"]\n+ },\n+ \"release_notes\": \"No updates\"\n+ },\n+\n+ \"835.8.0\": {\n+ \"release_date\": \"2015-12-01 23:03:20 +0000\",\n+ \"major_software\": {\n+ \"kernel\": [\"4.2.2\"],\n+ \"docker\": [\"1.8.3\"],\n+ \"etcd\": [\"0.4.9\",\"2.2.0\"],\n+ \"fleet\": [\"0.11.5\"]\n+ },\n+ \"release_notes\": \"No updates\"\n+ },\n+\n+ \"766.5.0\": {\n+ \"release_date\": \"2015-11-05 20:53:12 +0000\",\n+ \"major_software\": {\n+ \"kernel\": [\"4.1.7\"],\n+ \"docker\": [\"1.7.1\"],\n+ \"etcd\": [\"0.4.9\",\"2.1.2\"],\n+ \"fleet\": [\"0.10.2\"]\n+ },\n+ \"release_notes\": \"No updates\"\n+ },\n+\n+ \"766.4.0\": {\n+ \"release_date\": \"2015-09-16 20:23:53 +0000\",\n+ \"major_software\": {\n+ \"kernel\": [\"4.1.7\"],\n+ \"docker\": [\"1.7.1\"],\n+ \"etcd\": [\"0.4.9\",\"2.1.2\"],\n+ \"fleet\": [\"0.10.2\"]\n+ },\n+ \"release_notes\": \"No updates\"\n+ },\n+\n+ \"766.3.0\": {\n+ \"release_date\": \"2015-09-02 18:14:20 +0000\",\n+ \"major_software\": {\n+ \"kernel\": [\"4.1.6\"],\n+ \"docker\": [\"1.7.1\"],\n+ \"etcd\": [\"0.4.9\",\"2.1.2\"],\n+ \"fleet\": [\"0.10.2\"]\n+ },\n+ \"release_notes\": \"No updates\"\n+ },\n+\n+ \"717.3.0\": {\n+ \"release_date\": \"2015-07-10 00:50:39 +0000\",\n+ \"major_software\": {\n+ \"kernel\": [\"4.0.5\"],\n+ \"docker\": [\"1.6.2\"],\n+ \"etcd\": [\"0.4.9\",\"2.0.10\"],\n+ \"fleet\": [\"0.10.2\"]\n+ },\n+ \"release_notes\": \"No updates\"\n+ },\n+\n+ \"723.3.0\": {\n+ \"release_date\": \"2015-07-09 20:30:21 +0000\",\n+ \"major_software\": {\n+ \"kernel\": [\"4.0.5\"],\n+ \"docker\": [\"1.6.2\"],\n+ \"etcd\": [\"0.4.9\",\"2.0.12\"],\n+ \"fleet\": [\"0.10.2\"]\n+ },\n+ \"release_notes\": \"No updates\"\n+ },\n+\n+ \"717.1.0\": {\n+ \"release_date\": \"2015-06-24 17:11:24 +0000\",\n+ \"major_software\": {\n+ \"kernel\": [\"4.0.5\"],\n+ \"docker\": [\"1.6.2\"],\n+ \"etcd\": [\"0.4.9\",\"2.0.10\"],\n+ \"fleet\": [\"0.10.2\"]\n+ },\n+ \"release_notes\": \"No updates\"\n+ },\n+\n+ \"681.2.0\": {\n+ \"release_date\": \"2015-06-18 15:19:35 +0000\",\n+ \"major_software\": {\n+ \"kernel\": [\"4.0.5\"],\n+ \"docker\": [\"1.6.2\"],\n+ \"etcd\": [\"0.4.9\",\"2.0.10\"],\n+ \"fleet\": [\"0.10.2\"]\n+ },\n+ \"release_notes\": \"No updates\"\n+ },\n+\n+ \"681.1.0\": {\n+ \"release_date\": \"2015-06-17 18:33:42 +0000\",\n+ \"major_software\": {\n+ \"kernel\": [\"4.0.5\"],\n+ \"docker\": [\"1.6.2\"],\n+ \"etcd\": [\"0.4.9\",\"2.0.10\"],\n+ \"fleet\": [\"0.10.1\"]\n+ },\n+ \"release_notes\": \"No updates\"\n+ },\n+\n+ \"647.2.0\": {\n+ \"release_date\": \"2015-05-26 22:32:06 +0000\",\n+ \"major_software\": {\n+ \"kernel\": [\"4.0.1\"],\n+ \"docker\": [\"1.5.0\"],\n+ \"etcd\": [\"0.4.9\"],\n+ \"fleet\": [\"0.9.2\"]\n+ },\n+ \"release_notes\": \"No updates\"\n+ },\n+\n+ \"681.0.0\": {\n+ \"release_date\": \"2015-05-14 16:49:27 +0000\",\n+ \"major_software\": {\n+ \"kernel\": [\"4.0.3\"],\n+ \"docker\": [\"1.6.2\"],\n+ \"etcd\": [\"0.4.9\",\"2.0.10\"],\n+ \"fleet\": [\"0.10.1\"]\n+ },\n+ \"release_notes\": \"No updates\"\n+ },\n+\n+ \"647.0.0\": {\n+ \"release_date\": \"2015-04-09 16:34:27 +0000\",\n+ \"major_software\": {\n+ \"kernel\": [\"3.19.3\"],\n+ \"docker\": [\"1.5.0\"],\n+ \"etcd\": [\"0.4.9\"],\n+ \"fleet\": [\"0.9.2\"]\n+ },\n+ \"release_notes\": \"No updates\"\n+ },\n+\n+ \"633.1.0\": {\n+ \"release_date\": \"2015-03-26 17:00:41 +0000\",\n+ \"major_software\": {\n+ \"kernel\": [\"3.19\"],\n+ \"docker\": [\"1.5.0\"],\n+ \"etcd\": [\"0.4.8\"],\n+ \"fleet\": [\"0.9.1\"]\n+ },\n+ \"release_notes\": \"No updates\"\n+ },\n+\n+ \"607.0.0\": {\n+ \"release_date\": \"2015-02-28 19:49:51 +0000\",\n+ \"major_software\": {\n+ \"kernel\": [\"3.18.6\"],\n+ \"docker\": [\"1.5.0\"],\n+ \"etcd\": [\"0.4.7\"],\n+ \"fleet\": [\"0.9.1\"]\n+ },\n+ \"release_notes\": \"No updates\"\n+ },\n+\n+ \"557.2.0\": {\n+ \"release_date\": \"2015-02-04 18:19:14 +0000\",\n+ \"major_software\": {\n+ \"kernel\": [\"3.18.1\"],\n+ \"docker\": [\"1.4.1\"],\n+ \"etcd\": [\"0.4.6\"],\n+ \"fleet\": [\"0.9.0\"]\n+ },\n+ \"release_notes\": \"No updates\"\n+ },\n+\n+ \"522.6.0\": {\n+ \"release_date\": \"2015-01-28 17:48:19 +0000\",\n+ \"major_software\": {\n+ \"kernel\": [\"3.17.8\"],\n+ \"docker\": [\"1.3.3\"],\n+ \"etcd\": [\"0.4.6\"],\n+ \"fleet\": [\"0.8.3\"]\n+ },\n+ \"release_notes\": \"No updates\"\n+ },\n+\n+ \"522.5.0\": {\n+ \"release_date\": \"2015-01-12 21:58:31 +0000\",\n+ \"major_software\": {\n+ \"kernel\": [\"3.17.8\"],\n+ \"docker\": [\"1.3.3\"],\n+ \"etcd\": [\"0.4.6\"],\n+ \"fleet\": [\"0.8.3\"]\n+ },\n+ \"release_notes\": \"No updates\"\n+ },\n+\n+ \"522.4.0\": {\n+ \"release_date\": \"2015-01-06 21:27:42 +0000\",\n+ \"major_software\": {\n+ \"kernel\": [\"3.17.7\"],\n+ \"docker\": [\"1.3.3\"],\n+ \"etcd\": [\"0.4.6\"],\n+ \"fleet\": [\"0.8.3\"]\n+ },\n+ \"release_notes\": \"No updates\"\n+ }\n+ }`\n+)\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
adds missing release files
|
596,240 |
25.04.2019 10:54:33
| -7,200 |
7bcb640a80dcd63c0fd4d7ebdeb10de0c920a80e
|
considers successfull nodes without timeout
|
[
{
"change_type": "MODIFY",
"old_path": "pkg/controller/servicing/lister.go",
"new_path": "pkg/controller/servicing/lister.go",
"diff": "@@ -254,7 +254,6 @@ func (d *NodeLister) Successful() []*core_v1.Node {\nvar found []*core_v1.Node\n// Node must have updating annotation\n- // is not timed out\n// Node must not be in the list of nodes to be rebooted\n// Node must not be in the list of nodes to be replaced\n// Node must be ready\n@@ -262,17 +261,6 @@ func (d *NodeLister) Successful() []*core_v1.Node {\nfor _, node := range d.Updating() {\nfailure := false\n- for _, r := range d.updateTimeout() {\n- if r == node {\n- failure = true\n- break\n- }\n- }\n-\n- if failure {\n- continue\n- }\n-\nfor _, r := range d.Reboot() {\nif r == node {\nfailure = true\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
considers successfull nodes without timeout
|
596,240 |
25.04.2019 11:17:28
| -7,200 |
f018ba24a45c1dcab706dd263590060685235c87
|
fixes test for successful detection
|
[
{
"change_type": "MODIFY",
"old_path": "pkg/controller/servicing/lister_test.go",
"new_path": "pkg/controller/servicing/lister_test.go",
"diff": "@@ -305,7 +305,7 @@ func TestServicingListerUpdateSuccessful(t *testing.T) {\n},\n)\nlister := NewFakeNodeLister(t, TestLogger(), kluster, nodes)\n- assert.Len(t, lister.Successful(), 1)\n+ assert.Len(t, lister.Successful(), 2)\n}\nfunc TestServicingListerUpdateFailed(t *testing.T) {\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
fixes test for successful detection
|
596,240 |
27.04.2019 13:00:21
| -7,200 |
8c01e91ecf0e6decbed319d0aa41d57e5341dcc3
|
emergency disable servicing
|
[
{
"change_type": "MODIFY",
"old_path": "pkg/cmd/kubernikus/operator.go",
"new_path": "pkg/cmd/kubernikus/operator.go",
"diff": "@@ -51,7 +51,7 @@ func NewOperatorOptions() *Options {\noptions.KubernikusDomain = \"kluster.staging.cloud.sap\"\noptions.Namespace = \"kubernikus\"\noptions.MetricPort = 9091\n- options.Controllers = []string{\"groundctl\", \"launchctl\", \"deorbiter\", \"routegc\", \"flight\", \"migration\", \"hammertime\", \"servicing\"}\n+ options.Controllers = []string{\"groundctl\", \"launchctl\", \"deorbiter\", \"routegc\", \"flight\", \"migration\", \"hammertime\"}\nreturn options\n}\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
emergency disable servicing
|
596,240 |
02.05.2019 10:26:35
| -7,200 |
f5a1a7b6e95603e343ca3422bba45e0ecaaede0b
|
reenable auto-updates
|
[
{
"change_type": "MODIFY",
"old_path": "pkg/cmd/kubernikus/operator.go",
"new_path": "pkg/cmd/kubernikus/operator.go",
"diff": "@@ -51,7 +51,7 @@ func NewOperatorOptions() *Options {\noptions.KubernikusDomain = \"kluster.staging.cloud.sap\"\noptions.Namespace = \"kubernikus\"\noptions.MetricPort = 9091\n- options.Controllers = []string{\"groundctl\", \"launchctl\", \"deorbiter\", \"routegc\", \"flight\", \"migration\", \"hammertime\"}\n+ options.Controllers = []string{\"groundctl\", \"launchctl\", \"deorbiter\", \"routegc\", \"flight\", \"migration\", \"hammertime\", \"servicing\"}\nreturn options\n}\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
reenable auto-updates
|
596,226 |
03.05.2019 16:32:14
| -7,200 |
e9dfebd4811191eb1d5041acca3d883d58007862
|
Fix running node state
|
[
{
"change_type": "MODIFY",
"old_path": "pkg/client/openstack/kluster/node.go",
"new_path": "pkg/client/openstack/kluster/node.go",
"diff": "@@ -29,7 +29,7 @@ func (n *Node) Starting() bool {\n}\nfunc (n *Node) Stopping() bool {\n- if n.TaskState == \"spawning\" || n.TaskState == \"scheduling\" || n.TaskState == \"networking\" || n.TaskState == \"block_device_mapping\" {\n+ if n.TaskState == \"spawning\" || n.TaskState == \"scheduling\" || n.TaskState == \"networking\" || n.TaskState == \"block_device_mapping\" || n.TaskState == \"image_uploading\" {\nreturn false\n}\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
Fix running node state (#425)
|
596,229 |
14.05.2019 11:01:55
| -7,200 |
4a5c2ac99ca6d2d024f59bea3d44c735ba9b245c
|
CNI and calico folder creation in ExecStartPre
|
[
{
"change_type": "MODIFY",
"old_path": "pkg/templates/node_1.10.go",
"new_path": "pkg/templates/node_1.10.go",
"diff": "@@ -100,6 +100,11 @@ systemd:\nEnvironment=\"KUBELET_IMAGE_TAG={{ .HyperkubeImageTag }}\"\nEnvironment=\"KUBELET_IMAGE_URL=docker://{{ .HyperkubeImage }}\"\nEnvironment=\"KUBELET_IMAGE_ARGS=--name=kubelet --exec=/kubelet\"\n+{{- if .CalicoNetworking }}\n+ ExecStartPre=/bin/mkdir -p /etc/cni\n+ ExecStartPre=/bin/mkdir -p /opt/cni\n+ ExecStartPre=/bin/mkdir -p /var/lib/calico\n+{{- end }}\nExecStartPre=/bin/mkdir -p /etc/kubernetes/manifests\nExecStartPre=/bin/mkdir -p /var/lib/cni\nExecStartPre=-/usr/bin/rkt rm --uuid-file=/var/run/kubelet-pod.uuid\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
CNI and calico folder creation in ExecStartPre (#415)
|
596,240 |
14.05.2019 14:39:15
| -7,200 |
c44010074eba86db39f9ee5fc985ca06deb6f122
|
adds kluster specific dashboard
|
[
{
"change_type": "MODIFY",
"old_path": "charts/kubernikus-system/values.yaml",
"new_path": "charts/kubernikus-system/values.yaml",
"diff": "@@ -197,6 +197,638 @@ grafana:\nenabled = false\nserverDashboardFiles:\n+ klusters.json: |\n+ {\n+ \"__inputs\": [\n+ {\n+ \"name\": \"default\",\n+ \"label\": \"Prometheus\",\n+ \"description\": \"\",\n+ \"type\": \"datasource\",\n+ \"pluginId\": \"prometheus\",\n+ \"pluginName\": \"Prometheus\"\n+ }\n+ ],\n+ \"__requires\": [\n+ {\n+ \"type\": \"grafana\",\n+ \"id\": \"grafana\",\n+ \"name\": \"Grafana\",\n+ \"version\": \"5.0.4\"\n+ },\n+ {\n+ \"type\": \"panel\",\n+ \"id\": \"graph\",\n+ \"name\": \"Graph\",\n+ \"version\": \"5.0.0\"\n+ },\n+ {\n+ \"type\": \"datasource\",\n+ \"id\": \"prometheus\",\n+ \"name\": \"Prometheus\",\n+ \"version\": \"5.0.0\"\n+ }\n+ ],\n+ \"annotations\": {\n+ \"list\": [\n+ {\n+ \"builtIn\": 1,\n+ \"datasource\": \"-- Grafana --\",\n+ \"enable\": true,\n+ \"hide\": true,\n+ \"iconColor\": \"rgba(0, 211, 255, 1)\",\n+ \"name\": \"Annotations & Alerts\",\n+ \"type\": \"dashboard\"\n+ }\n+ ]\n+ },\n+ \"editable\": true,\n+ \"gnetId\": null,\n+ \"graphTooltip\": 0,\n+ \"id\": null,\n+ \"iteration\": 1557837204455,\n+ \"links\": [],\n+ \"panels\": [\n+ {\n+ \"aliasColors\": {},\n+ \"bars\": false,\n+ \"dashLength\": 10,\n+ \"dashes\": false,\n+ \"datasource\": \"default\",\n+ \"decimals\": 1,\n+ \"fill\": 1,\n+ \"gridPos\": {\n+ \"h\": 8,\n+ \"w\": 12,\n+ \"x\": 0,\n+ \"y\": 0\n+ },\n+ \"id\": 2,\n+ \"legend\": {\n+ \"alignAsTable\": true,\n+ \"avg\": false,\n+ \"current\": false,\n+ \"max\": true,\n+ \"min\": false,\n+ \"rightSide\": true,\n+ \"show\": true,\n+ \"sort\": \"max\",\n+ \"sortDesc\": true,\n+ \"total\": false,\n+ \"values\": true\n+ },\n+ \"lines\": true,\n+ \"linewidth\": 1,\n+ \"links\": [],\n+ \"nullPointMode\": \"connected\",\n+ \"percentage\": false,\n+ \"pointradius\": 5,\n+ \"points\": false,\n+ \"renderer\": \"flot\",\n+ \"seriesOverrides\": [],\n+ \"spaceLength\": 10,\n+ \"stack\": false,\n+ \"steppedLine\": false,\n+ \"targets\": [\n+ {\n+ \"expr\": \"max(container_memory_usage_bytes{pod_name=~\\\"$kluster-.*\\\",container_name!=\\\"POD\\\"}) by (container_name)\",\n+ \"format\": \"time_series\",\n+ \"intervalFactor\": 1,\n+ \"legendFormat\": \"{{container_name}}\",\n+ \"refId\": \"A\"\n+ }\n+ ],\n+ \"thresholds\": [],\n+ \"timeFrom\": null,\n+ \"timeShift\": null,\n+ \"title\": \"Container Memory Usage\",\n+ \"tooltip\": {\n+ \"shared\": true,\n+ \"sort\": 2,\n+ \"value_type\": \"individual\"\n+ },\n+ \"type\": \"graph\",\n+ \"xaxis\": {\n+ \"buckets\": null,\n+ \"mode\": \"time\",\n+ \"name\": null,\n+ \"show\": true,\n+ \"values\": []\n+ },\n+ \"yaxes\": [\n+ {\n+ \"decimals\": null,\n+ \"format\": \"decbytes\",\n+ \"label\": null,\n+ \"logBase\": 1,\n+ \"max\": null,\n+ \"min\": null,\n+ \"show\": true\n+ },\n+ {\n+ \"format\": \"short\",\n+ \"label\": null,\n+ \"logBase\": 1,\n+ \"max\": null,\n+ \"min\": null,\n+ \"show\": true\n+ }\n+ ]\n+ },\n+ {\n+ \"aliasColors\": {},\n+ \"bars\": false,\n+ \"dashLength\": 10,\n+ \"dashes\": false,\n+ \"datasource\": \"default\",\n+ \"fill\": 1,\n+ \"gridPos\": {\n+ \"h\": 8,\n+ \"w\": 12,\n+ \"x\": 12,\n+ \"y\": 0\n+ },\n+ \"id\": 7,\n+ \"legend\": {\n+ \"alignAsTable\": true,\n+ \"avg\": false,\n+ \"current\": false,\n+ \"max\": true,\n+ \"min\": false,\n+ \"rightSide\": true,\n+ \"show\": true,\n+ \"sort\": \"max\",\n+ \"sortDesc\": true,\n+ \"total\": false,\n+ \"values\": true\n+ },\n+ \"lines\": true,\n+ \"linewidth\": 1,\n+ \"links\": [],\n+ \"nullPointMode\": \"connected\",\n+ \"percentage\": false,\n+ \"pointradius\": 5,\n+ \"points\": false,\n+ \"renderer\": \"flot\",\n+ \"seriesOverrides\": [],\n+ \"spaceLength\": 10,\n+ \"stack\": false,\n+ \"steppedLine\": false,\n+ \"targets\": [\n+ {\n+ \"expr\": \"max(container_memory_usage_bytes{pod_name=~\\\"$kluster-.*\\\",container_name!=\\\"POD\\\"}) by (container_name) / max(label_replace(kube_pod_container_resource_limits_memory_bytes{pod=~\\\"$kluster-.*\\\",container!=\\\"POD\\\"}, \\\"container_name\\\", \\\"$1\\\", \\\"container\\\", \\\"(.*)\\\")) by (container_name)\",\n+ \"format\": \"time_series\",\n+ \"interval\": \"\",\n+ \"intervalFactor\": 1,\n+ \"legendFormat\": \"{{container_name}}\",\n+ \"refId\": \"A\"\n+ }\n+ ],\n+ \"thresholds\": [],\n+ \"timeFrom\": null,\n+ \"timeShift\": null,\n+ \"title\": \"Container Memory Utilization\",\n+ \"tooltip\": {\n+ \"shared\": true,\n+ \"sort\": 2,\n+ \"value_type\": \"individual\"\n+ },\n+ \"type\": \"graph\",\n+ \"xaxis\": {\n+ \"buckets\": null,\n+ \"mode\": \"time\",\n+ \"name\": null,\n+ \"show\": true,\n+ \"values\": []\n+ },\n+ \"yaxes\": [\n+ {\n+ \"decimals\": null,\n+ \"format\": \"percentunit\",\n+ \"label\": null,\n+ \"logBase\": 1,\n+ \"max\": \"1\",\n+ \"min\": \"0\",\n+ \"show\": true\n+ },\n+ {\n+ \"format\": \"short\",\n+ \"label\": null,\n+ \"logBase\": 1,\n+ \"max\": null,\n+ \"min\": null,\n+ \"show\": true\n+ }\n+ ]\n+ },\n+ {\n+ \"aliasColors\": {},\n+ \"bars\": false,\n+ \"dashLength\": 10,\n+ \"dashes\": false,\n+ \"datasource\": \"default\",\n+ \"fill\": 1,\n+ \"gridPos\": {\n+ \"h\": 8,\n+ \"w\": 12,\n+ \"x\": 0,\n+ \"y\": 8\n+ },\n+ \"id\": 3,\n+ \"legend\": {\n+ \"alignAsTable\": true,\n+ \"avg\": false,\n+ \"current\": false,\n+ \"max\": true,\n+ \"min\": false,\n+ \"rightSide\": true,\n+ \"show\": true,\n+ \"sort\": \"max\",\n+ \"sortDesc\": true,\n+ \"total\": false,\n+ \"values\": true\n+ },\n+ \"lines\": true,\n+ \"linewidth\": 1,\n+ \"links\": [],\n+ \"nullPointMode\": \"connected\",\n+ \"percentage\": false,\n+ \"pointradius\": 5,\n+ \"points\": false,\n+ \"renderer\": \"flot\",\n+ \"seriesOverrides\": [],\n+ \"spaceLength\": 10,\n+ \"stack\": false,\n+ \"steppedLine\": false,\n+ \"targets\": [\n+ {\n+ \"expr\": \"sum(irate(container_cpu_usage_seconds_total{pod_name=~\\\"$kluster-.*\\\",container_name!=\\\"POD\\\"}[5m])) by (container_name)\",\n+ \"format\": \"time_series\",\n+ \"intervalFactor\": 1,\n+ \"legendFormat\": \"{{container_name}}\",\n+ \"refId\": \"A\"\n+ }\n+ ],\n+ \"thresholds\": [],\n+ \"timeFrom\": null,\n+ \"timeShift\": null,\n+ \"title\": \"Container CPU Usage\",\n+ \"tooltip\": {\n+ \"shared\": true,\n+ \"sort\": 2,\n+ \"value_type\": \"individual\"\n+ },\n+ \"type\": \"graph\",\n+ \"xaxis\": {\n+ \"buckets\": null,\n+ \"mode\": \"time\",\n+ \"name\": null,\n+ \"show\": true,\n+ \"values\": []\n+ },\n+ \"yaxes\": [\n+ {\n+ \"decimals\": null,\n+ \"format\": \"none\",\n+ \"label\": null,\n+ \"logBase\": 1,\n+ \"max\": null,\n+ \"min\": null,\n+ \"show\": true\n+ },\n+ {\n+ \"format\": \"short\",\n+ \"label\": null,\n+ \"logBase\": 1,\n+ \"max\": null,\n+ \"min\": null,\n+ \"show\": true\n+ }\n+ ]\n+ },\n+ {\n+ \"aliasColors\": {},\n+ \"bars\": false,\n+ \"dashLength\": 10,\n+ \"dashes\": false,\n+ \"datasource\": \"default\",\n+ \"fill\": 1,\n+ \"gridPos\": {\n+ \"h\": 8,\n+ \"w\": 12,\n+ \"x\": 12,\n+ \"y\": 8\n+ },\n+ \"id\": 8,\n+ \"legend\": {\n+ \"alignAsTable\": true,\n+ \"avg\": false,\n+ \"current\": false,\n+ \"max\": true,\n+ \"min\": false,\n+ \"rightSide\": true,\n+ \"show\": true,\n+ \"sort\": \"max\",\n+ \"sortDesc\": true,\n+ \"total\": false,\n+ \"values\": true\n+ },\n+ \"lines\": true,\n+ \"linewidth\": 1,\n+ \"links\": [],\n+ \"nullPointMode\": \"connected\",\n+ \"percentage\": false,\n+ \"pointradius\": 5,\n+ \"points\": false,\n+ \"renderer\": \"flot\",\n+ \"seriesOverrides\": [],\n+ \"spaceLength\": 10,\n+ \"stack\": false,\n+ \"steppedLine\": false,\n+ \"targets\": [\n+ {\n+ \"expr\": \"sum(irate(container_cpu_usage_seconds_total{pod_name=~\\\"$kluster-.*\\\",container_name!=\\\"POD\\\"}[5m])) by (container_name) / max(label_replace(kube_pod_container_resource_limits_cpu_cores{pod=~\\\"$kluster-.*\\\",container!=\\\"POD\\\"}, \\\"container_name\\\", \\\"$1\\\", \\\"container\\\", \\\"(.*)\\\")) by (container_name)\",\n+ \"format\": \"time_series\",\n+ \"intervalFactor\": 1,\n+ \"legendFormat\": \"{{container_name}}\",\n+ \"refId\": \"B\"\n+ }\n+ ],\n+ \"thresholds\": [],\n+ \"timeFrom\": null,\n+ \"timeShift\": null,\n+ \"title\": \"Container CPU Utilization\",\n+ \"tooltip\": {\n+ \"shared\": true,\n+ \"sort\": 2,\n+ \"value_type\": \"individual\"\n+ },\n+ \"type\": \"graph\",\n+ \"xaxis\": {\n+ \"buckets\": null,\n+ \"mode\": \"time\",\n+ \"name\": null,\n+ \"show\": true,\n+ \"values\": []\n+ },\n+ \"yaxes\": [\n+ {\n+ \"decimals\": null,\n+ \"format\": \"percentunit\",\n+ \"label\": null,\n+ \"logBase\": 1,\n+ \"max\": \"1\",\n+ \"min\": null,\n+ \"show\": true\n+ },\n+ {\n+ \"format\": \"short\",\n+ \"label\": null,\n+ \"logBase\": 1,\n+ \"max\": null,\n+ \"min\": null,\n+ \"show\": true\n+ }\n+ ]\n+ },\n+ {\n+ \"aliasColors\": {},\n+ \"bars\": false,\n+ \"dashLength\": 10,\n+ \"dashes\": false,\n+ \"datasource\": \"default\",\n+ \"decimals\": 0,\n+ \"fill\": 1,\n+ \"gridPos\": {\n+ \"h\": 8,\n+ \"w\": 12,\n+ \"x\": 0,\n+ \"y\": 16\n+ },\n+ \"id\": 6,\n+ \"legend\": {\n+ \"alignAsTable\": true,\n+ \"avg\": false,\n+ \"current\": false,\n+ \"max\": true,\n+ \"min\": false,\n+ \"rightSide\": true,\n+ \"show\": true,\n+ \"sort\": \"max\",\n+ \"sortDesc\": true,\n+ \"total\": false,\n+ \"values\": true\n+ },\n+ \"lines\": true,\n+ \"linewidth\": 1,\n+ \"links\": [],\n+ \"nullPointMode\": \"connected\",\n+ \"percentage\": false,\n+ \"pointradius\": 5,\n+ \"points\": false,\n+ \"renderer\": \"flot\",\n+ \"seriesOverrides\": [],\n+ \"spaceLength\": 10,\n+ \"stack\": false,\n+ \"steppedLine\": true,\n+ \"targets\": [\n+ {\n+ \"expr\": \"max(label_replace(kube_pod_container_status_restarts_total{pod=~\\\"$kluster.*\\\"}, \\\"app\\\", \\\"$2\\\", \\\"pod\\\", \\\"(.*)-(apiserver|etcd|scheduler|cmanager|backup)-[^-]+\\\")) by (app)\",\n+ \"format\": \"time_series\",\n+ \"interval\": \"\",\n+ \"intervalFactor\": 1,\n+ \"legendFormat\": \"{{app}}\",\n+ \"refId\": \"A\"\n+ }\n+ ],\n+ \"thresholds\": [],\n+ \"timeFrom\": null,\n+ \"timeShift\": null,\n+ \"title\": \"Pod Restarts\",\n+ \"tooltip\": {\n+ \"shared\": true,\n+ \"sort\": 2,\n+ \"value_type\": \"individual\"\n+ },\n+ \"type\": \"graph\",\n+ \"xaxis\": {\n+ \"buckets\": null,\n+ \"mode\": \"time\",\n+ \"name\": null,\n+ \"show\": true,\n+ \"values\": []\n+ },\n+ \"yaxes\": [\n+ {\n+ \"decimals\": null,\n+ \"format\": \"none\",\n+ \"label\": null,\n+ \"logBase\": 1,\n+ \"max\": null,\n+ \"min\": \"0\",\n+ \"show\": true\n+ },\n+ {\n+ \"format\": \"short\",\n+ \"label\": null,\n+ \"logBase\": 1,\n+ \"max\": null,\n+ \"min\": null,\n+ \"show\": true\n+ }\n+ ]\n+ },\n+ {\n+ \"aliasColors\": {},\n+ \"bars\": false,\n+ \"dashLength\": 10,\n+ \"dashes\": false,\n+ \"datasource\": \"default\",\n+ \"fill\": 1,\n+ \"gridPos\": {\n+ \"h\": 8,\n+ \"w\": 12,\n+ \"x\": 12,\n+ \"y\": 16\n+ },\n+ \"id\": 4,\n+ \"legend\": {\n+ \"alignAsTable\": true,\n+ \"avg\": false,\n+ \"current\": false,\n+ \"hideEmpty\": true,\n+ \"hideZero\": true,\n+ \"max\": true,\n+ \"min\": false,\n+ \"rightSide\": true,\n+ \"show\": true,\n+ \"sort\": \"max\",\n+ \"sortDesc\": true,\n+ \"total\": false,\n+ \"values\": true\n+ },\n+ \"lines\": true,\n+ \"linewidth\": 1,\n+ \"links\": [],\n+ \"nullPointMode\": \"connected\",\n+ \"percentage\": false,\n+ \"pointradius\": 5,\n+ \"points\": false,\n+ \"renderer\": \"flot\",\n+ \"seriesOverrides\": [],\n+ \"spaceLength\": 10,\n+ \"stack\": false,\n+ \"steppedLine\": false,\n+ \"targets\": [\n+ {\n+ \"expr\": \"sum(irate(container_cpu_cfs_throttled_seconds_total{pod_name=~\\\"$kluster-.*\\\",container_name!=\\\"POD\\\"}[5m])) by (container_name) \",\n+ \"format\": \"time_series\",\n+ \"interval\": \"\",\n+ \"intervalFactor\": 1,\n+ \"legendFormat\": \"{{container_name}}\",\n+ \"refId\": \"A\"\n+ }\n+ ],\n+ \"thresholds\": [],\n+ \"timeFrom\": null,\n+ \"timeShift\": null,\n+ \"title\": \"Container CPU Throttled\",\n+ \"tooltip\": {\n+ \"shared\": true,\n+ \"sort\": 2,\n+ \"value_type\": \"individual\"\n+ },\n+ \"type\": \"graph\",\n+ \"xaxis\": {\n+ \"buckets\": null,\n+ \"mode\": \"time\",\n+ \"name\": null,\n+ \"show\": true,\n+ \"values\": []\n+ },\n+ \"yaxes\": [\n+ {\n+ \"decimals\": null,\n+ \"format\": \"none\",\n+ \"label\": null,\n+ \"logBase\": 1,\n+ \"max\": null,\n+ \"min\": null,\n+ \"show\": true\n+ },\n+ {\n+ \"format\": \"short\",\n+ \"label\": null,\n+ \"logBase\": 1,\n+ \"max\": null,\n+ \"min\": null,\n+ \"show\": true\n+ }\n+ ]\n+ }\n+ ],\n+ \"refresh\": \"1m\",\n+ \"schemaVersion\": 16,\n+ \"style\": \"dark\",\n+ \"tags\": [],\n+ \"templating\": {\n+ \"list\": [\n+ {\n+ \"allValue\": null,\n+ \"current\": {},\n+ \"datasource\": \"default\",\n+ \"hide\": 0,\n+ \"includeAll\": false,\n+ \"label\": \"Kluster\",\n+ \"multi\": false,\n+ \"name\": \"kluster\",\n+ \"options\": [],\n+ \"query\": \"label_values(kubernikus_kluster_info{phase=\\\"Running\\\",kluster_name!~\\\"kubernikus-test-.*|e2e-.*\\\"},kluster_name)\",\n+ \"refresh\": 1,\n+ \"regex\": \"\",\n+ \"sort\": 1,\n+ \"tagValuesQuery\": \"\",\n+ \"tags\": [],\n+ \"tagsQuery\": \"\",\n+ \"type\": \"query\",\n+ \"useTags\": false\n+ }\n+ ]\n+ },\n+ \"time\": {\n+ \"from\": \"now-2d\",\n+ \"to\": \"now\"\n+ },\n+ \"timepicker\": {\n+ \"refresh_intervals\": [\n+ \"5s\",\n+ \"10s\",\n+ \"30s\",\n+ \"1m\",\n+ \"5m\",\n+ \"15m\",\n+ \"30m\",\n+ \"1h\",\n+ \"2h\",\n+ \"1d\"\n+ ],\n+ \"time_options\": [\n+ \"5m\",\n+ \"15m\",\n+ \"1h\",\n+ \"6h\",\n+ \"12h\",\n+ \"24h\",\n+ \"2d\",\n+ \"7d\",\n+ \"30d\"\n+ ]\n+ },\n+ \"timezone\": \"\",\n+ \"title\": \"Klusters\",\n+ \"uid\": \"kKjSTjmWk\",\n+ \"version\": 12\n+ }\nkubernikus.json: |\n{\n\"__inputs\": [\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
adds kluster specific dashboard
|
596,226 |
06.06.2019 11:33:31
| -7,200 |
1f601fdf3cbd34f4d7e501d32c5a9499541d3340
|
Add support for token auth to kubernikusctl
|
[
{
"change_type": "MODIFY",
"old_path": "pkg/cmd/kubernikusctl/auth/init.go",
"new_path": "pkg/cmd/kubernikusctl/auth/init.go",
"diff": "@@ -79,7 +79,7 @@ func (o *InitOptions) Complete(args []string) (err error) {\nfunc (o *InitOptions) Run(c *cobra.Command) (err error) {\nstorePasswordInKeyRing := false\n- if o.openstack.Password == \"\" && o.openstack.ApplicationCredentialSecret == \"\" {\n+ if o.openstack.Password == \"\" && o.openstack.ApplicationCredentialSecret == \"\" && o.openstack.TokenID == \"\" {\nfmt.Printf(\"Password: \")\nif password, err := gopass.GetPasswdMasked(); err != nil {\nreturn err\n"
},
{
"change_type": "MODIFY",
"old_path": "pkg/cmd/kubernikusctl/auth/refresh.go",
"new_path": "pkg/cmd/kubernikusctl/auth/refresh.go",
"diff": "@@ -49,6 +49,7 @@ func NewRefreshCommand() *cobra.Command {\nfunc (o *RefreshOptions) BindFlags(flags *pflag.FlagSet) {\ncommon.BindLogFlags(flags)\nflags.StringVar(&o.openstack.Password, \"password\", \"\", \"User password [OS_PASSWORD]\")\n+ flags.StringVar(&o.openstack.TokenID, \"token\", \"\", \"Token to authenticate with [OS_TOKEN]\")\nflags.StringVar(&o.kubeconfigPath, \"kubeconfig\", o.kubeconfigPath, \"Overwrites kubeconfig auto-detection with explicit path\")\nflags.StringVar(&o.context, \"context\", o.context, \"Overwrites current-context in kubeconfig\")\nflags.BoolVar(&o.force, \"force\", o.force, \"Force refresh\")\n@@ -114,6 +115,8 @@ func (o *RefreshOptions) Run(c *cobra.Command) error {\nreturn errors.Wrap(err, \"Couldn't parse Kubernikus URL. Rerun init.\")\n}\n+ storePasswordInKeyRing := false\n+ if o.openstack.TokenID == \"\" {\nif o.openstack.Username, err = ktx.Username(); err != nil {\nreturn errors.Wrap(err, \"Failed to extract username from certificate\")\n}\n@@ -126,7 +129,6 @@ func (o *RefreshOptions) Run(c *cobra.Command) error {\nglog.V(2).Infof(\"Detected domain-name: %v\", o.openstack.DomainName)\no.openstack.DomainID = \"\" //Ignore conflicting value from environment\n- storePasswordInKeyRing := false\nif o.openstack.Password == \"\" {\nfmt.Printf(\"Password: \")\nif password, err := gopass.GetPasswdMasked(); err != nil {\n@@ -136,6 +138,7 @@ func (o *RefreshOptions) Run(c *cobra.Command) error {\nstorePasswordInKeyRing = true\n}\n}\n+ }\nif err := o.setupClients(); err != nil {\n"
},
{
"change_type": "MODIFY",
"old_path": "pkg/cmd/kubernikusctl/common/openstack.go",
"new_path": "pkg/cmd/kubernikusctl/common/openstack.go",
"diff": "@@ -48,9 +48,18 @@ func (o *OpenstackClient) BindFlags(flags *pflag.FlagSet) {\nflags.StringVar(&o.ApplicationCredentialName, \"application-credential-name\", o.ApplicationCredentialName, \"Project application credential name [OS_APPLICATION_CREDENTIAL_NAME]\")\nflags.StringVar(&o.ApplicationCredentialID, \"application-credential-id\", o.ApplicationCredentialName, \"Project application credential id [OS_APPLICATION_CREDENTIAL_ID]\")\nflags.StringVar(&o.ApplicationCredentialSecret, \"application-credential-secret\", \"\", \"Project application credential secret [OS_APPLICATION_CREDENTIAL_SECRET]\")\n+ flags.StringVar(&o.TokenID, \"token\", \"\", \"Token to authenticate with [OS_TOKEN]\")\n+\n}\nfunc (o *OpenstackClient) Validate(c *cobra.Command, args []string) error {\n+ if o.TokenID == \"\" {\n+ o.TokenID = os.Getenv(\"OS_TOKEN\")\n+ }\n+ if o.TokenID != \"\" {\n+ return nil\n+ }\n+\nif o.IdentityEndpoint == \"\" {\nreturn errors.Errorf(\"You need to provide --auth-url or OS_AUTH_URL\")\n} else {\n@@ -139,7 +148,11 @@ func (o *OpenstackClient) Complete(args []string) error {\nfunc (o *OpenstackClient) Setup() error {\nvar err error\n- if o.Password == \"\" && o.ApplicationCredentialSecret == \"\" {\n+ if o.TokenID == \"\" && os.Getenv(\"OS_TOKEN\") != \"\" {\n+ o.TokenID = os.Getenv(\"OS_TOKEN\")\n+ }\n+\n+ if o.Password == \"\" && o.ApplicationCredentialSecret == \"\" && o.TokenID == \"\" {\nif os.Getenv(\"OS_PASSWORD\") != \"\" {\no.Password = os.Getenv(\"OS_PASSWORD\")\n} else {\n@@ -215,6 +228,7 @@ func (o *OpenstackClient) PrintDebugAuthInfo() string {\nDomainName: {{ .DomainName }}\nApplicationCredentialID: {{ .ApplicationCredentialID }}\nApplicationCredentialName: {{ .ApplicationCredentialName }}\n+ Token: {{ .TokenID }}\nScope:\nProjectID: {{ .Scope.ProjectID }}\nProjectName: {{ .Scope.ProjectName }}\n@@ -232,6 +246,11 @@ func (o *OpenstackClient) PrintDebugAuthInfo() string {\n}\nfunc (o *OpenstackClient) Authenticate() error {\n+ if o.TokenID != \"\" {\n+ o.Provider.TokenID = o.TokenID\n+ return nil\n+ }\n+\nreturn openstack.AuthenticateV3(o.Provider, o, gophercloud.EndpointOpts{})\n}\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
Add support for token auth to kubernikusctl
|
596,240 |
06.06.2019 15:21:55
| -7,200 |
5edd8afa1ffb0e507aa1b6cb4832e7c1bd39833f
|
stubs coreos releases and versions
|
[
{
"change_type": "MODIFY",
"old_path": "pkg/controller/servicing/controller_test.go",
"new_path": "pkg/controller/servicing/controller_test.go",
"diff": "@@ -303,8 +303,8 @@ func TestServicingControllerReconcile(t *testing.T) {\nlisters := &NodeListerFactory{\nLogger: logger,\nNodeObservatory: nodeobservatory.NewFakeController(kluster, nodes...),\n- CoreOSVersion: &coreos.Version{},\n- CoreOSRelease: coreos.NewFakeRelease(t, \"2079.3.0\"),\n+ CoreOSVersion: coreos.NewFakeVersion(t, \"2023.4.0\"),\n+ CoreOSRelease: coreos.NewFakeRelease(t, \"2023.4.0\"),\n}\nreconcilers := &KlusterReconcilerFactory{\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
stubs coreos releases and versions
|
596,240 |
06.06.2019 16:14:19
| -7,200 |
a78569c46e770f42ad4224ae92eafdfb27d4b7d1
|
use force to drain nodes
|
[
{
"change_type": "MODIFY",
"old_path": "pkg/controller/servicing/lifecycler.go",
"new_path": "pkg/controller/servicing/lifecycler.go",
"diff": "@@ -144,7 +144,7 @@ func (lc *NodeLifeCycler) Drain(node *core_v1.Node) error {\n}\noptions := &drain.DrainOptions{\n- Force: false,\n+ Force: true,\nIgnoreDaemonsets: true,\nGracePeriodSeconds: -1,\nTimeout: EvictionTimeout,\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
use force to drain nodes
|
596,240 |
07.06.2019 17:06:40
| -7,200 |
1c55acfb983a3616ff0d1e47bbbed26da7dda6d2
|
fixes update.conf to reboot off again
|
[
{
"change_type": "ADD",
"old_path": null,
"new_path": "pkg/migration/16_fix_update_conf.go",
"diff": "+package migration\n+\n+import (\n+ v1 \"github.com/sapcc/kubernikus/pkg/apis/kubernikus/v1\"\n+ \"github.com/sapcc/kubernikus/pkg/controller/config\"\n+)\n+\n+const (\n+ fixUpdateConf = `#!/bin/bash\n+cat <<EOF > /etc/coreos/update.conf\n+REBOOT_STRATEGY=\"off\"\n+EOF\n+\n+/usr/bin/pkill update_engine\n+/usr/bin/pkill locksmithd\n+`\n+)\n+\n+func FixUpdateConf(rawKluster []byte, current *v1.Kluster, clients config.Clients, factories config.Factories) (err error) {\n+ client, err := clients.Satellites.ClientFor(current)\n+ if err != nil {\n+ return err\n+ }\n+\n+ return ApplySuppository(fixUpdateConf, client)\n+}\n"
},
{
"change_type": "MODIFY",
"old_path": "pkg/migration/register.go",
"new_path": "pkg/migration/register.go",
"diff": "@@ -22,6 +22,7 @@ func init() {\nFixRootCertificate,\nCleanupSuppositoryNamespaces,\nReconcileNodePoolConfigDefaults,\n+ FixUpdateConf,\n// <-- Insert new migrations at the end only!\n}\n}\n"
},
{
"change_type": "MODIFY",
"old_path": "pkg/templates/node_1.10.go",
"new_path": "pkg/templates/node_1.10.go",
"diff": "@@ -11,8 +11,6 @@ passwd:\nssh_authorized_keys:\n- {{ .LoginPublicKey | quote }}\n{{- end }}\n-locksmith:\n- reboot_strategy: \"reboot\"\nsystemd:\nunits:\n"
},
{
"change_type": "MODIFY",
"old_path": "pkg/templates/node_1.11.go",
"new_path": "pkg/templates/node_1.11.go",
"diff": "@@ -11,8 +11,6 @@ passwd:\nssh_authorized_keys:\n- {{ .LoginPublicKey | quote }}\n{{- end }}\n-locksmith:\n- reboot_strategy: \"reboot\"\nsystemd:\nunits:\n"
},
{
"change_type": "MODIFY",
"old_path": "pkg/templates/node_1.7.go",
"new_path": "pkg/templates/node_1.7.go",
"diff": "@@ -11,8 +11,6 @@ passwd:\nssh_authorized_keys:\n- {{ .LoginPublicKey | quote }}\n{{- end }}\n-locksmith:\n- reboot_strategy: \"reboot\"\nsystemd:\nunits:\n"
},
{
"change_type": "MODIFY",
"old_path": "pkg/templates/node_1.8.go",
"new_path": "pkg/templates/node_1.8.go",
"diff": "@@ -11,8 +11,6 @@ passwd:\nssh_authorized_keys:\n- {{ .LoginPublicKey | quote }}\n{{- end }}\n-locksmith:\n- reboot_strategy: \"reboot\"\nsystemd:\nunits:\n"
},
{
"change_type": "MODIFY",
"old_path": "pkg/templates/node_1.9.go",
"new_path": "pkg/templates/node_1.9.go",
"diff": "@@ -11,8 +11,6 @@ passwd:\nssh_authorized_keys:\n- {{ .LoginPublicKey | quote }}\n{{- end }}\n-locksmith:\n- reboot_strategy: \"reboot\"\nsystemd:\nunits:\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
fixes update.conf to reboot off again
|
596,245 |
17.06.2019 10:51:40
| -7,200 |
f4df4e57ff517315662b3944709b645d5a1da470
|
bump cni plugins in hyperkube
|
[
{
"change_type": "MODIFY",
"old_path": "contrib/hyperkube/Makefile",
"new_path": "contrib/hyperkube/Makefile",
"diff": "IMAGE?=sapcc/hyperkube\nARCH=amd64\n-CNI_RELEASE=v0.6.0\n-CNI_PLUGINS_RELEASE=v0.7.1\n+CNI_PLUGINS_RELEASE=v0.8.1\nCNI_DEDUP_BRIDGE_RELEASE=v0.1.0\n-MULTUS_RELEASE=v3.1\n+MULTUS_RELEASE=3.2\nBASEIMAGE?=k8s.gcr.io/hyperkube-$(ARCH):$(VERSION)\nTEMP_DIR:=$(shell mktemp -d -t hyperkubeXXXXXX)\n@@ -23,9 +22,8 @@ push: build\ncni:\nmkdir -p ${TEMP_DIR}/cni-bin/bin\n- curl -sSL --retry 5 https://github.com/containernetworking/cni/releases/download/${CNI_RELEASE}/cni-${ARCH}-${CNI_RELEASE}.tgz | tar -xz -C ${TEMP_DIR}/cni-bin/bin\n- curl -sSL --retry 5 https://github.com/containernetworking/plugins/releases/download/${CNI_PLUGINS_RELEASE}/cni-plugins-${ARCH}-${CNI_PLUGINS_RELEASE}.tgz | tar -xz -C ${TEMP_DIR}/cni-bin/bin\n- curl -sSL --retry 5 https://github.com/intel/multus-cni/releases/download/${MULTUS_RELEASE}/multus-cni_${MULTUS_RELEASE}_linux_amd64.tar.gz | tar --strip 1 -xz -C ${TEMP_DIR}/cni-bin/bin\n+ curl -sSL --retry 5 https://github.com/containernetworking/plugins/releases/download/${CNI_PLUGINS_RELEASE}/cni-plugins-linux-${ARCH}-${CNI_PLUGINS_RELEASE}.tgz | tar -xz -C ${TEMP_DIR}/cni-bin/bin\n+ curl -sSL --retry 5 https://github.com/intel/multus-cni/releases/download/v${MULTUS_RELEASE}/multus-cni_${MULTUS_RELEASE}_linux_amd64.tar.gz | tar --strip 1 -xz -C ${TEMP_DIR}/cni-bin/bin\ncurl -o ${TEMP_DIR}/cni-bin/bin/dedup-bridge -sSL --retry 5 https://github.com/sapcc/cni-dedup-bridge/releases/download/${CNI_DEDUP_BRIDGE_RELEASE}/dedup-bridge\nchmod +x ${TEMP_DIR}/cni-bin/bin/dedup-bridge\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
bump cni plugins in hyperkube (#432)
|
596,240 |
27.06.2019 15:45:50
| -7,200 |
96b3375c6dd78251b49a5ae4f00f376dd109f0fc
|
listing avz needs the member roles
|
[
{
"change_type": "MODIFY",
"old_path": "terraform/kubernikus.tf",
"new_path": "terraform/kubernikus.tf",
"diff": "@@ -229,6 +229,12 @@ resource \"openstack_identity_role_assignment_v3\" \"terraform_kubernetes_admin\" {\nrole_id = \"${openstack_identity_role_v3.kubernetes_admin.id}\"\n}\n+resource \"openstack_identity_role_assignment_v3\" \"terraform_member\" {\n+ user_id = \"${data.openstack_identity_user_v3.kubernikus_terraform.id}\"\n+ project_id = \"${openstack_identity_project_v3.kubernikus.id}\"\n+ role_id = \"${openstack_identity_role_v3.member.id}\"\n+}\n+\nresource \"openstack_identity_role_assignment_v3\" \"pipeline_kubernetes_admin\" {\nuser_id = \"${openstack_identity_user_v3.kubernikus_pipeline.id}\"\nproject_id = \"${openstack_identity_project_v3.kubernikus.id}\"\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
listing avz needs the member roles
|
596,240 |
27.06.2019 15:45:59
| -7,200 |
24ffe48e725999ffba464125ce13c3abe5a5c1c8
|
adding k-na-us-2
|
[
{
"change_type": "MODIFY",
"old_path": "ci/pipeline.yaml",
"new_path": "ci/pipeline.yaml",
"diff": "@@ -443,6 +443,40 @@ auth_e2e_na-us-1: &auth_e2e_na-us-1\n+auth_na-us-2: &auth_na-us-2\n+\n+\n+ OS_AUTH_URL: https://identity-3.na-us-2.cloud.sap/v3\n+ OS_USERNAME: kubernikus-pipeline\n+ OS_PASSWORD: ((kubernikus-pipeline-password))\n+ OS_USER_DOMAIN_NAME: Default\n+ OS_PROJECT_NAME: kubernikus\n+ OS_PROJECT_DOMAIN_NAME: ccadmin\n+ KUBERNIKUS_NAME: k-na-us-2\n+ KUBERNIKUS_URL: https://k-na-us-2.admin.cloud.sap\n+\n+\n+\n+auth_e2e_na-us-2: &auth_e2e_na-us-2\n+ CP_KLUSTER: k-na-us-2\n+ CP_KUBERNIKUS_URL: https://k-na-us-2.admin.cloud.sap\n+ CP_OS_PROJECT_DOMAIN_NAME: ccadmin\n+ KUBERNIKUS_NAME: e2e\n+ OS_PROJECT_DOMAIN_NAME: ccadmin\n+\n+\n+ OS_AUTH_URL: https://identity-3.na-us-2.cloud.sap/v3\n+ OS_USERNAME: kubernikus-pipeline\n+ OS_PASSWORD: ((kubernikus-pipeline-password))\n+ OS_USER_DOMAIN_NAME: Default\n+ OS_PROJECT_NAME: kubernikus_e2e\n+ OS_REGION_NAME: na-us-2\n+ KUBERNIKUS_URL: https://kubernikus.na-us-2.cloud.sap\n+ CP_OS_PROJECT_NAME: kubernikus\n+\n+\n+\n+\nauth_na-us-3: &auth_na-us-3\n@@ -2850,6 +2884,14 @@ jobs:\nKUBERNIKUS_NAME: k-na-us-1\nGITHUB_TOKEN: ((github-access-token))\n+ - task: k-na-us-2\n+ config:\n+ <<: *task_helm-admin_kubernikus\n+ params:\n+ REGION: admin\n+ KUBERNIKUS_NAME: k-na-us-2\n+ GITHUB_TOKEN: ((github-access-token))\n+\n- task: k-na-us-3\nconfig:\n<<: *task_helm-admin_kubernikus\n@@ -3281,6 +3323,148 @@ jobs:\n+ - name: seed_na-us-2\n+ serial: true\n+ plan:\n+ - aggregate:\n+ - get: AMERICAS.tag\n+ passed: [admin_AMERICAS]\n+ trigger: true\n+ - get: secrets.git\n+ passed: [admin_AMERICAS]\n+ - get: helm-charts.git\n+ passed: [AMERICAS]\n+ - task: seed\n+ config:\n+ <<: *task_helm_seed\n+ input_mapping:\n+ charts.git: helm-charts.git\n+ params:\n+ REGION: na-us-2\n+ GITHUB_TOKEN: ((github-access-token))\n+ <<: *slack\n+\n+\n+ - name: terraform_na-us-2\n+ serial: true\n+ plan:\n+ - aggregate:\n+ - get: AMERICAS.tag\n+ trigger: true\n+ passed: [seed_na-us-2]\n+ - get: secrets.git\n+ passed: [seed_na-us-2]\n+ - get: terraform.git\n+ passed: [AMERICAS]\n+ - task: terraform\n+ config:\n+ <<: *task_terraform\n+ timeout: 45m\n+ attempts: 5\n+ params:\n+ TF_REGION: na-us-2\n+ TF_PASSWORD: ((kubernikus-terraform-password))\n+ <<: *slack\n+\n+\n+ - name: tiller_na-us-2\n+ plan:\n+ - aggregate:\n+ - get: AMERICAS.tag\n+ trigger: true\n+ passed: [terraform_na-us-2]\n+ - get: tiller-version.git\n+ trigger: true\n+ - aggregate:\n+ - task: k-na-us-2\n+ config:\n+ <<: *task_tiller\n+ timeout: 10m\n+ params:\n+ <<: *auth_na-us-2\n+ <<: *slack\n+\n+ - name: prometheus-operator_na-us-2\n+ plan:\n+ - aggregate:\n+ - get: helm-charts-prometheus-operator.git\n+ trigger: true\n+ - get: secrets.git\n+ - aggregate:\n+ - task: prometheus-operator_na-us-2\n+ input_mapping: {helm-charts.git: helm-charts-prometheus-operator.git}\n+ config:\n+ <<: *task_helm_prometheus-operator\n+ params:\n+ <<: *auth_na-us-2\n+\n+ - name: kubernikus-monitoring_na-us-2\n+ serial: true\n+ plan:\n+ - aggregate:\n+ - get: secrets.git\n+ passed: [prometheus-operator_na-us-2]\n+ - get: kubernikus.builds\n+ trigger: true\n+ resource: master.builds\n+ - aggregate:\n+ - task: kubernikus-monitoring_na-us-2\n+ config:\n+ <<: *task_helm_kubernikus-monitoring\n+ params:\n+ <<: *auth_na-us-2\n+\n+ - name: deploy_na-us-2\n+ serial: true\n+ plan:\n+ - aggregate:\n+ - get: AMERICAS.tag\n+ trigger: true\n+ passed: [tiller_na-us-2]\n+ - get: secrets.git\n+ passed: [terraform_na-us-2]\n+ - get: kubernikus.builds\n+ resource: master.builds\n+ passed: [admin_AMERICAS]\n+ - aggregate:\n+ - task: kubernikus_na-us-2\n+ config:\n+ <<: *task_helm_kubernikus\n+ params:\n+ <<: *auth_na-us-2\n+ - task: kubernikus-system_na-us-2\n+ config:\n+ <<: *task_helm_kubernikus-system\n+ params:\n+ <<: *auth_na-us-2\n+ <<: *slack\n+\n+ - name: soak_na-us-2\n+ serial: true\n+ build_logs_to_retain: 168\n+ plan:\n+ - aggregate:\n+ - get: AMERICAS.tag\n+ trigger: true\n+ passed: [deploy_na-us-2]\n+ - get: kubernikus.builds\n+ resource: master.builds\n+ passed: [deploy_na-us-2]\n+ - get: hourly\n+ trigger: true\n+ - task: e2e_tests\n+ config:\n+ <<: *task_e2e_tests\n+ timeout: 45m\n+ params:\n+ <<: *auth_e2e_na-us-2\n+ SENTRY_DSN: ((sentry-soak-tests))\n+ <<: *slack\n+\n+\n+\n+\n+\n- name: seed_na-us-3\nserial: true\nplan:\n@@ -3769,6 +3953,16 @@ groups:\n- deploy_na-us-1\n- soak_na-us-1\n+ - name: na-us-2\n+ jobs:\n+ - AMERICAS\n+ - admin_AMERICAS\n+ - seed_na-us-2\n+ - terraform_na-us-2\n+ - tiller_na-us-2\n+ - deploy_na-us-2\n+ - soak_na-us-2\n+\n- name: na-us-3\njobs:\n- AMERICAS\n@@ -3889,6 +4083,12 @@ groups:\n- deploy_na-us-1\n- soak_na-us-1\n+ - seed_na-us-2\n+ - terraform_na-us-2\n+ - tiller_na-us-2\n+ - deploy_na-us-2\n+ - soak_na-us-2\n+\n- seed_na-us-3\n- terraform_na-us-3\n- tiller_na-us-3\n@@ -3944,6 +4144,9 @@ groups:\n- kubernikus-monitoring_na-us-1\n- prometheus-operator_na-us-1\n+ - kubernikus-monitoring_na-us-2\n+ - prometheus-operator_na-us-2\n+\n- kubernikus-monitoring_na-us-3\n- prometheus-operator_na-us-3\n"
},
{
"change_type": "MODIFY",
"old_path": "ci/pipeline.yaml.erb",
"new_path": "ci/pipeline.yaml.erb",
"diff": "@@ -13,6 +13,7 @@ REGIONS = {\n\"la-br-1\": { continent: 'AMERICAS', install: 'terraform', e2e: 'terraform' },\n\"na-ca-1\": { continent: 'AMERICAS', install: 'terraform', e2e: 'terraform' },\n\"na-us-1\": { continent: 'AMERICAS', install: 'manual' , e2e: 'manual' },\n+ \"na-us-2\": { continent: 'AMERICAS', install: 'terraform', e2e: 'terraform' },\n\"na-us-3\": { continent: 'AMERICAS', install: 'terraform', e2e: 'terraform' },\n\"qa-de-1\": { continent: 'INFRA', install: 'terraform', e2e: 'terraform', nodepool_avz: 'qa-de-1a' },\n\"master\": { continent: \"INFRA\", install: 'virtual' , e2e: 'virtual' }\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
adding k-na-us-2
|
596,240 |
27.06.2019 15:52:45
| -7,200 |
4643eb1ee844a22ace431d2f79c6aaa400d5eb4c
|
fix missing data provider
|
[
{
"change_type": "MODIFY",
"old_path": "terraform/kubernikus.tf",
"new_path": "terraform/kubernikus.tf",
"diff": "@@ -232,7 +232,7 @@ resource \"openstack_identity_role_assignment_v3\" \"terraform_kubernetes_admin\" {\nresource \"openstack_identity_role_assignment_v3\" \"terraform_member\" {\nuser_id = \"${data.openstack_identity_user_v3.kubernikus_terraform.id}\"\nproject_id = \"${openstack_identity_project_v3.kubernikus.id}\"\n- role_id = \"${openstack_identity_role_v3.member.id}\"\n+ role_id = \"${data.openstack_identity_role_v3.member.id}\"\n}\nresource \"openstack_identity_role_assignment_v3\" \"pipeline_kubernetes_admin\" {\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
fix missing data provider
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.