author
int64 658
755k
| date
stringlengths 19
19
| timezone
int64 -46,800
43.2k
| hash
stringlengths 40
40
| message
stringlengths 5
490
| mods
list | language
stringclasses 20
values | license
stringclasses 3
values | repo
stringlengths 5
68
| original_message
stringlengths 12
491
|
---|---|---|---|---|---|---|---|---|---|
596,240 |
27.11.2017 16:35:39
| -3,600 |
7464dae111b19d9bc709f8838d261a9b885b71c9
|
adds instrumentation to poolmanagers
|
[
{
"change_type": "MODIFY",
"old_path": "pkg/controller/launch/controller.go",
"new_path": "pkg/controller/launch/controller.go",
"diff": "@@ -138,6 +138,12 @@ func (lr *LaunchReconciler) newNodePoolManager(kluster *v1.Kluster, pool *models\npm = &ConcretePoolManager{lr.Clients, kluster, pool, logger}\npm = &EventingPoolManager{pm, kluster, lr.Recorder}\npm = &LoggingPoolManager{pm, logger}\n+ pm = &InstrumentingPoolManager{pm,\n+ metrics.LaunchOperationsLatency,\n+ metrics.LaunchOperationsTotal,\n+ metrics.LaunchSuccessfulOperationsTotal,\n+ metrics.LaunchFailedOperationsTotal,\n+ }\nreturn pm\n}\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
adds instrumentation to poolmanagers
|
596,240 |
27.11.2017 16:43:41
| -3,600 |
59039075265c856351c48fcd3e93fb871245612f
|
detangles reconciler and pool manager
|
[
{
"change_type": "MODIFY",
"old_path": "pkg/controller/launch/controller.go",
"new_path": "pkg/controller/launch/controller.go",
"diff": "@@ -3,24 +3,14 @@ package launch\nimport (\n\"github.com/sapcc/kubernikus/pkg/api/models\"\n\"github.com/sapcc/kubernikus/pkg/apis/kubernikus/v1\"\n- \"github.com/sapcc/kubernikus/pkg/client/openstack\"\n\"github.com/sapcc/kubernikus/pkg/controller/base\"\n\"github.com/sapcc/kubernikus/pkg/controller/config\"\n\"github.com/sapcc/kubernikus/pkg/controller/metrics\"\n- \"github.com/sapcc/kubernikus/pkg/templates\"\n\"github.com/go-kit/kit/log\"\n- metav1 \"k8s.io/apimachinery/pkg/apis/meta/v1\"\n\"k8s.io/client-go/tools/record\"\n)\n-type PoolManager interface {\n- GetStatus() (*PoolStatus, error)\n- SetStatus(*PoolStatus) error\n- CreateNode() (string, error)\n- DeleteNode(string) error\n-}\n-\ntype LaunchReconciler struct {\nconfig.Clients\n@@ -28,23 +18,6 @@ type LaunchReconciler struct {\nLogger log.Logger\n}\n-type PoolStatus struct {\n- Nodes []string\n- Running int\n- Starting int\n- Stopping int\n- Needed int\n- UnNeeded int\n-}\n-\n-type ConcretePoolManager struct {\n- config.Clients\n-\n- Kluster *v1.Kluster\n- Pool *models.NodePool\n- Logger log.Logger\n-}\n-\nfunc NewController(factories config.Factories, clients config.Clients, recorder record.EventRecorder, logger log.Logger) base.Controller {\nlogger = log.With(logger,\n\"controller\", \"launch\")\n@@ -85,7 +58,7 @@ func (lr *LaunchReconciler) Reconcile(kluster *v1.Kluster) (requeueRequested boo\nfunc (lr *LaunchReconciler) reconcilePool(kluster *v1.Kluster, pool *models.NodePool) (status *PoolStatus, requeue bool, err error) {\n- pm := lr.newNodePoolManager(kluster, pool)\n+ pm := lr.newPoolManager(kluster, pool)\nstatus, err = pm.GetStatus()\nif err != nil {\nreturn\n@@ -127,149 +100,3 @@ func (lr *LaunchReconciler) reconcilePool(kluster *v1.Kluster, pool *models.Node\nerr = pm.SetStatus(status)\nreturn\n}\n-\n-func (lr *LaunchReconciler) newNodePoolManager(kluster *v1.Kluster, pool *models.NodePool) PoolManager {\n- logger := log.With(lr.Logger,\n- \"kluster\", kluster.Spec.Name,\n- \"project\", kluster.Account(),\n- \"pool\", pool.Name)\n-\n- var pm PoolManager\n- pm = &ConcretePoolManager{lr.Clients, kluster, pool, logger}\n- pm = &EventingPoolManager{pm, kluster, lr.Recorder}\n- pm = &LoggingPoolManager{pm, logger}\n- pm = &InstrumentingPoolManager{pm,\n- metrics.LaunchOperationsLatency,\n- metrics.LaunchOperationsTotal,\n- metrics.LaunchSuccessfulOperationsTotal,\n- metrics.LaunchFailedOperationsTotal,\n- }\n-\n- return pm\n-}\n-\n-func (cpm *ConcretePoolManager) GetStatus() (status *PoolStatus, err error) {\n- status = &PoolStatus{}\n- nodes, err := cpm.Clients.Openstack.GetNodes(cpm.Kluster, cpm.Pool)\n- if err != nil {\n- return status, err\n- }\n-\n- return &PoolStatus{\n- Nodes: cpm.nodeIDs(nodes),\n- Running: cpm.running(nodes),\n- Starting: cpm.starting(nodes),\n- Stopping: cpm.stopping(nodes),\n- Needed: cpm.needed(nodes),\n- UnNeeded: cpm.unNeeded(nodes),\n- }, nil\n-}\n-\n-func (cpm *ConcretePoolManager) SetStatus(status *PoolStatus) error {\n- newInfo := models.NodePoolInfo{\n- Name: cpm.Pool.Name,\n- Size: cpm.Pool.Size,\n- Running: int64(status.Running + status.Starting),\n- Healthy: int64(status.Running),\n- Schedulable: int64(status.Running),\n- }\n-\n- copy, err := cpm.Clients.Kubernikus.Kubernikus().Klusters(cpm.Kluster.Namespace).Get(cpm.Kluster.Name, metav1.GetOptions{})\n- if err != nil {\n- return err\n- }\n-\n- for i, curInfo := range copy.Status.NodePools {\n- if curInfo.Name == newInfo.Name {\n- if curInfo == newInfo {\n- return nil\n- }\n-\n- copy.Status.NodePools[i] = newInfo\n- _, err = cpm.Clients.Kubernikus.Kubernikus().Klusters(copy.Namespace).Update(copy)\n- return err\n- }\n- }\n-\n- return nil\n-}\n-\n-func (cpm *ConcretePoolManager) CreateNode() (id string, err error) {\n- secret, err := cpm.Clients.Kubernetes.CoreV1().Secrets(cpm.Kluster.Namespace).Get(cpm.Kluster.GetName(), metav1.GetOptions{})\n- if err != nil {\n- return \"\", err\n- }\n-\n- userdata, err := templates.Ignition.GenerateNode(cpm.Kluster, secret)\n- if err != nil {\n- return \"\", err\n- }\n-\n- id, err = cpm.Clients.Openstack.CreateNode(cpm.Kluster, cpm.Pool, userdata)\n- if err != nil {\n- return \"\", err\n- }\n-\n- return id, nil\n-}\n-\n-func (cpm *ConcretePoolManager) DeleteNode(id string) (err error) {\n- if err = cpm.Clients.Openstack.DeleteNode(cpm.Kluster, id); err != nil {\n- return err\n- }\n- return nil\n-}\n-\n-func (cpm *ConcretePoolManager) nodeIDs(nodes []openstack.Node) []string {\n- result := []string{}\n- for _, n := range nodes {\n- result = append(result, n.ID)\n- }\n- return result\n-}\n-\n-func (cpm *ConcretePoolManager) starting(nodes []openstack.Node) int {\n- var count int = 0\n- for _, n := range nodes {\n- if n.Starting() {\n- count = count + 1\n- }\n- }\n- return count\n-}\n-\n-func (cpm *ConcretePoolManager) stopping(nodes []openstack.Node) int {\n- var count int = 0\n- for _, n := range nodes {\n- if n.Stopping() {\n- count = count + 1\n- }\n- }\n- return count\n-}\n-\n-func (cpm *ConcretePoolManager) running(nodes []openstack.Node) int {\n- var count int = 0\n- for _, n := range nodes {\n- if n.Running() {\n- count = count + 1\n- }\n- }\n- return count\n-}\n-\n-func (cpm *ConcretePoolManager) needed(nodes []openstack.Node) int {\n- needed := int(cpm.Pool.Size) - cpm.running(nodes) - cpm.starting(nodes)\n- if needed < 0 {\n- return 0\n- }\n- return needed\n-}\n-\n-func (cpm ConcretePoolManager) unNeeded(nodes []openstack.Node) int {\n- unneeded := cpm.running(nodes) + cpm.starting(nodes) - int(cpm.Pool.Size)\n- if unneeded < 0 {\n- return 0\n- }\n- return unneeded\n-}\n"
},
{
"change_type": "ADD",
"old_path": null,
"new_path": "pkg/controller/launch/pool_manager.go",
"diff": "+package launch\n+\n+import (\n+ \"github.com/sapcc/kubernikus/pkg/api/models\"\n+ \"github.com/sapcc/kubernikus/pkg/apis/kubernikus/v1\"\n+ \"github.com/sapcc/kubernikus/pkg/client/openstack\"\n+ \"github.com/sapcc/kubernikus/pkg/controller/config\"\n+ \"github.com/sapcc/kubernikus/pkg/controller/metrics\"\n+ \"github.com/sapcc/kubernikus/pkg/templates\"\n+\n+ \"github.com/go-kit/kit/log\"\n+ metav1 \"k8s.io/apimachinery/pkg/apis/meta/v1\"\n+)\n+\n+type PoolManager interface {\n+ GetStatus() (*PoolStatus, error)\n+ SetStatus(*PoolStatus) error\n+ CreateNode() (string, error)\n+ DeleteNode(string) error\n+}\n+\n+type PoolStatus struct {\n+ Nodes []string\n+ Running int\n+ Starting int\n+ Stopping int\n+ Needed int\n+ UnNeeded int\n+}\n+\n+type ConcretePoolManager struct {\n+ config.Clients\n+\n+ Kluster *v1.Kluster\n+ Pool *models.NodePool\n+ Logger log.Logger\n+}\n+\n+func (lr *LaunchReconciler) newPoolManager(kluster *v1.Kluster, pool *models.NodePool) PoolManager {\n+ logger := log.With(lr.Logger,\n+ \"kluster\", kluster.Spec.Name,\n+ \"project\", kluster.Account(),\n+ \"pool\", pool.Name)\n+\n+ var pm PoolManager\n+ pm = &ConcretePoolManager{lr.Clients, kluster, pool, logger}\n+ pm = &EventingPoolManager{pm, kluster, lr.Recorder}\n+ pm = &LoggingPoolManager{pm, logger}\n+ pm = &InstrumentingPoolManager{pm,\n+ metrics.LaunchOperationsLatency,\n+ metrics.LaunchOperationsTotal,\n+ metrics.LaunchSuccessfulOperationsTotal,\n+ metrics.LaunchFailedOperationsTotal,\n+ }\n+\n+ return pm\n+}\n+\n+func (cpm *ConcretePoolManager) GetStatus() (status *PoolStatus, err error) {\n+ status = &PoolStatus{}\n+ nodes, err := cpm.Clients.Openstack.GetNodes(cpm.Kluster, cpm.Pool)\n+ if err != nil {\n+ return status, err\n+ }\n+\n+ return &PoolStatus{\n+ Nodes: cpm.nodeIDs(nodes),\n+ Running: cpm.running(nodes),\n+ Starting: cpm.starting(nodes),\n+ Stopping: cpm.stopping(nodes),\n+ Needed: cpm.needed(nodes),\n+ UnNeeded: cpm.unNeeded(nodes),\n+ }, nil\n+}\n+\n+func (cpm *ConcretePoolManager) SetStatus(status *PoolStatus) error {\n+ newInfo := models.NodePoolInfo{\n+ Name: cpm.Pool.Name,\n+ Size: cpm.Pool.Size,\n+ Running: int64(status.Running + status.Starting),\n+ Healthy: int64(status.Running),\n+ Schedulable: int64(status.Running),\n+ }\n+\n+ copy, err := cpm.Clients.Kubernikus.Kubernikus().Klusters(cpm.Kluster.Namespace).Get(cpm.Kluster.Name, metav1.GetOptions{})\n+ if err != nil {\n+ return err\n+ }\n+\n+ for i, curInfo := range copy.Status.NodePools {\n+ if curInfo.Name == newInfo.Name {\n+ if curInfo == newInfo {\n+ return nil\n+ }\n+\n+ copy.Status.NodePools[i] = newInfo\n+ _, err = cpm.Clients.Kubernikus.Kubernikus().Klusters(copy.Namespace).Update(copy)\n+ return err\n+ }\n+ }\n+\n+ return nil\n+}\n+\n+func (cpm *ConcretePoolManager) CreateNode() (id string, err error) {\n+ secret, err := cpm.Clients.Kubernetes.CoreV1().Secrets(cpm.Kluster.Namespace).Get(cpm.Kluster.GetName(), metav1.GetOptions{})\n+ if err != nil {\n+ return \"\", err\n+ }\n+\n+ userdata, err := templates.Ignition.GenerateNode(cpm.Kluster, secret)\n+ if err != nil {\n+ return \"\", err\n+ }\n+\n+ id, err = cpm.Clients.Openstack.CreateNode(cpm.Kluster, cpm.Pool, userdata)\n+ if err != nil {\n+ return \"\", err\n+ }\n+\n+ return id, nil\n+}\n+\n+func (cpm *ConcretePoolManager) DeleteNode(id string) (err error) {\n+ if err = cpm.Clients.Openstack.DeleteNode(cpm.Kluster, id); err != nil {\n+ return err\n+ }\n+ return nil\n+}\n+\n+func (cpm *ConcretePoolManager) nodeIDs(nodes []openstack.Node) []string {\n+ result := []string{}\n+ for _, n := range nodes {\n+ result = append(result, n.ID)\n+ }\n+ return result\n+}\n+\n+func (cpm *ConcretePoolManager) starting(nodes []openstack.Node) int {\n+ var count int = 0\n+ for _, n := range nodes {\n+ if n.Starting() {\n+ count = count + 1\n+ }\n+ }\n+ return count\n+}\n+\n+func (cpm *ConcretePoolManager) stopping(nodes []openstack.Node) int {\n+ var count int = 0\n+ for _, n := range nodes {\n+ if n.Stopping() {\n+ count = count + 1\n+ }\n+ }\n+ return count\n+}\n+\n+func (cpm *ConcretePoolManager) running(nodes []openstack.Node) int {\n+ var count int = 0\n+ for _, n := range nodes {\n+ if n.Running() {\n+ count = count + 1\n+ }\n+ }\n+ return count\n+}\n+\n+func (cpm *ConcretePoolManager) needed(nodes []openstack.Node) int {\n+ needed := int(cpm.Pool.Size) - cpm.running(nodes) - cpm.starting(nodes)\n+ if needed < 0 {\n+ return 0\n+ }\n+ return needed\n+}\n+\n+func (cpm ConcretePoolManager) unNeeded(nodes []openstack.Node) int {\n+ unneeded := cpm.running(nodes) + cpm.starting(nodes) - int(cpm.Pool.Size)\n+ if unneeded < 0 {\n+ return 0\n+ }\n+ return unneeded\n+}\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
detangles reconciler and pool manager
|
596,240 |
27.11.2017 16:50:24
| -3,600 |
3ea6413899488bfd9441d663994577d3852ea2d3
|
break out of switch and loop to persist status for display
|
[
{
"change_type": "MODIFY",
"old_path": "pkg/controller/launch/controller.go",
"new_path": "pkg/controller/launch/controller.go",
"diff": "@@ -69,26 +69,26 @@ func (lr *LaunchReconciler) reconcilePool(kluster *v1.Kluster, pool *models.Node\nfor _, node := range status.Nodes {\nrequeue = true\nif err = pm.DeleteNode(node); err != nil {\n- return\n+ break\n}\n}\n- return\n+ break\ncase status.Needed > 0:\nfor i := 0; i < int(status.Needed); i++ {\nrequeue = true\nif _, err = pm.CreateNode(); err != nil {\n- return\n+ break\n}\n}\n- return\n+ break\ncase status.UnNeeded > 0:\nfor i := 0; i < int(status.UnNeeded); i++ {\nrequeue = true\nif err = pm.DeleteNode(status.Nodes[i]); err != nil {\n- return\n+ break\n}\n}\n- return\n+ break\ncase status.Starting > 0:\nrequeue = true\ncase status.Stopping > 0:\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
break out of switch and loop to persist status for display
|
596,240 |
27.11.2017 17:00:00
| -3,600 |
a804121a1280abfef5bc8feeadd318426472471c
|
reverts optimization
|
[
{
"change_type": "MODIFY",
"old_path": "pkg/controller/launch/controller.go",
"new_path": "pkg/controller/launch/controller.go",
"diff": "@@ -69,26 +69,26 @@ func (lr *LaunchReconciler) reconcilePool(kluster *v1.Kluster, pool *models.Node\nfor _, node := range status.Nodes {\nrequeue = true\nif err = pm.DeleteNode(node); err != nil {\n- break\n+ return\n}\n}\n- break\n+ return\ncase status.Needed > 0:\nfor i := 0; i < int(status.Needed); i++ {\nrequeue = true\nif _, err = pm.CreateNode(); err != nil {\n- break\n+ return\n}\n}\n- break\n+ return\ncase status.UnNeeded > 0:\nfor i := 0; i < int(status.UnNeeded); i++ {\nrequeue = true\nif err = pm.DeleteNode(status.Nodes[i]); err != nil {\n- break\n+ return\n}\n}\n- break\n+ return\ncase status.Starting > 0:\nrequeue = true\ncase status.Stopping > 0:\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
reverts optimization
|
596,240 |
27.11.2017 17:01:32
| -3,600 |
f4c7ac5e99721fef724ec532e3b7a71bff6e77ff
|
removes fake event
|
[
{
"change_type": "MODIFY",
"old_path": "pkg/controller/base/logging.go",
"new_path": "pkg/controller/base/logging.go",
"diff": "package base\nimport (\n- \"fmt\"\n\"time\"\n\"github.com/go-kit/kit/log\"\n@@ -15,15 +14,6 @@ type LoggingReconciler struct {\nLogger log.Logger\n}\n-type EventingReconciler struct {\n- Reconciler\n-}\n-\n-func (r *EventingReconciler) Reconcile(kluster *v1.Kluster) (requeue bool, err error) {\n- fmt.Printf(\"EVENT: Reconciled %v\\n\", kluster.Name)\n- return r.Reconciler.Reconcile(kluster)\n-}\n-\nfunc (r *LoggingReconciler) Reconcile(kluster *v1.Kluster) (requeue bool, err error) {\ndefer func(begin time.Time) {\nr.Logger.Log(\n"
},
{
"change_type": "MODIFY",
"old_path": "pkg/controller/launch/controller.go",
"new_path": "pkg/controller/launch/controller.go",
"diff": "@@ -25,7 +25,6 @@ func NewController(factories config.Factories, clients config.Clients, recorder\nvar reconciler base.Reconciler\nreconciler = &LaunchReconciler{clients, recorder, logger}\nreconciler = &base.LoggingReconciler{reconciler, logger}\n- reconciler = &base.EventingReconciler{reconciler}\nreconciler = &base.InstrumentingReconciler{\nreconciler,\nmetrics.LaunchOperationsLatency,\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
removes fake event
|
596,240 |
27.11.2017 17:02:01
| -3,600 |
4322d8214325b0428561750e9295c4d77a4c0738
|
unifies log message
|
[
{
"change_type": "MODIFY",
"old_path": "pkg/controller/launch/logging.go",
"new_path": "pkg/controller/launch/logging.go",
"diff": "@@ -30,7 +30,7 @@ func (npm *LoggingPoolManager) GetStatus() (status *PoolStatus, err error) {\nfunc (npm *LoggingPoolManager) SetStatus(status *PoolStatus) (err error) {\ndefer func(begin time.Time) {\nnpm.Logger.Log(\n- \"msg\", \"wrote node pool status\",\n+ \"msg\", \"wrote status\",\n\"running\", status.Running,\n\"starting\", status.Starting,\n\"stopping\", status.Stopping,\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
unifies log message
|
596,240 |
28.11.2017 10:41:25
| -3,600 |
3d9c90dfbf60ed56a24bd1a12643402f583a7daa
|
removes default return. this caused the status to always be out-of-date
|
[
{
"change_type": "MODIFY",
"old_path": "pkg/controller/launch/controller.go",
"new_path": "pkg/controller/launch/controller.go",
"diff": "@@ -92,8 +92,6 @@ func (lr *LaunchReconciler) reconcilePool(kluster *v1.Kluster, pool *models.Node\nrequeue = true\ncase status.Stopping > 0:\nrequeue = true\n- default:\n- return\n}\nerr = pm.SetStatus(status)\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
removes default return. this caused the status to always be out-of-date
|
596,240 |
28.11.2017 12:43:16
| -3,600 |
35270210edab92db968a376323a5a3b25b1e4cda
|
unswallows error
|
[
{
"change_type": "MODIFY",
"old_path": "pkg/client/openstack/client.go",
"new_path": "pkg/client/openstack/client.go",
"diff": "@@ -552,7 +552,7 @@ func (c *client) GetNodes(kluster *kubernikus_v1.Kluster, pool *models.NodePool)\nprefix := fmt.Sprintf(\"%v-%v-\", kluster.Spec.Name, pool_id)\nopts := servers.ListOpts{Name: prefix}\n- servers.List(client, opts).EachPage(func(page pagination.Page) (bool, error) {\n+ err = servers.List(client, opts).EachPage(func(page pagination.Page) (bool, error) {\nnodes, err = ExtractServers(page)\nif err != nil {\nglog.V(5).Infof(\"Couldn't extract server %v\", err)\n@@ -561,6 +561,9 @@ func (c *client) GetNodes(kluster *kubernikus_v1.Kluster, pool *models.NodePool)\nreturn true, nil\n})\n+ if err != nil {\n+ return nodes, err\n+ }\nreturn nodes, nil\n}\n"
},
{
"change_type": "MODIFY",
"old_path": "pkg/controller/launch/controller.go",
"new_path": "pkg/controller/launch/controller.go",
"diff": "@@ -56,7 +56,6 @@ func (lr *LaunchReconciler) Reconcile(kluster *v1.Kluster) (requeueRequested boo\n}\nfunc (lr *LaunchReconciler) reconcilePool(kluster *v1.Kluster, pool *models.NodePool) (status *PoolStatus, requeue bool, err error) {\n-\npm := lr.newPoolManager(kluster, pool)\nstatus, err = pm.GetStatus()\nif err != nil {\n"
},
{
"change_type": "MODIFY",
"old_path": "pkg/controller/launch/pool_manager.go",
"new_path": "pkg/controller/launch/pool_manager.go",
"diff": "@@ -60,7 +60,7 @@ func (cpm *ConcretePoolManager) GetStatus() (status *PoolStatus, err error) {\nstatus = &PoolStatus{}\nnodes, err := cpm.Clients.Openstack.GetNodes(cpm.Kluster, cpm.Pool)\nif err != nil {\n- return status, err\n+ return nil, err\n}\nreturn &PoolStatus{\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
unswallows error
|
596,240 |
04.12.2017 11:37:47
| -3,600 |
b9ca0b5d3c0c3fb08896e79406e28956964b295b
|
defaults to v=0 for log levels
|
[
{
"change_type": "MODIFY",
"old_path": "pkg/cmd/kubernikus/operator.go",
"new_path": "pkg/cmd/kubernikus/operator.go",
"diff": "@@ -93,7 +93,7 @@ func (o *Options) Run(c *cobra.Command) error {\nvar logger log.Logger\nlogger = log.NewLogfmtLogger(log.NewSyncWriter(os.Stderr))\nlogger = logutil.NewTrailingNilFilter(logger)\n- logger = log.With(logger, \"ts\", log.DefaultTimestampUTC, \"caller\", log.DefaultCaller)\n+ logger = log.With(logger, \"ts\", log.DefaultTimestampUTC, \"caller\", log.DefaultCaller, \"v\", 0)\nsigs := make(chan os.Signal, 1)\nstop := make(chan struct{})\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
defaults to v=0 for log levels
|
596,240 |
05.12.2017 15:26:12
| -3,600 |
689fe983a4f4769b11be79435a156eedad48ebcd
|
fixes logging panic
|
[
{
"change_type": "MODIFY",
"old_path": "pkg/controller/launch/pool_manager.go",
"new_path": "pkg/controller/launch/pool_manager.go",
"diff": "@@ -60,7 +60,7 @@ func (cpm *ConcretePoolManager) GetStatus() (status *PoolStatus, err error) {\nstatus = &PoolStatus{}\nnodes, err := cpm.Clients.Openstack.GetNodes(cpm.Kluster, cpm.Pool)\nif err != nil {\n- return nil, err\n+ return status, err\n}\nreturn &PoolStatus{\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
fixes logging panic
|
596,240 |
05.12.2017 18:17:29
| -3,600 |
9ffa1c45868f470e0579115869c76907144db47d
|
abort abort. doesn't work
|
[
{
"change_type": "MODIFY",
"old_path": "pkg/cmd/kubernikus/operator.go",
"new_path": "pkg/cmd/kubernikus/operator.go",
"diff": "@@ -93,7 +93,7 @@ func (o *Options) Run(c *cobra.Command) error {\nvar logger log.Logger\nlogger = log.NewLogfmtLogger(log.NewSyncWriter(os.Stderr))\nlogger = logutil.NewTrailingNilFilter(logger)\n- logger = log.With(logger, \"ts\", log.DefaultTimestampUTC, \"caller\", log.DefaultCaller, \"v\", 0)\n+ logger = log.With(logger, \"ts\", log.DefaultTimestampUTC, \"caller\", log.DefaultCaller)\nsigs := make(chan os.Signal, 1)\nstop := make(chan struct{})\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
abort abort. doesn't work
|
596,240 |
05.12.2017 18:19:05
| -3,600 |
3f08e02d497fe1d0714b24771dd85e7e8bcab634
|
removes glog
|
[
{
"change_type": "MODIFY",
"old_path": "pkg/cmd/kubernikus/operator.go",
"new_path": "pkg/cmd/kubernikus/operator.go",
"diff": "@@ -9,7 +9,6 @@ import (\n\"syscall\"\n\"github.com/go-kit/kit/log\"\n- \"github.com/golang/glog\"\n\"github.com/spf13/cobra\"\n\"github.com/spf13/pflag\"\n@@ -104,7 +103,7 @@ func (o *Options) Run(c *cobra.Command) error {\ngo metrics.ExposeMetrics(o.MetricPort, stop, wg)\n<-sigs // Wait for signals (this hangs until a signal arrives)\n- glog.Info(\"Shutting down...\")\n+ logger.Log(\"msg\", \"shutting down\", \"v\", 1)\nclose(stop) // Tell goroutines to stop themselves\nwg.Wait() // Wait for all to be stopped\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
removes glog
|
596,230 |
06.12.2017 13:36:36
| -3,600 |
db9b613ec6b214a9cf95b7a7c38a1b693b397c23
|
disable non-masquerade-cidr and apply masquerade via iptables-restore
|
[
{
"change_type": "MODIFY",
"old_path": "pkg/templates/node.go",
"new_path": "pkg/templates/node.go",
"diff": "@@ -15,6 +15,8 @@ locksmith:\nsystemd:\nunits:\n+ - name: iptables-restore.service\n+ enable: true\n- name: ccloud-metadata.service\ncontents: |\n[Unit]\n@@ -79,7 +81,7 @@ systemd:\n--cluster-dns={{ .ClusterDNSAddress }} \\\n--cluster-domain={{ .ClusterDomain }} \\\n--client-ca-file=/etc/kubernetes/certs/kubelet-clients-ca.pem \\\n- --non-masquerade-cidr={{ .ClusterCIDR }} \\\n+ --non-masquerade-cidr=0.0.0.0/0 \\\n--anonymous-auth=false\nExecStop=-/usr/bin/rkt stop --uuid-file=/var/run/kubelet-pod.uuid\nRestart=always\n@@ -166,6 +168,18 @@ networkd:\nstorage:\nfiles:\n+ - path: /var/lib/iptables/rules-save\n+ mode: 0644\n+ contents:\n+ inline: |-\n+ *nat\n+ :PREROUTING ACCEPT [0:0]\n+ :INPUT ACCEPT [0:0]\n+ :OUTPUT ACCEPT [0:0]\n+ :POSTROUTING ACCEPT [0:0]\n+ -A POSTROUTING -p tcp ! -d {{ .ClusterCIDR }} -m addrtype ! --dst-type LOCAL -j MASQUERADE --to-ports 32000-65000\n+ -A POSTROUTING -p udp ! -d {{ .ClusterCIDR }} -m addrtype ! --dst-type LOCAL -j MASQUERADE --to-ports 32000-65000\n+ COMMIT\n- path: /etc/sysctl.d/10-enable-icmp-redirects\nfilesystem: root\nmode: 0644\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
-- disable non-masquerade-cidr and apply masquerade via iptables-restore
|
596,240 |
06.12.2017 13:53:54
| -3,600 |
30e4be72793c48aaa039d22333c03fcc4d570a2b
|
sets log levels
|
[
{
"change_type": "MODIFY",
"old_path": "pkg/controller/base/controller.go",
"new_path": "pkg/controller/base/controller.go",
"diff": "@@ -77,6 +77,7 @@ func (c *controller) Run(threadiness int, stopCh <-chan struct{}, wg *sync.WaitG\nc.logger.Log(\n\"msg\", \"starting run loop\",\n\"threadiness\", threadiness,\n+ \"v\", 2,\n)\ndefer c.queue.ShutDown()\n@@ -107,6 +108,7 @@ func (c *controller) requeueAllKlusters() (err error) {\ndefer func() {\nc.logger.Log(\n\"msg\", \"requeued all\",\n+ \"v\", 1,\n\"err\", err,\n)\n}()\n@@ -128,6 +130,7 @@ func (c *controller) requeueKluster(kluster *v1.Kluster) {\n\"msg\", \"queuing\",\n\"kluster\", kluster.Spec.Name,\n\"project\", kluster.Account(),\n+ \"v\", 2,\n)\nkey, err := cache.MetaNamespaceKeyFunc(kluster)\n"
},
{
"change_type": "MODIFY",
"old_path": "pkg/controller/base/logging.go",
"new_path": "pkg/controller/base/logging.go",
"diff": "@@ -22,6 +22,7 @@ func (r *LoggingReconciler) Reconcile(kluster *v1.Kluster) (requeue bool, err er\n\"project\", kluster.Account(),\n\"requeue\", requeue,\n\"took\", time.Since(begin),\n+ \"v\", 1,\n\"err\", err)\n}(time.Now())\nreturn r.Reconciler.Reconcile(kluster)\n"
},
{
"change_type": "MODIFY",
"old_path": "pkg/controller/launch/logging.go",
"new_path": "pkg/controller/launch/logging.go",
"diff": "@@ -21,6 +21,7 @@ func (npm *LoggingPoolManager) GetStatus() (status *PoolStatus, err error) {\n\"needed\", status.Needed,\n\"unneeded\", status.UnNeeded,\n\"took\", time.Since(begin),\n+ \"v\", 1,\n\"err\", err,\n)\n}(time.Now())\n@@ -37,6 +38,7 @@ func (npm *LoggingPoolManager) SetStatus(status *PoolStatus) (err error) {\n\"needed\", status.Needed,\n\"unneeded\", status.UnNeeded,\n\"took\", time.Since(begin),\n+ \"v\", 1,\n\"err\", err,\n)\n}(time.Now())\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
sets log levels
|
596,240 |
06.12.2017 14:22:40
| -3,600 |
28c60ebe580ece5f54e972c0ee9dde6188a952c9
|
enriches caller with package path
|
[
{
"change_type": "MODIFY",
"old_path": "pkg/cmd/kubernikus/operator.go",
"new_path": "pkg/cmd/kubernikus/operator.go",
"diff": "@@ -3,12 +3,14 @@ package kubernikus\nimport (\n\"errors\"\ngoflag \"flag\"\n+ \"fmt\"\n\"os\"\n\"os/signal\"\n\"sync\"\n\"syscall\"\n\"github.com/go-kit/kit/log\"\n+ \"github.com/go-stack/stack\"\n\"github.com/spf13/cobra\"\n\"github.com/spf13/pflag\"\n@@ -92,7 +94,7 @@ func (o *Options) Run(c *cobra.Command) error {\nvar logger log.Logger\nlogger = log.NewLogfmtLogger(log.NewSyncWriter(os.Stderr))\nlogger = logutil.NewTrailingNilFilter(logger)\n- logger = log.With(logger, \"ts\", log.DefaultTimestampUTC, \"caller\", log.DefaultCaller)\n+ logger = log.With(logger, \"ts\", log.DefaultTimestampUTC, \"caller\", Caller(3))\nsigs := make(chan os.Signal, 1)\nstop := make(chan struct{})\n@@ -109,3 +111,7 @@ func (o *Options) Run(c *cobra.Command) error {\nreturn nil\n}\n+\n+func Caller(depth int) log.Valuer {\n+ return func() interface{} { return fmt.Sprintf(\"%+v\", stack.Caller(depth)) }\n+}\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
enriches caller with package path
|
596,240 |
07.12.2017 09:46:27
| -3,600 |
e9fb5890f7e01179eb2082b074df39dd0f98ca16
|
locks docker down to 1.2 until upgrade to k8s 1.8
|
[
{
"change_type": "MODIFY",
"old_path": "pkg/templates/node.go",
"new_path": "pkg/templates/node.go",
"diff": "@@ -188,6 +188,10 @@ storage:\ncontents:\ninline: |-\nnet.ipv4.conf.all.accept_redirects=1\n+ - path: /etc/coreos/docker-1.12\n+ filesystem: root\n+ contents:\n+ inline: yes\n- path: /etc/kubernetes/certs/kubelet-clients-ca.pem\nfilesystem: root\nmode: 0644\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
locks docker down to 1.2 until upgrade to k8s 1.8
|
596,230 |
11.12.2017 10:31:20
| -3,600 |
c97fcc303948987bb8a9fd3a859e0421556a89f1
|
add icmp to masquerade rules
|
[
{
"change_type": "MODIFY",
"old_path": "pkg/templates/node.go",
"new_path": "pkg/templates/node.go",
"diff": "@@ -180,6 +180,7 @@ storage:\n:POSTROUTING ACCEPT [0:0]\n-A POSTROUTING -p tcp ! -d {{ .ClusterCIDR }} -m addrtype ! --dst-type LOCAL -j MASQUERADE --to-ports 32000-65000\n-A POSTROUTING -p udp ! -d {{ .ClusterCIDR }} -m addrtype ! --dst-type LOCAL -j MASQUERADE --to-ports 32000-65000\n+ -A POSTROUTING -p icmp ! -d {{ .ClusterCIDR }} -m addrtype ! --dst-type LOCAL -j MASQUERADE\nCOMMIT\n- path: /etc/sysctl.d/10-enable-icmp-redirects\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
-- add icmp to masquerade rules
|
596,240 |
06.12.2017 13:53:54
| -3,600 |
34c02deb60b3a6cf3c2388b76319ed6a74b519d1
|
implements structured logging
|
[
{
"change_type": "MODIFY",
"old_path": "cmd/apiserver/main.go",
"new_path": "cmd/apiserver/main.go",
"diff": "@@ -3,16 +3,17 @@ package main\nimport (\ngoflag \"flag\"\n\"fmt\"\n- \"log\"\n\"os\"\n- \"github.com/golang/glog\"\n+ kitLog \"github.com/go-kit/kit/log\"\n+ \"github.com/go-stack/stack\"\n\"github.com/spf13/pflag\"\napipkg \"github.com/sapcc/kubernikus/pkg/api\"\n\"github.com/sapcc/kubernikus/pkg/api/rest\"\n\"github.com/sapcc/kubernikus/pkg/api/rest/operations\"\n\"github.com/sapcc/kubernikus/pkg/api/spec\"\n+ logutil \"github.com/sapcc/kubernikus/pkg/util/log\"\n\"github.com/sapcc/kubernikus/pkg/version\"\n)\n@@ -23,10 +24,17 @@ func init() {\n}\nfunc main() {\n+ var logger kitLog.Logger\n+ logger = kitLog.NewLogfmtLogger(kitLog.NewSyncWriter(os.Stderr))\n+ logger = logutil.NewTrailingNilFilter(logger)\n+ logger = kitLog.With(logger, \"ts\", kitLog.DefaultTimestampUTC, \"caller\", Caller(3))\nswaggerSpec, err := spec.Spec()\nif err != nil {\n- log.Fatalln(err)\n+ logger.Log(\n+ \"msg\", \"failed to spec swagger spec\",\n+ \"err\", err)\n+ os.Exit(1)\n}\nvar server *rest.Server // make sure init is called\n@@ -53,12 +61,21 @@ func main() {\napi := operations.NewKubernikusAPI(swaggerSpec)\n- rt := &apipkg.Runtime{Namespace: namespace}\n+ rt := &apipkg.Runtime{\n+ Namespace: namespace,\n+ Logger: logger,\n+ }\nrt.Kubernikus, rt.Kubernetes = rest.NewKubeClients()\nif err := rest.Configure(api, rt); err != nil {\n- glog.Fatalf(\"Failed to configure apiserver :%s\", err)\n+ logger.Log(\n+ \"msg\", \"failed to configure API server\",\n+ \"err\", err)\n+ os.Exit(1)\n}\n- glog.Infof(\"Starting kubernikus apiserver v%v. Using namespace %s\", version.GitCommit, namespace)\n+ logger.Log(\n+ \"msg\", \"starting Kubernikus API\",\n+ \"namespace\", namespace,\n+ \"version\", version.GitCommit)\n// get server with flag values filled out\nserver = rest.NewServer(api)\n@@ -67,7 +84,14 @@ func main() {\nserver.ConfigureAPI()\nif err := server.Serve(); err != nil {\n- log.Fatalln(err)\n+ logger.Log(\n+ \"msg\", \"failed to start API server\",\n+ \"err\", err)\n+ os.Exit(1)\n+ }\n+\n}\n+func Caller(depth int) kitLog.Valuer {\n+ return func() interface{} { return fmt.Sprintf(\"%+v\", stack.Caller(depth)) }\n}\n"
},
{
"change_type": "MODIFY",
"old_path": "pkg/api/handlers/get_openstack_metadata.go",
"new_path": "pkg/api/handlers/get_openstack_metadata.go",
"diff": "@@ -6,7 +6,7 @@ import (\n\"github.com/sapcc/kubernikus/pkg/api\"\n\"github.com/sapcc/kubernikus/pkg/api/models\"\n\"github.com/sapcc/kubernikus/pkg/api/rest/operations\"\n- \"github.com/sapcc/kubernikus/pkg/client/openstack\"\n+ \"github.com/sapcc/kubernikus/pkg/client/openstack/scoped\"\n)\nfunc NewGetOpenstackMetadata(rt *api.Runtime) operations.GetOpenstackMetadataHandler {\n@@ -29,7 +29,7 @@ func (d *getOpenstackMetadata) Handle(params operations.GetOpenstackMetadataPara\n},\n}\n- client, err := openstack.NewScopedClient(authOptions)\n+ client, err := scoped.NewClient(authOptions, d.Logger)\nif err != nil {\nreturn NewErrorResponse(&operations.GetOpenstackMetadataDefault{}, 500, err.Error())\n}\n"
},
{
"change_type": "MODIFY",
"old_path": "pkg/api/rest/api_test.go",
"new_path": "pkg/api/rest/api_test.go",
"diff": "@@ -9,6 +9,7 @@ import (\n\"strings\"\n\"testing\"\n+ kitlog \"github.com/go-kit/kit/log\"\nerrors \"github.com/go-openapi/errors\"\n\"github.com/stretchr/testify/assert\"\nmetav1 \"k8s.io/apimachinery/pkg/apis/meta/v1\"\n@@ -57,6 +58,7 @@ func createTestHandler(t *testing.T) (http.Handler, *apipkg.Runtime) {\nNamespace: NAMESPACE,\nKubernikus: kubernikusfake.NewSimpleClientset(),\nKubernetes: fake.NewSimpleClientset(),\n+ Logger: kitlog.NewNopLogger(),\n}\nif err := Configure(api, rt); err != nil {\nt.Fatal(err)\n"
},
{
"change_type": "MODIFY",
"old_path": "pkg/api/rest/configure.go",
"new_path": "pkg/api/rest/configure.go",
"diff": "@@ -2,16 +2,21 @@ package rest\nimport (\n\"fmt\"\n+ \"net/http\"\n+ \"strings\"\n\"github.com/go-openapi/errors\"\nruntime \"github.com/go-openapi/runtime\"\n- \"github.com/golang/glog\"\n+ \"github.com/go-openapi/runtime/middleware\"\n+ \"github.com/justinas/alice\"\n+ \"github.com/rs/cors\"\napipkg \"github.com/sapcc/kubernikus/pkg/api\"\n\"github.com/sapcc/kubernikus/pkg/api/auth\"\n\"github.com/sapcc/kubernikus/pkg/api/handlers\"\n\"github.com/sapcc/kubernikus/pkg/api/rest/operations\"\n\"github.com/sapcc/kubernikus/pkg/api/spec\"\n+ logutil \"github.com/sapcc/kubernikus/pkg/util/log\"\n)\nfunc Configure(api *operations.KubernikusAPI, rt *apipkg.Runtime) error {\n@@ -23,7 +28,7 @@ func Configure(api *operations.KubernikusAPI, rt *apipkg.Runtime) error {\n//\n// Example:\napi.Logger = func(msg string, args ...interface{}) {\n- glog.InfoDepth(2, fmt.Sprintf(msg, args...))\n+ rt.Logger.Log(\"msg\", fmt.Sprintf(msg, args...))\n}\napi.JSONConsumer = runtime.JSONConsumer()\n@@ -60,5 +65,39 @@ func Configure(api *operations.KubernikusAPI, rt *apipkg.Runtime) error {\napi.GetClusterEventsHandler = handlers.NewGetClusterEvents(rt)\napi.ServerShutdown = func() {}\n+\n+ api.Middleware = func(builder middleware.Builder) http.Handler {\n+ return setupGlobalMiddleware(api.Context().APIHandler(builder), rt)\n+ }\nreturn nil\n}\n+\n+// The middleware configuration happens before anything, this middleware also applies to serving the swagger.json document.\n+// So this is a good place to plug in a panic handling middleware, logging and metrics\n+func setupGlobalMiddleware(handler http.Handler, rt *apipkg.Runtime) http.Handler {\n+ corsHandler := cors.New(cors.Options{\n+ AllowedHeaders: []string{\"X-Auth-Token\", \"Content-Type\", \"Accept\"},\n+ AllowedMethods: []string{\"GET\", \"HEAD\", \"POST\", \"DELETE\", \"PUT\"},\n+ MaxAge: 600,\n+ }).Handler\n+\n+ loggingHandler := func(next http.Handler) http.Handler {\n+ return logutil.LoggingHandler(rt.Logger, next)\n+ }\n+\n+ redocHandler := func(next http.Handler) http.Handler {\n+ return middleware.Redoc(middleware.RedocOpts{Path: \"swagger\"}, next)\n+ }\n+\n+ staticHandler := func(next http.Handler) http.Handler {\n+ return http.HandlerFunc(func(rw http.ResponseWriter, r *http.Request) {\n+ if strings.HasPrefix(r.URL.Path, \"/docs\") {\n+ http.StripPrefix(\"/docs\", http.FileServer(http.Dir(\"static/docs\"))).ServeHTTP(rw, r)\n+ return\n+ }\n+ next.ServeHTTP(rw, r)\n+ })\n+ }\n+\n+ return alice.New(loggingHandler, handlers.RootHandler, redocHandler, staticHandler, corsHandler).Then(handler)\n+}\n"
},
{
"change_type": "MODIFY",
"old_path": "pkg/api/rest/configure_kubernikus.go",
"new_path": "pkg/api/rest/configure_kubernikus.go",
"diff": "@@ -3,16 +3,9 @@ package rest\nimport (\n\"crypto/tls\"\n\"net/http\"\n- \"os\"\n- \"strings\"\n- \"github.com/go-openapi/runtime/middleware\"\n- gmiddleware \"github.com/gorilla/handlers\"\n- \"github.com/justinas/alice\"\n- \"github.com/rs/cors\"\ngraceful \"github.com/tylerb/graceful\"\n- \"github.com/sapcc/kubernikus/pkg/api/handlers\"\n\"github.com/sapcc/kubernikus/pkg/api/rest/operations\"\n)\n@@ -25,8 +18,9 @@ func configureFlags(api *operations.KubernikusAPI) {\n}\nfunc configureAPI(api *operations.KubernikusAPI) http.Handler {\n-\n- return setupGlobalMiddleware(api.Serve(setupMiddlewares))\n+ return api.Serve(func(handler http.Handler) http.Handler {\n+ return handler\n+ })\n}\n// The TLS configuration before HTTPS server starts.\n@@ -40,38 +34,3 @@ func configureTLS(tlsConfig *tls.Config) {\n// scheme value will be set accordingly: \"http\", \"https\" or \"unix\"\nfunc configureServer(s *graceful.Server, scheme, addr string) {\n}\n-\n-// The middleware configuration is for the handler executors. These do not apply to the swagger.json document.\n-// The middleware executes after routing but before authentication, binding and validation\n-func setupMiddlewares(handler http.Handler) http.Handler {\n- return handler\n-}\n-\n-// The middleware configuration happens before anything, this middleware also applies to serving the swagger.json document.\n-// So this is a good place to plug in a panic handling middleware, logging and metrics\n-func setupGlobalMiddleware(handler http.Handler) http.Handler {\n- corsHandler := cors.New(cors.Options{\n- AllowedHeaders: []string{\"X-Auth-Token\", \"Content-Type\", \"Accept\"},\n- AllowedMethods: []string{\"GET\", \"HEAD\", \"POST\", \"DELETE\", \"PUT\"},\n- MaxAge: 600,\n- }).Handler\n-\n- loggingHandler := func(next http.Handler) http.Handler {\n- return gmiddleware.LoggingHandler(os.Stdout, next)\n- }\n- redocHandler := func(next http.Handler) http.Handler {\n- return middleware.Redoc(middleware.RedocOpts{Path: \"swagger\"}, next)\n- }\n-\n- return alice.New(loggingHandler, handlers.RootHandler, redocHandler, StaticFiles, corsHandler).Then(handler)\n-}\n-\n-func StaticFiles(next http.Handler) http.Handler {\n- return http.HandlerFunc(func(rw http.ResponseWriter, r *http.Request) {\n- if strings.HasPrefix(r.URL.Path, \"/docs\") {\n- http.StripPrefix(\"/docs\", http.FileServer(http.Dir(\"static/docs\"))).ServeHTTP(rw, r)\n- return\n- }\n- next.ServeHTTP(rw, r)\n- })\n-}\n"
},
{
"change_type": "MODIFY",
"old_path": "pkg/api/runtime.go",
"new_path": "pkg/api/runtime.go",
"diff": "package api\nimport (\n+ \"github.com/go-kit/kit/log\"\n+\n\"github.com/sapcc/kubernikus/pkg/generated/clientset\"\n\"k8s.io/client-go/kubernetes\"\n)\n@@ -9,4 +11,5 @@ type Runtime struct {\nKubernikus clientset.Interface\nKubernetes kubernetes.Interface\nNamespace string\n+ Logger log.Logger\n}\n"
},
{
"change_type": "RENAME",
"old_path": "pkg/client/openstack/metadata.go",
"new_path": "pkg/client/openstack/scoped/client.go",
"diff": "-package openstack\n+package scoped\nimport (\n+ \"github.com/go-kit/kit/log\"\n\"github.com/gophercloud/gophercloud\"\n\"github.com/gophercloud/gophercloud/openstack\"\n\"github.com/gophercloud/gophercloud/openstack/compute/v2/extensions/keypairs\"\n@@ -13,73 +14,85 @@ import (\n\"github.com/gophercloud/gophercloud/openstack/networking/v2/subnets\"\n\"github.com/gophercloud/gophercloud/pagination\"\n\"github.com/sapcc/kubernikus/pkg/api/models\"\n+ \"github.com/sapcc/kubernikus/pkg/client/openstack/util\"\n+ utillog \"github.com/sapcc/kubernikus/pkg/util/log\"\n)\n-type scopedClient struct {\n- providerClient *gophercloud.ProviderClient\n- networkClient *gophercloud.ServiceClient\n- computeClient *gophercloud.ServiceClient\n- identityClient *gophercloud.ServiceClient\n+type client struct {\n+ util.AuthenticatedClient\n+ Logger log.Logger\n}\n-type ScopedClient interface {\n+type Client interface {\n+ Authenticate(*tokens.AuthOptions) error\nGetMetadata() (*models.OpenstackMetadata, error)\n}\n-func NewScopedClient(authOptions *tokens.AuthOptions) (ScopedClient, error) {\n- var err error\n- client := &scopedClient{}\n+func NewClient(authOptions *tokens.AuthOptions, logger log.Logger) (Client, error) {\n+ logger = utillog.NewAuthLogger(logger, authOptions)\n- if client.providerClient, err = openstack.NewClient(authOptions.IdentityEndpoint); err != nil {\n- return nil, err\n+ var c Client\n+ c = &client{Logger: logger}\n+ c = LoggingClient{c, logger}\n+\n+ return c, c.Authenticate(authOptions)\n}\n- if err := openstack.AuthenticateV3(client.providerClient, authOptions, gophercloud.EndpointOpts{}); err != nil {\n- return nil, err\n+func (c *client) Authenticate(authOptions *tokens.AuthOptions) error {\n+ providerClient, err := utillog.NewLoggingProviderClient(authOptions.IdentityEndpoint, c.Logger)\n+ if err != nil {\n+ return err\n}\n- if client.identityClient, err = openstack.NewIdentityV3(client.providerClient, gophercloud.EndpointOpts{}); err != nil {\n- return nil, err\n+ if err := openstack.AuthenticateV3(providerClient, authOptions, gophercloud.EndpointOpts{}); err != nil {\n+ return err\n}\n- if client.computeClient, err = openstack.NewComputeV2(client.providerClient, gophercloud.EndpointOpts{}); err != nil {\n- return nil, err\n+ if c.IdentityClient, err = openstack.NewIdentityV3(providerClient, gophercloud.EndpointOpts{}); err != nil {\n+ return err\n}\n- if client.networkClient, err = openstack.NewNetworkV2(client.providerClient, gophercloud.EndpointOpts{}); err != nil {\n- return nil, err\n+ if c.ComputeClient, err = openstack.NewComputeV2(providerClient, gophercloud.EndpointOpts{}); err != nil {\n+ return err\n+ }\n+\n+ if c.NetworkClient, err = openstack.NewNetworkV2(providerClient, gophercloud.EndpointOpts{}); err != nil {\n+ return err\n}\n- return client, nil\n+ return nil\n}\n-func (c *scopedClient) GetMetadata() (*models.OpenstackMetadata, error) {\n- var err error\n- metadata := &models.OpenstackMetadata{}\n+func (c *client) GetMetadata() (metadata *models.OpenstackMetadata, err error) {\n+ metadata = &models.OpenstackMetadata{\n+ Flavors: make([]*models.Flavor, 0),\n+ KeyPairs: make([]*models.KeyPair, 0),\n+ Routers: make([]*models.Router, 0),\n+ SecurityGroups: make([]*models.SecurityGroup, 0),\n+ }\nif metadata.Routers, err = c.getRouters(); err != nil {\n- return nil, err\n+ return metadata, err\n}\nif metadata.KeyPairs, err = c.getKeyPairs(); err != nil {\n- return nil, err\n+ return metadata, err\n}\nif metadata.SecurityGroups, err = c.getSecurityGroups(); err != nil {\n- return nil, err\n+ return metadata, err\n}\nif metadata.Flavors, err = c.getFlavors(); err != nil {\n- return nil, err\n+ return metadata, err\n}\n-\nreturn metadata, nil\n}\n-func (c *scopedClient) getRouters() ([]*models.Router, error) {\n+func (c *client) getRouters() ([]*models.Router, error) {\nresult := []*models.Router{}\n- err := routers.List(c.networkClient, routers.ListOpts{}).EachPage(func(page pagination.Page) (bool, error) {\n+ err := routers.List(c.NetworkClient, routers.ListOpts{}).EachPage(func(page pagination.Page) (bool, error) {\nif routerList, err := routers.ExtractRouters(page); err != nil {\nreturn false, err\n} else {\n@@ -91,30 +104,30 @@ func (c *scopedClient) getRouters() ([]*models.Router, error) {\n})\nif err != nil {\n- return nil, err\n+ return result, err\n}\nfor _, router := range result {\nif router.Networks, err = c.getNetworks(router); err != nil {\n- return nil, err\n+ return result, err\n}\n}\nreturn result, nil\n}\n-func (c *scopedClient) getNetworks(router *models.Router) ([]*models.Network, error) {\n+func (c *client) getNetworks(router *models.Router) ([]*models.Network, error) {\nresult := []*models.Network{}\nnetworkIDs, err := c.getRouterNetworkIDs(router)\nif err != nil {\n- return nil, err\n+ return result, err\n}\nfor _, networkID := range networkIDs {\n- network, err := networks.Get(c.networkClient, networkID).Extract()\n+ network, err := networks.Get(c.NetworkClient, networkID).Extract()\nif err != nil {\n- return nil, err\n+ return result, err\n}\nresult = append(result, &models.Network{ID: network.ID, Name: network.Name})\n}\n@@ -128,10 +141,10 @@ func (c *scopedClient) getNetworks(router *models.Router) ([]*models.Network, er\nreturn result, nil\n}\n-func (c *scopedClient) getRouterNetworkIDs(router *models.Router) ([]string, error) {\n+func (c *client) getRouterNetworkIDs(router *models.Router) ([]string, error) {\nresult := []string{}\n- err := ports.List(c.networkClient, ports.ListOpts{DeviceID: router.ID, DeviceOwner: \"network:router_interface\"}).EachPage(func(page pagination.Page) (bool, error) {\n+ err := ports.List(c.NetworkClient, ports.ListOpts{DeviceID: router.ID, DeviceOwner: \"network:router_interface\"}).EachPage(func(page pagination.Page) (bool, error) {\nportList, err := ports.ExtractPorts(page)\nif err != nil {\nreturn false, err\n@@ -145,27 +158,27 @@ func (c *scopedClient) getRouterNetworkIDs(router *models.Router) ([]string, err\nreturn result, err\n}\n-func (c *scopedClient) getSubnetIDs(network *models.Network) ([]string, error) {\n- result, err := networks.Get(c.networkClient, network.ID).Extract()\n+func (c *client) getSubnetIDs(network *models.Network) ([]string, error) {\n+ result, err := networks.Get(c.NetworkClient, network.ID).Extract()\nif err != nil {\n- return nil, err\n+ return []string{}, err\n}\nreturn result.Subnets, nil\n}\n-func (c *scopedClient) getSubnets(network *models.Network) ([]*models.Subnet, error) {\n+func (c *client) getSubnets(network *models.Network) ([]*models.Subnet, error) {\nresult := []*models.Subnet{}\nsubnetIDs, err := c.getSubnetIDs(network)\nif err != nil {\n- return nil, err\n+ return result, err\n}\nfor _, subnetID := range subnetIDs {\n- subnet, err := subnets.Get(c.networkClient, subnetID).Extract()\n+ subnet, err := subnets.Get(c.NetworkClient, subnetID).Extract()\nif err != nil {\n- return nil, err\n+ return result, err\n}\nresult = append(result, &models.Subnet{ID: subnet.ID, Name: subnet.Name, CIDR: subnet.CIDR})\n}\n@@ -173,17 +186,17 @@ func (c *scopedClient) getSubnets(network *models.Network) ([]*models.Subnet, er\nreturn result, nil\n}\n-func (c *scopedClient) getKeyPairs() ([]*models.KeyPair, error) {\n+func (c *client) getKeyPairs() ([]*models.KeyPair, error) {\nresult := []*models.KeyPair{}\n- pager, err := keypairs.List(c.computeClient).AllPages()\n+ pager, err := keypairs.List(c.ComputeClient).AllPages()\nif err != nil {\n- return nil, err\n+ return result, err\n}\nkeyList, err := keypairs.ExtractKeyPairs(pager)\nif err != nil {\n- return nil, err\n+ return result, err\n}\nfor _, key := range keyList {\n@@ -193,10 +206,10 @@ func (c *scopedClient) getKeyPairs() ([]*models.KeyPair, error) {\nreturn result, nil\n}\n-func (c *scopedClient) getSecurityGroups() ([]*models.SecurityGroup, error) {\n+func (c *client) getSecurityGroups() ([]*models.SecurityGroup, error) {\nresult := []*models.SecurityGroup{}\n- err := secgroups.List(c.computeClient).EachPage(func(page pagination.Page) (bool, error) {\n+ err := secgroups.List(c.ComputeClient).EachPage(func(page pagination.Page) (bool, error) {\nsecGroupList, err := secgroups.ExtractSecurityGroups(page)\nif err != nil {\nreturn false, err\n@@ -210,10 +223,10 @@ func (c *scopedClient) getSecurityGroups() ([]*models.SecurityGroup, error) {\nreturn result, err\n}\n-func (c *scopedClient) getFlavors() ([]*models.Flavor, error) {\n+func (c *client) getFlavors() ([]*models.Flavor, error) {\nresult := []*models.Flavor{}\n- err := flavors.ListDetail(c.computeClient, &flavors.ListOpts{}).EachPage(func(page pagination.Page) (bool, error) {\n+ err := flavors.ListDetail(c.ComputeClient, &flavors.ListOpts{}).EachPage(func(page pagination.Page) (bool, error) {\nlist, err := flavors.ExtractFlavors(page)\nif err != nil {\nreturn false, err\n"
},
{
"change_type": "ADD",
"old_path": null,
"new_path": "pkg/client/openstack/scoped/logging.go",
"diff": "+package scoped\n+\n+import (\n+ \"time\"\n+\n+ \"github.com/go-kit/kit/log\"\n+ \"github.com/gophercloud/gophercloud/openstack/identity/v3/tokens\"\n+\n+ \"github.com/sapcc/kubernikus/pkg/api/models\"\n+)\n+\n+type LoggingClient struct {\n+ Client Client\n+ Logger log.Logger\n+}\n+\n+func (c LoggingClient) GetMetadata() (metadata *models.OpenstackMetadata, err error) {\n+ defer func(begin time.Time) {\n+ c.Logger.Log(\n+ \"msg\", \"fetched metadata\",\n+ \"flavors\", len(metadata.Flavors),\n+ \"keypairs\", len(metadata.KeyPairs),\n+ \"routers\", len(metadata.Routers),\n+ \"security_groups\", len(metadata.SecurityGroups),\n+ \"took\", time.Since(begin),\n+ \"v\", 1,\n+ \"err\", err,\n+ )\n+ }(time.Now())\n+\n+ return c.Client.GetMetadata()\n+}\n+\n+func (c LoggingClient) Authenticate(authOptions *tokens.AuthOptions) (err error) {\n+ defer func(begin time.Time) {\n+ v := 2\n+ if err != nil {\n+ v = 0\n+ }\n+ c.Logger.Log(\n+ \"msg\", \"authenticated\",\n+ \"took\", time.Since(begin),\n+ \"v\", v,\n+ \"err\", err,\n+ )\n+ }(time.Now())\n+ return c.Client.Authenticate(authOptions)\n+}\n"
},
{
"change_type": "ADD",
"old_path": null,
"new_path": "pkg/client/openstack/util/factory.go",
"diff": "+package util\n+\n+import (\n+ \"github.com/gophercloud/gophercloud\"\n+)\n+\n+type AuthenticatedClient struct {\n+ providerClient *gophercloud.ProviderClient\n+ NetworkClient *gophercloud.ServiceClient\n+ ComputeClient *gophercloud.ServiceClient\n+ IdentityClient *gophercloud.ServiceClient\n+}\n"
},
{
"change_type": "ADD",
"old_path": null,
"new_path": "pkg/util/log/authlogger.go",
"diff": "+package log\n+\n+import (\n+ \"fmt\"\n+\n+ kitlog \"github.com/go-kit/kit/log\"\n+ \"github.com/gophercloud/gophercloud/openstack/identity/v3/tokens\"\n+)\n+\n+func NewAuthLogger(logger kitlog.Logger, authOptions *tokens.AuthOptions) kitlog.Logger {\n+ if project := getProject(authOptions); project != \"\" {\n+ logger = kitlog.With(logger, \"project\", project)\n+ }\n+\n+ if authMethod := getAuthMethod(authOptions); authMethod != \"\" {\n+ logger = kitlog.With(logger, \"auth\", authMethod)\n+ }\n+\n+ if principal := getPrincipal(authOptions); principal != \"\" {\n+ logger = kitlog.With(logger, \"principal\", principal)\n+ }\n+ return logger\n+}\n+\n+func getProject(authOptions *tokens.AuthOptions) string {\n+ if authOptions.Scope.ProjectID != \"\" {\n+ return authOptions.Scope.ProjectID\n+ }\n+\n+ domain := \"\"\n+ if authOptions.Scope.DomainID != \"\" {\n+ domain = authOptions.Scope.DomainID\n+ } else {\n+ domain = authOptions.Scope.DomainName\n+ }\n+\n+ return fmt.Sprintf(\"%s/%s\", domain, authOptions.Scope.ProjectName)\n+}\n+\n+func getAuthMethod(authOptions *tokens.AuthOptions) string {\n+ if authOptions.TokenID != \"\" {\n+ return \"token\"\n+ }\n+\n+ if authOptions.Password != \"\" {\n+ return \"password\"\n+ }\n+\n+ return \"\"\n+}\n+\n+func getPrincipal(authOptions *tokens.AuthOptions) string {\n+ if authOptions.TokenID != \"\" {\n+ return \"\"\n+ }\n+\n+ if authOptions.UserID != \"\" {\n+ return authOptions.UserID\n+ }\n+\n+ domain := \"\"\n+ if authOptions.DomainID != \"\" {\n+ domain = authOptions.DomainID\n+ } else {\n+ domain = authOptions.DomainName\n+ }\n+\n+ return fmt.Sprintf(\"%s/%s\", domain, authOptions.Username)\n+}\n"
},
{
"change_type": "ADD",
"old_path": null,
"new_path": "pkg/util/log/gophercloud.go",
"diff": "+package log\n+\n+import (\n+ \"net/http\"\n+ \"strings\"\n+ \"time\"\n+\n+ kitlog \"github.com/go-kit/kit/log\"\n+ \"github.com/gophercloud/gophercloud\"\n+ \"github.com/gophercloud/gophercloud/openstack\"\n+)\n+\n+func NewLoggingProviderClient(endpoint string, logger kitlog.Logger) (*gophercloud.ProviderClient, error) {\n+ providerClient, err := openstack.NewClient(endpoint)\n+ if err != nil {\n+ return nil, err\n+ }\n+\n+ transport := providerClient.HTTPClient.Transport\n+ if transport == nil {\n+ transport = http.DefaultTransport\n+ }\n+ providerClient.HTTPClient.Transport = &loggingRoundTripper{\n+ transport,\n+ kitlog.With(logger, \"api\", \"egress\"),\n+ }\n+\n+ return providerClient, err\n+}\n+\n+type loggingRoundTripper struct {\n+ rt http.RoundTripper\n+ Logger kitlog.Logger\n+}\n+\n+func (lrt *loggingRoundTripper) RoundTrip(request *http.Request) (response *http.Response, err error) {\n+ defer func(begin time.Time) {\n+ keyvals := make([]interface{}, 0, 6)\n+\n+ if response != nil {\n+ keyvals = append(keyvals,\n+ \"status\", response.StatusCode,\n+ \"openstack_id\", strings.Join(requestIds(response), \",\"))\n+ }\n+\n+ keyvals = append(keyvals,\n+ \"took\", time.Since(begin),\n+ \"v\", 2,\n+ \"err\", err,\n+ )\n+\n+ log(lrt.Logger, request, keyvals...)\n+ }(time.Now())\n+\n+ return lrt.rt.RoundTrip(request)\n+}\n+\n+func requestIds(response *http.Response) []string {\n+ ids := []string{}\n+\n+ if id := response.Header.Get(\"X-Openstack-Request-ID\"); id != \"\" {\n+ ids = append(ids, id)\n+ }\n+\n+ if id := response.Header.Get(\"X-Compute-Request-ID\"); id != \"\" {\n+ ids = append(ids, id)\n+ }\n+\n+ return ids\n+}\n"
},
{
"change_type": "ADD",
"old_path": null,
"new_path": "pkg/util/log/middleware.go",
"diff": "+package log\n+\n+import (\n+ \"bufio\"\n+ \"fmt\"\n+ \"net\"\n+ \"net/http\"\n+ \"strings\"\n+ \"time\"\n+\n+ kitlog \"github.com/go-kit/kit/log\"\n+)\n+\n+func LoggingHandler(logger kitlog.Logger, next http.Handler) http.Handler {\n+ logger = kitlog.With(logger, \"api\", \"ingress\")\n+ return http.HandlerFunc(func(rw http.ResponseWriter, request *http.Request) {\n+ wrapper := makeWrapper(rw)\n+\n+ defer func(begin time.Time) {\n+ log(logger, request,\n+ \"status\", wrapper.Status(),\n+ \"size\", wrapper.Size(),\n+ \"took\", time.Since(begin))\n+ }(time.Now())\n+\n+ next.ServeHTTP(wrapper, request)\n+ })\n+}\n+\n+func log(logger kitlog.Logger, request *http.Request, extra ...interface{}) {\n+ var keyvals []interface{}\n+\n+ source_ip, _, err := net.SplitHostPort(request.RemoteAddr)\n+ if err != nil {\n+ source_ip = request.RemoteAddr\n+ }\n+\n+ if source_ip != \"\" {\n+ keyvals = append(keyvals, \"source_ip\", source_ip)\n+ }\n+\n+ keyvals = append(keyvals, \"method\", request.Method)\n+\n+ host, host_port, err := net.SplitHostPort(request.Host)\n+ if err == nil {\n+ if host != \"\" {\n+ keyvals = append(keyvals,\n+ \"host\", host)\n+ }\n+ if host_port != \"\" {\n+ keyvals = append(keyvals,\n+ \"port\", host_port)\n+ }\n+ }\n+\n+ keyvals = append(keyvals, \"path\", request.URL.EscapedPath())\n+\n+ for i, k := range request.URL.Query() {\n+ keyvals = append(keyvals, i, strings.Join(k, \",\"))\n+ }\n+\n+ keyvals = append(keyvals, \"user_agent\", request.UserAgent())\n+ keyvals = append(keyvals, extra...)\n+ logger.Log(keyvals...)\n+}\n+\n+// this stuff is copied from gorilla\n+\n+func makeWrapper(w http.ResponseWriter) loggingResponseWriter {\n+ var logger loggingResponseWriter = &responseLogger{w: w, status: http.StatusOK}\n+ if _, ok := w.(http.Hijacker); ok {\n+ logger = &hijackLogger{responseLogger{w: w, status: http.StatusOK}}\n+ }\n+ h, ok1 := logger.(http.Hijacker)\n+ c, ok2 := w.(http.CloseNotifier)\n+ if ok1 && ok2 {\n+ return hijackCloseNotifier{logger, h, c}\n+ }\n+ if ok2 {\n+ return &closeNotifyWriter{logger, c}\n+ }\n+ return logger\n+}\n+\n+type hijackLogger struct {\n+ responseLogger\n+}\n+\n+func (l *hijackLogger) Hijack() (net.Conn, *bufio.ReadWriter, error) {\n+ h := l.responseLogger.w.(http.Hijacker)\n+ conn, rw, err := h.Hijack()\n+ if err == nil && l.responseLogger.status == 0 {\n+ // The status will be StatusSwitchingProtocols if there was no error and\n+ // WriteHeader has not been called yet\n+ l.responseLogger.status = http.StatusSwitchingProtocols\n+ }\n+ return conn, rw, err\n+}\n+\n+type closeNotifyWriter struct {\n+ loggingResponseWriter\n+ http.CloseNotifier\n+}\n+\n+type hijackCloseNotifier struct {\n+ loggingResponseWriter\n+ http.Hijacker\n+ http.CloseNotifier\n+}\n+\n+type loggingResponseWriter interface {\n+ commonLoggingResponseWriter\n+ http.Pusher\n+}\n+\n+type commonLoggingResponseWriter interface {\n+ http.ResponseWriter\n+ http.Flusher\n+ Status() int\n+ Size() int\n+}\n+\n+type responseLogger struct {\n+ w http.ResponseWriter\n+ status int\n+ size int\n+}\n+\n+func (l *responseLogger) Header() http.Header {\n+ return l.w.Header()\n+}\n+\n+func (l *responseLogger) Write(b []byte) (int, error) {\n+ size, err := l.w.Write(b)\n+ l.size += size\n+ return size, err\n+}\n+\n+func (l *responseLogger) WriteHeader(s int) {\n+ l.w.WriteHeader(s)\n+ l.status = s\n+}\n+\n+func (l *responseLogger) Status() int {\n+ return l.status\n+}\n+\n+func (l *responseLogger) Size() int {\n+ return l.size\n+}\n+\n+func (l *responseLogger) Flush() {\n+ f, ok := l.w.(http.Flusher)\n+ if ok {\n+ f.Flush()\n+ }\n+}\n+\n+func (l *responseLogger) Push(target string, opts *http.PushOptions) error {\n+ p, ok := l.w.(http.Pusher)\n+ if !ok {\n+ return fmt.Errorf(\"responseLogger does not implement http.Pusher\")\n+ }\n+ return p.Push(target, opts)\n+}\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
implements structured logging
|
596,240 |
11.12.2017 15:28:16
| -3,600 |
53fac3a585e9c1e6916ecc98b03ba973a8ea7538
|
implements tracing id
|
[
{
"change_type": "MODIFY",
"old_path": "pkg/api/handlers/get_openstack_metadata.go",
"new_path": "pkg/api/handlers/get_openstack_metadata.go",
"diff": "@@ -30,7 +30,7 @@ func (d *getOpenstackMetadata) Handle(params operations.GetOpenstackMetadataPara\n},\n}\n- client, err := scoped.NewClient(authOptions, d.Logger)\n+ client, err := scoped.NewClient(authOptions, getTracingLogger(params.HTTPRequest))\nif err != nil {\nreturn NewErrorResponse(&operations.GetOpenstackMetadataDefault{}, 500, err.Error())\n}\n"
},
{
"change_type": "MODIFY",
"old_path": "pkg/api/handlers/util.go",
"new_path": "pkg/api/handlers/util.go",
"diff": "@@ -2,8 +2,10 @@ package handlers\nimport (\n\"fmt\"\n+ \"net/http\"\n\"strings\"\n+ kitlog \"github.com/go-kit/kit/log\"\nmetav1 \"k8s.io/apimachinery/pkg/apis/meta/v1\"\n\"k8s.io/apimachinery/pkg/labels\"\n@@ -52,3 +54,11 @@ func klusterFromCRD(k *v1.Kluster) *models.Kluster {\nStatus: k.Status,\n}\n}\n+\n+func getTracingLogger(request *http.Request) kitlog.Logger {\n+ logger, ok := request.Context().Value(\"logger\").(kitlog.Logger)\n+ if !ok {\n+ logger = kitlog.NewNopLogger()\n+ }\n+ return logger\n+}\n"
},
{
"change_type": "MODIFY",
"old_path": "pkg/api/rest/configure.go",
"new_path": "pkg/api/rest/configure.go",
"diff": "@@ -81,6 +81,10 @@ func setupGlobalMiddleware(handler http.Handler, rt *apipkg.Runtime) http.Handle\nMaxAge: 600,\n}).Handler\n+ requestIDHandler := func(next http.Handler) http.Handler {\n+ return logutil.RequestIDHandler(next)\n+ }\n+\nloggingHandler := func(next http.Handler) http.Handler {\nreturn logutil.LoggingHandler(rt.Logger, next)\n}\n@@ -99,5 +103,5 @@ func setupGlobalMiddleware(handler http.Handler, rt *apipkg.Runtime) http.Handle\n})\n}\n- return alice.New(loggingHandler, handlers.RootHandler, redocHandler, staticHandler, corsHandler).Then(handler)\n+ return alice.New(requestIDHandler, loggingHandler, handlers.RootHandler, redocHandler, staticHandler, corsHandler).Then(handler)\n}\n"
},
{
"change_type": "MODIFY",
"old_path": "pkg/util/log/gophercloud.go",
"new_path": "pkg/util/log/gophercloud.go",
"diff": "package log\nimport (\n+ \"fmt\"\n\"net/http\"\n\"strings\"\n\"time\"\n@@ -43,6 +44,10 @@ func (lrt *loggingRoundTripper) RoundTrip(request *http.Request) (response *http\n\"openstack_id\", strings.Join(requestIds(response), \",\"))\n}\n+ if id := request.Context().Value(KubernikusRequestID); id != nil {\n+ keyvals = append(keyvals, \"id\", fmt.Sprintf(\"%s\", id))\n+ }\n+\nkeyvals = append(keyvals,\n\"took\", time.Since(begin),\n\"v\", 2,\n"
},
{
"change_type": "MODIFY",
"old_path": "pkg/util/log/middleware.go",
"new_path": "pkg/util/log/middleware.go",
"diff": "@@ -2,6 +2,7 @@ package log\nimport (\n\"bufio\"\n+ \"context\"\n\"fmt\"\n\"net\"\n\"net/http\"\n@@ -9,18 +10,50 @@ import (\n\"time\"\nkitlog \"github.com/go-kit/kit/log\"\n+ uuid \"github.com/satori/go.uuid\"\n)\n+type key int\n+\n+const (\n+ KubernikusRequestID key = 0\n+)\n+\n+func RequestIDHandler(next http.Handler) http.Handler {\n+ return http.HandlerFunc(func(rw http.ResponseWriter, request *http.Request) {\n+ if id := request.Context().Value(KubernikusRequestID); id == nil {\n+ request = request.WithContext(context.WithValue(request.Context(), KubernikusRequestID, uuid.NewV4()))\n+ }\n+ next.ServeHTTP(rw, request)\n+ })\n+}\n+\nfunc LoggingHandler(logger kitlog.Logger, next http.Handler) http.Handler {\n- logger = kitlog.With(logger, \"api\", \"ingress\")\n+ ingress_logger := kitlog.With(logger, \"api\", \"ingress\")\nreturn http.HandlerFunc(func(rw http.ResponseWriter, request *http.Request) {\nwrapper := makeWrapper(rw)\n+ id := \"\"\n+ if reqId := request.Context().Value(KubernikusRequestID); reqId != nil {\n+ id = fmt.Sprintf(\"%s\", reqId)\n+ logger = kitlog.With(logger, \"id\", id)\n+ }\n+ request = request.WithContext(context.WithValue(request.Context(), \"logger\", logger))\n+\ndefer func(begin time.Time) {\n- log(logger, request,\n+ var keyvals = make([]interface{}, 0, 4)\n+\n+ keyvals = append(keyvals,\n\"status\", wrapper.Status(),\n\"size\", wrapper.Size(),\n- \"took\", time.Since(begin))\n+ \"took\", time.Since(begin),\n+ )\n+\n+ if id != \"\" {\n+ keyvals = append(keyvals, \"id\", id)\n+ }\n+\n+ log(ingress_logger, request, keyvals...)\n}(time.Now())\nnext.ServeHTTP(wrapper, request)\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
implements tracing id
|
596,240 |
13.12.2017 17:23:52
| -3,600 |
207b04340d83d023a12adb9c1dd20bdf9b3a2a35
|
adds special self-signed root cert
|
[
{
"change_type": "MODIFY",
"old_path": "pkg/templates/node.go",
"new_path": "pkg/templates/node.go",
"diff": "@@ -149,6 +149,20 @@ systemd:\nRestartSec=10s\n[Install]\nWantedBy=multi-user.target\n+ - name: updatecertificates.service\n+ command: start\n+ enable: true\n+ contents: |\n+ [Unit]\n+ Description=Update the certificates w/ self-signed root CAs\n+ ConditionPathIsSymbolicLink=!/etc/ssl/certs/48b11003.0\n+ Before=early-docker.service docker.service\n+ [Service]\n+ ExecStart=/usr/sbin/update-ca-certificates\n+ RemainAfterExit=yes\n+ Type=oneshot\n+ [Install]\n+ WantedBy=multi-user.target\nnetworkd:\nunits:\n@@ -168,6 +182,45 @@ networkd:\nstorage:\nfiles:\n+ - path: /etc/ssl/certs/SAPNetCA_G2.pem\n+ mode: 0644\n+ contents: |\n+ -----BEGIN CERTIFICATE-----\n+ MIIGPTCCBCWgAwIBAgIKYQ4GNwAAAAAADDANBgkqhkiG9w0BAQsFADBOMQswCQYD\n+ VQQGEwJERTERMA8GA1UEBwwIV2FsbGRvcmYxDzANBgNVBAoMBlNBUCBBRzEbMBkG\n+ A1UEAwwSU0FQIEdsb2JhbCBSb290IENBMB4XDTE1MDMxNzA5MjQ1MVoXDTI1MDMx\n+ NzA5MzQ1MVowRDELMAkGA1UEBhMCREUxETAPBgNVBAcMCFdhbGxkb3JmMQwwCgYD\n+ VQQKDANTQVAxFDASBgNVBAMMC1NBUE5ldENBX0cyMIICIjANBgkqhkiG9w0BAQEF\n+ AAOCAg8AMIICCgKCAgEAjuP7Hj/1nVWfsCr8M/JX90s88IhdTLaoekrxpLNJ1W27\n+ ECUQogQF6HCu/RFD4uIoanH0oGItbmp2p8I0XVevHXnisxQGxBdkjz+a6ZyOcEVk\n+ cEGTcXev1i0R+MxM8Y2WW/LGDKKkYOoVRvA5ChhTLtX2UXnBLcRdf2lMMvEHd/nn\n+ KWEQ47ENC+uXd6UPxzE+JqVSVaVN+NNbXBJrI1ddNdEE3/++PSAmhF7BSeNWscs7\n+ w0MoPwHAGMvMHe9pas1xD3RsRFQkV01XiJqqUbf1OTdYAoUoXo9orPPrO7FMfXjZ\n+ RbzwzFtdKRlAFnKZOVf95MKlSo8WzhffKf7pQmuabGSLqSSXzIuCpxuPlNy7kwCX\n+ j5m8U1xGN7L2vlalKEG27rCLx/n6ctXAaKmQo3FM+cHim3ko/mOy+9GDwGIgToX3\n+ 5SQPnmCSR19H3nYscT06ff5lgWfBzSQmBdv//rjYkk2ZeLnTMqDNXsgT7ac6LJlj\n+ WXAdfdK2+gvHruf7jskio29hYRb2//ti5jD3NM6LLyovo1GOVl0uJ0NYLsmjDUAJ\n+ dqqNzBocy/eV3L2Ky1L6DvtcQ1otmyvroqsL5JxziP0/gRTj/t170GC/aTxjUnhs\n+ 7vDebVOT5nffxFsZwmolzTIeOsvM4rAnMu5Gf4Mna/SsMi9w/oeXFFc/b1We1a0C\n+ AwEAAaOCASUwggEhMAsGA1UdDwQEAwIBBjAdBgNVHQ4EFgQUOCSvjXUS/Dg/N4MQ\n+ r5A8/BshWv8wHwYDVR0jBBgwFoAUg8dB/Q4mTynBuHmOhnrhv7XXagMwSwYDVR0f\n+ BEQwQjBAoD6gPIY6aHR0cDovL2NkcC5wa2kuY28uc2FwLmNvbS9jZHAvU0FQJTIw\n+ R2xvYmFsJTIwUm9vdCUyMENBLmNybDBWBggrBgEFBQcBAQRKMEgwRgYIKwYBBQUH\n+ MAKGOmh0dHA6Ly9haWEucGtpLmNvLnNhcC5jb20vYWlhL1NBUCUyMEdsb2JhbCUy\n+ MFJvb3QlMjBDQS5jcnQwGQYJKwYBBAGCNxQCBAweCgBTAHUAYgBDAEEwEgYDVR0T\n+ AQH/BAgwBgEB/wIBADANBgkqhkiG9w0BAQsFAAOCAgEAGdBNALO509FQxcPhMCwE\n+ /eymAe9f2u6hXq0hMlQAuuRbpnxr0+57lcw/1eVFsT4slceh7+CHGCTCVHK1ELAd\n+ XQeibeQovsVx80BkugEG9PstCJpHnOAoWGjlZS2uWz89Y4O9nla+L9SCuK7tWI5Y\n+ +QuVhyGCD6FDIUCMlVADOLQV8Ffcm458q5S6eGViVa8Y7PNpvMyFfuUTLcUIhrZv\n+ eh4yjPSpz5uvQs7p/BJLXilEf3VsyXX5Q4ssibTS2aH2z7uF8gghfMvbLi7sS7oj\n+ XBEylxyaegwOBLtlmcbII8PoUAEAGJzdZ4kFCYjqZBMgXK9754LMpvkXDTVzy4OP\n+ emK5Il+t+B0VOV73T4yLamXG73qqt8QZndJ3ii7NGutv4SWhVYQ4s7MfjRwbFYlB\n+ z/N5eH3veBx9lJbV6uXHuNX3liGS8pNVNKPycfwlaGEbD2qZE0aZRU8OetuH1kVp\n+ jGqvWloPjj45iCGSCbG7FcY1gPVTEAreLjyINVH0pPve1HXcrnCV4PALT6HvoZoF\n+ bCuBKVgkSSoGgmasxjjjVIfMiOhkevDya52E5m0WnM1LD3ZoZzavsDSYguBP6MOV\n+ ViWNsVHocptphbEgdwvt3B75CDN4kf6MNZg2/t8bRhEQyK1FRy8NMeBnbRFnnEPe\n+ 7HJNBB1ZTjnrxJAgCQgNBIQ=\n+ -----END CERTIFICATE-----\n- path: /var/lib/iptables/rules-save\nfilesystem: root\nmode: 0644\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
adds special self-signed root cert
|
596,240 |
13.12.2017 17:24:47
| -3,600 |
f5b1deaca381da09a70a0422f0ae2332a464cf16
|
fixes iptables-restore error
|
[
{
"change_type": "MODIFY",
"old_path": "pkg/templates/node.go",
"new_path": "pkg/templates/node.go",
"diff": "@@ -224,8 +224,7 @@ storage:\n- path: /var/lib/iptables/rules-save\nfilesystem: root\nmode: 0644\n- contents:\n- inline: |-\n+ contents: |\n*nat\n:PREROUTING ACCEPT [0:0]\n:INPUT ACCEPT [0:0]\n@@ -235,7 +234,6 @@ storage:\n-A POSTROUTING -p udp ! -d {{ .ClusterCIDR }} -m addrtype ! --dst-type LOCAL -j MASQUERADE --to-ports 32000-65000\n-A POSTROUTING -p icmp ! -d {{ .ClusterCIDR }} -m addrtype ! --dst-type LOCAL -j MASQUERADE\nCOMMIT\n-\n- path: /etc/sysctl.d/10-enable-icmp-redirects\nfilesystem: root\nmode: 0644\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
fixes iptables-restore error
|
596,240 |
13.12.2017 18:23:03
| -3,600 |
a6abecb26a642cd2d99024a51292ba2b04992238
|
fixes inlines
|
[
{
"change_type": "MODIFY",
"old_path": "pkg/templates/node.go",
"new_path": "pkg/templates/node.go",
"diff": "@@ -184,7 +184,8 @@ storage:\nfiles:\n- path: /etc/ssl/certs/SAPNetCA_G2.pem\nmode: 0644\n- contents: |\n+ contents:\n+ inline: |-\n-----BEGIN CERTIFICATE-----\nMIIGPTCCBCWgAwIBAgIKYQ4GNwAAAAAADDANBgkqhkiG9w0BAQsFADBOMQswCQYD\nVQQGEwJERTERMA8GA1UEBwwIV2FsbGRvcmYxDzANBgNVBAoMBlNBUCBBRzEbMBkG\n@@ -224,7 +225,8 @@ storage:\n- path: /var/lib/iptables/rules-save\nfilesystem: root\nmode: 0644\n- contents: |\n+ contents:\n+ inline: |\n*nat\n:PREROUTING ACCEPT [0:0]\n:INPUT ACCEPT [0:0]\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
fixes inlines
|
596,240 |
14.12.2017 15:27:56
| -3,600 |
e5f2d7ecb4319492909a29b06bd6e02eaa245405
|
fixes local docs development
|
[
{
"change_type": "MODIFY",
"old_path": "Makefile",
"new_path": "Makefile",
"diff": "@@ -63,6 +63,11 @@ push:\ndocker push sapcc/kubernikus-kubectl:$(VERSION)\ndocker push sapcc/kubernikus-kubectl:latest\n+documentation:\n+ docker build $(BUILD_ARGS) -t sapcc/kubernikus-docs-builder:$(VERSION) --cache-from=sapcc/kubernikus-docs-builder:latest ./contrib/kubernikus-docs-builder\n+ docker build $(BUILD_ARGS) -t sapcc/kubernikus-docs:$(VERSION) -f Dockerfile.kubernikus-docs .\n+ docker tag sapcc/kubernikus-docs:$(VERSION) sapcc/kubernikus-docs:latest\n+\ngh-pages:\ndocker run --name gh-pages sapcc/kubernikus-docs:$(VERSION) /bin/true\ndocker cp gh-pages:/public/kubernikus gh-pages\n"
},
{
"change_type": "MODIFY",
"old_path": "docs/development/changing_docs.md",
"new_path": "docs/development/changing_docs.md",
"diff": "@@ -14,6 +14,9 @@ It constists of 2 parts:\nA live preview for development can be started with:\n```\n+make documentation\n+...\n+\ndocker run --rm -ti -p 1313:1313 \\\n-v $PWD/contrib/kubernikus-docs-builder/data:/live \\\n-v $PWD/docs/:/live/content \\\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
fixes local docs development
|
596,240 |
19.12.2017 14:22:39
| -3,600 |
857118af135aa46be3c87959ac1add2268eda124
|
fixes panic due to missing logger initialization
|
[
{
"change_type": "MODIFY",
"old_path": "pkg/wormhole/server.go",
"new_path": "pkg/wormhole/server.go",
"diff": "@@ -37,6 +37,9 @@ type Server struct {\nfunc NewServer(options *ServerOptions) (*Server, error) {\ns := &Server{Logger: log.With(options.Logger, \"wormhole\", \"server\")}\n+ if options.TunnelOptions.Logger == nil {\n+ options.TunnelOptions.Logger = options.Logger\n+ }\nclient, err := kube.NewClient(options.KubeConfig, options.Context, options.Logger)\nif err != nil {\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
fixes panic due to missing logger initialization
|
596,233 |
21.12.2017 00:41:07
| 28,800 |
4ed63464123c7df56da763b741fd1226e2a106db
|
fix default backend and use it only for 502,503,504
|
[
{
"change_type": "MODIFY",
"old_path": "charts/kubernikus-system/charts/nginx-ingress/templates/controller-configmap.yaml",
"new_path": "charts/kubernikus-system/charts/nginx-ingress/templates/controller-configmap.yaml",
"diff": "@@ -10,7 +10,7 @@ metadata:\nname: {{ template \"controller.fullname\" . }}\ndata:\nenable-vts-status: \"{{ .Values.controller.stats.enabled }}\"\n- custom-http-errors: 404,500\n+ custom-http-errors: {{.Values.defaultBackend.custom_http_errors}}\n{{- if .Values.controller.config }}\n{{ toYaml .Values.controller.config | indent 2 }}\n{{- end }}\n"
},
{
"change_type": "MODIFY",
"old_path": "charts/kubernikus-system/charts/nginx-ingress/values.yaml",
"new_path": "charts/kubernikus-system/charts/nginx-ingress/values.yaml",
"diff": "@@ -143,9 +143,11 @@ defaultBackend:\nname: default-backend\nimage:\nrepository: sapcc/custombackend\n- tag: \"0.1\"\n+ tag: \"0.3\"\npullPolicy: IfNotPresent\n+ custom_http_errors: 502,503,504\n+\nextraArgs: {}\n## Node tolerations for server scheduling to nodes with taints\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
fix default backend and use it only for 502,503,504 (#134)
|
596,240 |
22.12.2017 10:29:19
| -3,600 |
375b3c07cacd17909162e429d503c28e18a10a43
|
Kubernetes v1.9.0 Support
This adds support for Kubernetes v1.9.0. No major surprises so far. The
usual nginx smoke test and exposing via LB works.
The Openstack cloud-provider now only supports Cinder v3. The expected
major pain didn't actually kick in. It just works. TM. Famous last
words.
|
[
{
"change_type": "MODIFY",
"old_path": "charts/kube-master/Chart.yaml",
"new_path": "charts/kube-master/Chart.yaml",
"diff": "apiVersion: v1\ndescription: A Helm chart for Kubernetes\nname: kube-master\n-version: 1.8.5-kubernikus.0\n+version: 1.9.0-kubernikus.0\n"
},
{
"change_type": "MODIFY",
"old_path": "charts/kube-master/charts/etcd/values.yaml",
"new_path": "charts/kube-master/charts/etcd/values.yaml",
"diff": "# Declare variables to be passed into your templates.\nimage:\nrepository: gcr.io/google_containers/etcd\n- tag: 3.0.17\n+ tag: 3.1.10\npullPolicy: IfNotPresent\n## Persist data to a persitent volume\npersistence:\n"
},
{
"change_type": "MODIFY",
"old_path": "charts/kube-master/values.yaml",
"new_path": "charts/kube-master/values.yaml",
"diff": "# Declare variables to be passed into your templates.\nimage:\nrepository: quay.io/coreos/hyperkube\n- tag: v1.8.5_coreos.0\n+ tag: v1.9.0_coreos.0\npullPolicy: IfNotPresent\n# Settings for the openstack cloudprovider\n@@ -33,7 +33,7 @@ advertiseAddress: 198.18.128.1\nversion:\n# kubernikus:\n- kubernetes: 1.8.5\n+ kubernetes: 1.9.0\napi:\nreplicaCount: 1\n"
},
{
"change_type": "MODIFY",
"old_path": "pkg/controller/ground/bootstrap/dns/dns.go",
"new_path": "pkg/controller/ground/bootstrap/dns/dns.go",
"diff": "@@ -21,7 +21,7 @@ const (\nSERVICE_ACCOUNT = \"kube-dns\"\nCONFIGMAP = \"kube-dns\"\nDEFAULT_REPOSITORY = \"gcr.io/google_containers\"\n- DEFAULT_VERSION = \"1.14.5\"\n+ DEFAULT_VERSION = \"1.14.7\"\n)\nvar (\n"
},
{
"change_type": "MODIFY",
"old_path": "pkg/templates/ignition.go",
"new_path": "pkg/templates/ignition.go",
"diff": "@@ -39,6 +39,8 @@ var Ignition = &ignition{\nfunc (i *ignition) getIgnitionTemplate(kluster *kubernikusv1.Kluster) string {\nswitch {\n+ case strings.HasPrefix(kluster.Spec.Version, \"1.9\"):\n+ return Node_1_9\ncase strings.HasPrefix(kluster.Spec.Version, \"1.8\"):\nreturn Node_1_8\ndefault:\n"
},
{
"change_type": "ADD",
"old_path": null,
"new_path": "pkg/templates/node_1.9.go",
"diff": "+/* vim: set filetype=yaml : */\n+\n+package templates\n+\n+var Node_1_9 = `\n+passwd:\n+ users:\n+ - name: core\n+ password_hash: xyTGJkB462ewk\n+ ssh_authorized_keys:\n+ - \"ssh-rsa AAAAB3NzaC1yc2EAAAABIwAAAQEAvFapuevZeHFpFn438XMjvEQYd0wt7+tzUdAkMiSd007Tx1h79Xm9ZziDDUe4W6meinVOq93MAS/ER27hoVWGo2H/vn/Cz5M8xr2j5rQODnrF3RmfrJTbZAWaDN0JTq2lFjmCHhZJNhr+VQP1uw4z2ofMBP6MLybnLmm9ukzxFYZqCCyfEEUTCMA9SWywtTpGQp8VLM4INCxzBSCuyt3SO6PBvJSo4HoKg/sLvmRwpCVZth48PI0EUbJ72wp88Cw3bv8CLce2TOkLMwkE6NRN55w2aOyqP1G3vixHa6YcVaLlkQhJoJsBwE3rX5603y2KjOhMomqHfXxXn/3GKTWlsQ== michael.j.schmidt@gmail.com\"\n+\n+locksmith:\n+ reboot_strategy: \"reboot\"\n+\n+systemd:\n+ units:\n+ - name: iptables-restore.service\n+ enable: true\n+ - name: ccloud-metadata.service\n+ contents: |\n+ [Unit]\n+ Description=Converged Cloud Metadata Agent\n+\n+ [Service]\n+ Type=oneshot\n+ ExecStart=/usr/bin/coreos-metadata --provider=openstack-metadata --attributes=/run/metadata/coreos --ssh-keys=core --hostname=/etc/hostname\n+ - name: ccloud-metadata-hostname.service\n+ enable: true\n+ contents: |\n+ [Unit]\n+ Description=Workaround for coreos-metadata hostname bug\n+ Requires=ccloud-metadata.service\n+ After=ccloud-metadata.service\n+\n+ [Service]\n+ Type=oneshot\n+ EnvironmentFile=/run/metadata/coreos\n+ ExecStart=/usr/bin/hostnamectl set-hostname ${COREOS_OPENSTACK_HOSTNAME}\n+\n+ [Install]\n+ WantedBy=multi-user.target\n+ - name: docker.service\n+ enable: true\n+ dropins:\n+ - name: 20-docker-opts.conf\n+ contents: |\n+ [Service]\n+ Environment=\"DOCKER_OPTS=--log-opt max-size=5m --log-opt max-file=5 --ip-masq=false --iptables=false --bridge=none\"\n+ - name: kubelet.service\n+ enable: true\n+ contents: |\n+ [Unit]\n+ Description=Kubelet via Hyperkube ACI\n+\n+ [Service]\n+ Environment=\"RKT_RUN_ARGS=--uuid-file-save=/var/run/kubelet-pod.uuid \\\n+ --inherit-env \\\n+ --dns=host \\\n+ --net=host \\\n+ --volume var-lib-cni,kind=host,source=/var/lib/cni \\\n+ --volume var-log,kind=host,source=/var/log \\\n+ --mount volume=var-lib-cni,target=/var/lib/cni \\\n+ --mount volume=var-log,target=/var/log\"\n+ Environment=\"KUBELET_IMAGE_TAG=v1.9.0_coreos.0\"\n+ Environment=\"KUBELET_IMAGE_URL=quay.io/coreos/hyperkube\"\n+ ExecStartPre=/bin/mkdir -p /etc/kubernetes/manifests\n+ ExecStartPre=/bin/mkdir -p /var/lib/cni\n+ ExecStartPre=-/usr/bin/rkt rm --uuid-file=/var/run/kubelet-pod.uuid\n+ ExecStart=/usr/lib/coreos/kubelet-wrapper \\\n+ --cert-dir=/var/lib/kubelet/pki \\\n+ --cloud-config=/etc/kubernetes/openstack/openstack.config \\\n+ --cloud-provider=openstack \\\n+ --require-kubeconfig \\\n+ --bootstrap-kubeconfig=/etc/kubernetes/bootstrap/kubeconfig \\\n+ --network-plugin=kubenet \\\n+ --lock-file=/var/run/lock/kubelet.lock \\\n+ --exit-on-lock-contention \\\n+ --pod-manifest-path=/etc/kubernetes/manifests \\\n+ --allow-privileged \\\n+ --cluster-dns={{ .ClusterDNSAddress }} \\\n+ --cluster-domain={{ .ClusterDomain }} \\\n+ --client-ca-file=/etc/kubernetes/certs/kubelet-clients-ca.pem \\\n+ --non-masquerade-cidr=0.0.0.0/0 \\\n+ --anonymous-auth=false\n+ ExecStop=-/usr/bin/rkt stop --uuid-file=/var/run/kubelet-pod.uuid\n+ Restart=always\n+ RestartSec=10\n+\n+ [Install]\n+ WantedBy=multi-user.target\n+ - name: wormhole.service\n+ contents: |\n+ [Unit]\n+ Description=Kubernikus Wormhole\n+ Requires=network-online.target\n+ After=network-online.target\n+ [Service]\n+ Slice=machine.slice\n+ ExecStartPre=/usr/bin/rkt fetch --insecure-options=image --pull-policy=new docker://{{ .KubernikusImage }}:{{ .KubernikusImageTag }}\n+ ExecStart=/usr/bin/rkt run \\\n+ --inherit-env \\\n+ --net=host \\\n+ --dns=host \\\n+ --volume var-lib-kubelet,kind=host,source=/var/lib/kubelet,readOnly=true \\\n+ --mount volume=var-lib-kubelet,target=/var/lib/kubelet \\\n+ --volume etc-kubernetes-certs,kind=host,source=/etc/kubernetes/certs,readOnly=true \\\n+ --mount volume=etc-kubernetes-certs,target=/etc/kubernetes/certs \\\n+ docker://{{ .KubernikusImage }}:{{ .KubernikusImageTag }} \\\n+ --exec wormhole -- client --listen {{ .ApiserverIP }}:6443 --kubeconfig=/var/lib/kubelet/kubeconfig\n+ ExecStopPost=/usr/bin/rkt gc --mark-only\n+ KillMode=mixed\n+ Restart=always\n+ RestartSec=10s\n+ - name: wormhole.path\n+ enable: true\n+ contents: |\n+ [Path]\n+ PathExists=/var/lib/kubelet/kubeconfig\n+ [Install]\n+ WantedBy=multi-user.target\n+ - name: kube-proxy.service\n+ enable: true\n+ contents: |\n+ [Unit]\n+ Description=Kube-Proxy\n+ Requires=network-online.target\n+ After=network-online.target\n+ [Service]\n+ Slice=machine.slice\n+ ExecStart=/usr/bin/rkt run \\\n+ --trust-keys-from-https \\\n+ --inherit-env \\\n+ --net=host \\\n+ --dns=host \\\n+ --volume etc-kubernetes,kind=host,source=/etc/kubernetes,readOnly=true \\\n+ --mount volume=etc-kubernetes,target=/etc/kubernetes \\\n+ --volume lib-modules,kind=host,source=/lib/modules,readOnly=true \\\n+ --mount volume=lib-modules,target=/lib/modules \\\n+ --stage1-from-dir=stage1-fly.aci \\\n+ quay.io/coreos/hyperkube:v1.9.0_coreos.0 \\\n+ --exec=hyperkube \\\n+ -- \\\n+ proxy \\\n+ --config=/etc/kubernetes/kube-proxy/config\n+ ExecStopPost=/usr/bin/rkt gc --mark-only\n+ KillMode=mixed\n+ Restart=always\n+ RestartSec=10s\n+ [Install]\n+ WantedBy=multi-user.target\n+ - name: updatecertificates.service\n+ command: start\n+ enable: true\n+ contents: |\n+ [Unit]\n+ Description=Update the certificates w/ self-signed root CAs\n+ ConditionPathIsSymbolicLink=!/etc/ssl/certs/48b11003.0\n+ Before=early-docker.service docker.service\n+ [Service]\n+ ExecStart=/usr/sbin/update-ca-certificates\n+ RemainAfterExit=yes\n+ Type=oneshot\n+ [Install]\n+ WantedBy=multi-user.target\n+\n+networkd:\n+ units:\n+ - name: 50-kubernikus.netdev\n+ contents: |\n+ [NetDev]\n+ Description=Kubernikus Dummy Interface\n+ Name=kubernikus\n+ Kind=dummy\n+ - name: 51-kubernikus.network\n+ contents: |\n+ [Match]\n+ Name=kubernikus\n+ [Network]\n+ DHCP=no\n+ Address={{ .ApiserverIP }}/32\n+\n+storage:\n+ files:\n+ - path: /etc/ssl/certs/SAPNetCA_G2.pem\n+ filesystem: root\n+ mode: 0644\n+ contents:\n+ inline: |-\n+ -----BEGIN CERTIFICATE-----\n+ MIIGPTCCBCWgAwIBAgIKYQ4GNwAAAAAADDANBgkqhkiG9w0BAQsFADBOMQswCQYD\n+ VQQGEwJERTERMA8GA1UEBwwIV2FsbGRvcmYxDzANBgNVBAoMBlNBUCBBRzEbMBkG\n+ A1UEAwwSU0FQIEdsb2JhbCBSb290IENBMB4XDTE1MDMxNzA5MjQ1MVoXDTI1MDMx\n+ NzA5MzQ1MVowRDELMAkGA1UEBhMCREUxETAPBgNVBAcMCFdhbGxkb3JmMQwwCgYD\n+ VQQKDANTQVAxFDASBgNVBAMMC1NBUE5ldENBX0cyMIICIjANBgkqhkiG9w0BAQEF\n+ AAOCAg8AMIICCgKCAgEAjuP7Hj/1nVWfsCr8M/JX90s88IhdTLaoekrxpLNJ1W27\n+ ECUQogQF6HCu/RFD4uIoanH0oGItbmp2p8I0XVevHXnisxQGxBdkjz+a6ZyOcEVk\n+ cEGTcXev1i0R+MxM8Y2WW/LGDKKkYOoVRvA5ChhTLtX2UXnBLcRdf2lMMvEHd/nn\n+ KWEQ47ENC+uXd6UPxzE+JqVSVaVN+NNbXBJrI1ddNdEE3/++PSAmhF7BSeNWscs7\n+ w0MoPwHAGMvMHe9pas1xD3RsRFQkV01XiJqqUbf1OTdYAoUoXo9orPPrO7FMfXjZ\n+ RbzwzFtdKRlAFnKZOVf95MKlSo8WzhffKf7pQmuabGSLqSSXzIuCpxuPlNy7kwCX\n+ j5m8U1xGN7L2vlalKEG27rCLx/n6ctXAaKmQo3FM+cHim3ko/mOy+9GDwGIgToX3\n+ 5SQPnmCSR19H3nYscT06ff5lgWfBzSQmBdv//rjYkk2ZeLnTMqDNXsgT7ac6LJlj\n+ WXAdfdK2+gvHruf7jskio29hYRb2//ti5jD3NM6LLyovo1GOVl0uJ0NYLsmjDUAJ\n+ dqqNzBocy/eV3L2Ky1L6DvtcQ1otmyvroqsL5JxziP0/gRTj/t170GC/aTxjUnhs\n+ 7vDebVOT5nffxFsZwmolzTIeOsvM4rAnMu5Gf4Mna/SsMi9w/oeXFFc/b1We1a0C\n+ AwEAAaOCASUwggEhMAsGA1UdDwQEAwIBBjAdBgNVHQ4EFgQUOCSvjXUS/Dg/N4MQ\n+ r5A8/BshWv8wHwYDVR0jBBgwFoAUg8dB/Q4mTynBuHmOhnrhv7XXagMwSwYDVR0f\n+ BEQwQjBAoD6gPIY6aHR0cDovL2NkcC5wa2kuY28uc2FwLmNvbS9jZHAvU0FQJTIw\n+ R2xvYmFsJTIwUm9vdCUyMENBLmNybDBWBggrBgEFBQcBAQRKMEgwRgYIKwYBBQUH\n+ MAKGOmh0dHA6Ly9haWEucGtpLmNvLnNhcC5jb20vYWlhL1NBUCUyMEdsb2JhbCUy\n+ MFJvb3QlMjBDQS5jcnQwGQYJKwYBBAGCNxQCBAweCgBTAHUAYgBDAEEwEgYDVR0T\n+ AQH/BAgwBgEB/wIBADANBgkqhkiG9w0BAQsFAAOCAgEAGdBNALO509FQxcPhMCwE\n+ /eymAe9f2u6hXq0hMlQAuuRbpnxr0+57lcw/1eVFsT4slceh7+CHGCTCVHK1ELAd\n+ XQeibeQovsVx80BkugEG9PstCJpHnOAoWGjlZS2uWz89Y4O9nla+L9SCuK7tWI5Y\n+ +QuVhyGCD6FDIUCMlVADOLQV8Ffcm458q5S6eGViVa8Y7PNpvMyFfuUTLcUIhrZv\n+ eh4yjPSpz5uvQs7p/BJLXilEf3VsyXX5Q4ssibTS2aH2z7uF8gghfMvbLi7sS7oj\n+ XBEylxyaegwOBLtlmcbII8PoUAEAGJzdZ4kFCYjqZBMgXK9754LMpvkXDTVzy4OP\n+ emK5Il+t+B0VOV73T4yLamXG73qqt8QZndJ3ii7NGutv4SWhVYQ4s7MfjRwbFYlB\n+ z/N5eH3veBx9lJbV6uXHuNX3liGS8pNVNKPycfwlaGEbD2qZE0aZRU8OetuH1kVp\n+ jGqvWloPjj45iCGSCbG7FcY1gPVTEAreLjyINVH0pPve1HXcrnCV4PALT6HvoZoF\n+ bCuBKVgkSSoGgmasxjjjVIfMiOhkevDya52E5m0WnM1LD3ZoZzavsDSYguBP6MOV\n+ ViWNsVHocptphbEgdwvt3B75CDN4kf6MNZg2/t8bRhEQyK1FRy8NMeBnbRFnnEPe\n+ 7HJNBB1ZTjnrxJAgCQgNBIQ=\n+ -----END CERTIFICATE-----\n+ - path: /var/lib/iptables/rules-save\n+ filesystem: root\n+ mode: 0644\n+ contents:\n+ inline: |\n+ *nat\n+ :PREROUTING ACCEPT [0:0]\n+ :INPUT ACCEPT [0:0]\n+ :OUTPUT ACCEPT [0:0]\n+ :POSTROUTING ACCEPT [0:0]\n+ -A POSTROUTING -p tcp ! -d {{ .ClusterCIDR }} -m addrtype ! --dst-type LOCAL -j MASQUERADE --to-ports 32000-65000\n+ -A POSTROUTING -p udp ! -d {{ .ClusterCIDR }} -m addrtype ! --dst-type LOCAL -j MASQUERADE --to-ports 32000-65000\n+ -A POSTROUTING -p icmp ! -d {{ .ClusterCIDR }} -m addrtype ! --dst-type LOCAL -j MASQUERADE\n+ COMMIT\n+ - path: /etc/sysctl.d/10-enable-icmp-redirects\n+ filesystem: root\n+ mode: 0644\n+ contents:\n+ inline: |-\n+ net.ipv4.conf.all.accept_redirects=1\n+ - path: /etc/coreos/docker-1.12\n+ filesystem: root\n+ contents:\n+ inline: yes\n+ - path: /etc/kubernetes/certs/kubelet-clients-ca.pem\n+ filesystem: root\n+ mode: 0644\n+ contents:\n+ inline: |-\n+{{ .KubeletClientsCA | indent 10 }}\n+ - path: /etc/kubernetes/certs/apiserver-clients-system-kube-proxy-key.pem\n+ filesystem: root\n+ mode: 0644\n+ contents:\n+ inline: |-\n+{{ .ApiserverClientsSystemKubeProxyKey | indent 10 }}\n+ - path: /etc/kubernetes/certs/apiserver-clients-system-kube-proxy.pem\n+ filesystem: root\n+ mode: 0644\n+ contents:\n+ inline: |-\n+{{ .ApiserverClientsSystemKubeProxy | indent 10 }}\n+ - path: /etc/kubernetes/certs/tls-ca.pem\n+ filesystem: root\n+ mode: 0644\n+ contents:\n+ inline: |-\n+{{ .TLSCA | indent 10 }}\n+ - path: /etc/kubernetes/bootstrap/kubeconfig\n+ filesystem: root\n+ mode: 0644\n+ contents:\n+ inline: |-\n+ apiVersion: v1\n+ kind: Config\n+ clusters:\n+ - name: local\n+ cluster:\n+ certificate-authority: /etc/kubernetes/certs/tls-ca.pem\n+ server: {{ .ApiserverURL }}\n+ contexts:\n+ - name: local\n+ context:\n+ cluster: local\n+ user: local\n+ current-context: local\n+ users:\n+ - name: local\n+ user:\n+ token: {{ .BootstrapToken }}\n+ - path: /etc/kubernetes/kube-proxy/kubeconfig\n+ filesystem: root\n+ mode: 0644\n+ contents:\n+ inline: |-\n+ apiVersion: v1\n+ kind: Config\n+ clusters:\n+ - name: local\n+ cluster:\n+ certificate-authority: /etc/kubernetes/certs/tls-ca.pem\n+ server: {{ .ApiserverURL }}\n+ contexts:\n+ - name: local\n+ context:\n+ cluster: local\n+ user: local\n+ current-context: local\n+ users:\n+ - name: local\n+ user:\n+ client-certificate: /etc/kubernetes/certs/apiserver-clients-system-kube-proxy.pem\n+ client-key: /etc/kubernetes/certs/apiserver-clients-system-kube-proxy-key.pem\n+ - path: /etc/kubernetes/kube-proxy/config\n+ filesystem: root\n+ mode: 0644\n+ contents:\n+ inline: |-\n+ apiVersion: kubeproxy.config.k8s.io/v1alpha1\n+ kind: KubeProxyConfiguration\n+ bindAddress: 0.0.0.0\n+ clientConnection:\n+ acceptContentTypes: \"\"\n+ burst: 10\n+ contentType: application/vnd.kubernetes.protobuf\n+ kubeconfig: \"/etc/kubernetes/kube-proxy/kubeconfig\"\n+ qps: 5\n+ clusterCIDR: \"{{ .ClusterCIDR }}\"\n+ configSyncPeriod: 15m0s\n+ conntrack:\n+ max: 0\n+ maxPerCore: 32768\n+ min: 131072\n+ tcpCloseWaitTimeout: 1h0m0s\n+ tcpEstablishedTimeout: 24h0m0s\n+ enableProfiling: false\n+ featureGates: \"\"\n+ healthzBindAddress: 0.0.0.0:10256\n+ hostnameOverride: \"\"\n+ iptables:\n+ masqueradeAll: false\n+ masqueradeBit: 14\n+ minSyncPeriod: 0s\n+ syncPeriod: 30s\n+ metricsBindAddress: 127.0.0.1:10249\n+ mode: \"\"\n+ oomScoreAdj: -999\n+ portRange: \"\"\n+ resourceContainer: /kube-proxy\n+ udpTimeoutMilliseconds: 250ms\n+ - path: /etc/kubernetes/openstack/openstack.config\n+ filesystem: root\n+ mode: 0644\n+ contents:\n+ inline: |-\n+ [Global]\n+ auth-url = {{ .OpenstackAuthURL }}\n+ username = {{ .OpenstackUsername }}\n+ password = {{ .OpenstackPassword }}\n+ domain-name = {{ .OpenstackDomain }}\n+ region = {{ .OpenstackRegion }}\n+\n+ [LoadBalancer]\n+ lb-version=v2\n+ subnet-id = {{ .OpenstackLBSubnetID }}\n+ create-monitor = yes\n+ monitor-delay = 1m\n+ monitor-timeout = 30s\n+ monitor-max-retries = 3\n+\n+ [BlockStorage]\n+ trust-device-path = no\n+\n+ [Route]\n+ router-id = {{ .OpenstackRouterID }}\n+`\n"
},
{
"change_type": "MODIFY",
"old_path": "pkg/util/constants.go",
"new_path": "pkg/util/constants.go",
"diff": "@@ -5,5 +5,5 @@ const (\nCA_ISSUER_KUBERNIKUS_IDENTIFIER_1 = \"Kubernikus\"\n// This is the default Kubernetes version that clusters are created in\n- DEFAULT_KUBERNETES_VERSION = \"1.8.5\"\n+ DEFAULT_KUBERNETES_VERSION = \"1.9.0\"\n)\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
Kubernetes v1.9.0 Support (#135)
This adds support for Kubernetes v1.9.0. No major surprises so far. The
usual nginx smoke test and exposing via LB works.
The Openstack cloud-provider now only supports Cinder v3. The expected
major pain didn't actually kick in. It just works. TM. Famous last
words.
|
596,240 |
22.12.2017 11:28:51
| -3,600 |
e3df9babb0c4fd229980700fcf2e2f8f14646237
|
silences keystone auth log spam using a log adapter
|
[
{
"change_type": "MODIFY",
"old_path": "glide.lock",
"new_path": "glide.lock",
"diff": "@@ -50,7 +50,7 @@ imports:\nsubpackages:\n- group\n- name: github.com/databus23/keystone\n- version: 12c566d59fdb198f5a6d7ad7dfbf99f2a7e09929\n+ version: f642ecf9fb5eacb10ed79d7e7fc17e5933a296ed\nsubpackages:\n- cache/memory\n- name: github.com/databus23/requestutil\n"
},
{
"change_type": "MODIFY",
"old_path": "pkg/api/auth/keystone_auth.go",
"new_path": "pkg/api/auth/keystone_auth.go",
"diff": "@@ -7,6 +7,7 @@ import (\n\"github.com/databus23/keystone\"\n\"github.com/databus23/keystone/cache/memory\"\n+ \"github.com/go-kit/kit/log\"\nerrors \"github.com/go-openapi/errors\"\nflag \"github.com/spf13/pflag\"\n@@ -19,12 +20,15 @@ func init() {\nflag.StringVar(&authURL, \"auth-url\", \"\", \"Openstack identity v3 auth url\")\n}\n-func Keystone() func(token string) (*models.Principal, error) {\n+func Keystone(logger log.Logger) func(token string) (*models.Principal, error) {\nif !(strings.HasSuffix(authURL, \"/v3\") || strings.HasSuffix(authURL, \"/v3/\")) {\nauthURL = fmt.Sprintf(\"%s/%s\", strings.TrimRight(authURL, \"/\"), \"/v3\")\n}\n+ keystone.Log = func(format string, a ...interface{}) {\n+ logger.Log(\"library\", \"keystone\", \"msg\", fmt.Sprintf(format, a...))\n+ }\nauth := keystone.New(authURL)\nauth.TokenCache = memory.New(10 * time.Minute)\n"
},
{
"change_type": "MODIFY",
"old_path": "pkg/api/rest/configure.go",
"new_path": "pkg/api/rest/configure.go",
"diff": "@@ -35,7 +35,7 @@ func Configure(api *operations.KubernikusAPI, rt *apipkg.Runtime) error {\napi.JSONProducer = runtime.JSONProducer()\n// Applies when the \"x-auth-token\" header is set\n- api.KeystoneAuth = auth.Keystone()\n+ api.KeystoneAuth = auth.Keystone(rt.Logger)\n// Set your custom authorizer if needed. Default one is security.Authorized()\nrules, err := auth.LoadPolicy(auth.DefaultPolicyFile)\n"
},
{
"change_type": "MODIFY",
"old_path": "vendor/github.com/databus23/keystone/middleware.go",
"new_path": "vendor/github.com/databus23/keystone/middleware.go",
"diff": "@@ -12,11 +12,16 @@ import (\n\"encoding/json\"\n\"errors\"\n\"fmt\"\n+ \"log\"\n\"net/http\"\n\"strings\"\n\"time\"\n)\n+var Log func(string, ...interface{}) = func(format string, a ...interface{}) {\n+ log.Printf(format, a...)\n+}\n+\n// Cache provides the interface for cache implementations.\ntype Cache interface {\n//Set stores a value with the given ttl\n@@ -62,7 +67,7 @@ func (a *Auth) Validate(authToken string) (*Token, error) {\nif a.TokenCache != nil {\nvar cachedToken Token\nif ok := a.TokenCache.Get(authToken, &cachedToken); ok && cachedToken.Valid() {\n- fmt.Println(\"Found valid token in cache\")\n+ Log(\"Found valid token in cache\")\nreturn &cachedToken, nil\n}\n}\n@@ -150,7 +155,7 @@ func (h *handler) ServeHTTP(w http.ResponseWriter, req *http.Request) {\ncontext, err := h.Auth.Validate(authToken)\nif err != nil {\n//ToDo: How to handle logging, printing to stdout isn't the best thing\n- fmt.Println(\"Failed to validate token. \", err)\n+ Log(\"Failed to validate token: %v\", err)\nreturn\n}\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
silences keystone auth log spam using a log adapter
|
596,246 |
23.12.2017 15:28:42
| 25,200 |
913d2540e96cf7b972559385fb859288a436232d
|
Correct Test Option to use correct name.
Replace duplicate IsTestNetwork with proper IsTestVolume
|
[
{
"change_type": "MODIFY",
"old_path": "test/e2e/main.go",
"new_path": "test/e2e/main.go",
"diff": "@@ -20,7 +20,7 @@ func init() {\npflag.StringVar(&options.ConfigFile, \"configFile\", \"test/e2e/e2e_config.yaml\", \"Path to configuration file\")\npflag.BoolVar(&options.IsTestCreate, \"create\", false, \"Create a new cluster\")\npflag.BoolVar(&options.IsTestNetwork, \"network\", false, \"Run network tests\")\n- pflag.BoolVar(&options.IsTestNetwork, \"volume\", false, \"Run volume tests\")\n+ pflag.BoolVar(&options.IsTestVolume, \"volume\", false, \"Run volume tests\")\npflag.BoolVar(&options.IsTestDelete, \"delete\", false, \"Delete the cluster\")\npflag.BoolVar(&options.IsTestAll, \"all\", false, \"The whole show. Test everything\")\npflag.BoolVar(&options.IsTestAPI, \"api\", false, \"Test API\")\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
Correct Test Option to use correct name.
Replace duplicate IsTestNetwork with proper IsTestVolume
|
596,232 |
10.01.2018 15:56:51
| -3,600 |
2abfdb30e0773e9cc57ca95dd72148af97514d9a
|
add preseed RBAC RoleBinding kubernikus member
|
[
{
"change_type": "MODIFY",
"old_path": "pkg/controller/ground/bootstrap.go",
"new_path": "pkg/controller/ground/bootstrap.go",
"diff": "@@ -24,6 +24,9 @@ func SeedKluster(client clientset.Interface, kluster *v1.Kluster) error {\nif err := SeedKubernikusAdmin(client); err != nil {\nreturn err\n}\n+ if err := SeedKubernikusMember(client); err != nil {\n+ return err\n+ }\nif err := SeedCinderStorageClass(client); err != nil {\nreturn err\n}\n@@ -77,6 +80,26 @@ func SeedKubernikusAdmin(client clientset.Interface) error {\n})\n}\n+func SeedKubernikusMember(client clientset.Interface) error {\n+ return CreateOrUpdateRoleBinding(client, &rbac.RoleBinding{\n+ ObjectMeta: metav1.ObjectMeta{\n+ Name: \"kubernikus:member\",\n+ Namespace: \"default\",\n+ },\n+ RoleRef: rbac.RoleRef{\n+ APIGroup: rbac.GroupName,\n+ Kind: \"Role\",\n+ Name: \"edit\",\n+ },\n+ Subjects: []rbac.Subject{\n+ {\n+ Kind: rbac.GroupKind,\n+ Name: \"os:kubernetes_member\",\n+ },\n+ },\n+ })\n+}\n+\nfunc SeedAllowBootstrapTokensToPostCSRs(client clientset.Interface) error {\nreturn CreateOrUpdateClusterRoleBinding(client, &rbac.ClusterRoleBinding{\nObjectMeta: metav1.ObjectMeta{\n@@ -140,6 +163,19 @@ func CreateOrUpdateClusterRoleBinding(client clientset.Interface, clusterRoleBin\nreturn nil\n}\n+func CreateOrUpdateRoleBinding(client clientset.Interface, roleBinding *rbac.RoleBinding) error {\n+ if _, err := client.RbacV1beta1().RoleBindings(roleBinding.Namespace).Create(roleBinding); err != nil {\n+ if !apierrors.IsAlreadyExists(err) {\n+ return fmt.Errorf(\"unable to create RBAC rolebinding: %v\", err)\n+ }\n+\n+ if _, err := client.RbacV1beta1().RoleBindings(roleBinding.Namespace).Update(roleBinding); err != nil {\n+ return fmt.Errorf(\"unable to update RBAC rolebinding: %v\", err)\n+ }\n+ }\n+ return nil\n+}\n+\nfunc CreateOrUpdateClusterRole(client clientset.Interface, clusterRole *rbac.ClusterRole) error {\nif _, err := client.RbacV1beta1().ClusterRoles().Create(clusterRole); err != nil {\nif !apierrors.IsAlreadyExists(err) {\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
add preseed RBAC RoleBinding kubernikus member (#138)
|
596,240 |
10.01.2018 17:26:12
| -3,600 |
c06c3f388d02cb5b990e809cebef3ba73da1f8c3
|
Adds Public-Key to GetMetadata Call
|
[
{
"change_type": "MODIFY",
"old_path": "pkg/api/models/openstack_metadata.go",
"new_path": "pkg/api/models/openstack_metadata.go",
"diff": "@@ -220,6 +220,9 @@ type KeyPair struct {\n// name\nName string `json:\"name,omitempty\"`\n+\n+ // public key\n+ PublicKey string `json:\"publicKey,omitempty\"`\n}\n// Validate validates this key pair\n"
},
{
"change_type": "MODIFY",
"old_path": "pkg/api/spec/embedded_spec.go",
"new_path": "pkg/api/spec/embedded_spec.go",
"diff": "@@ -573,6 +573,9 @@ func init() {\n\"properties\": {\n\"name\": {\n\"type\": \"string\"\n+ },\n+ \"publicKey\": {\n+ \"type\": \"string\"\n}\n},\n\"x-go-name\": \"KeyPair\"\n"
},
{
"change_type": "MODIFY",
"old_path": "pkg/client/openstack/scoped/client.go",
"new_path": "pkg/client/openstack/scoped/client.go",
"diff": "@@ -201,7 +201,7 @@ func (c *client) getKeyPairs() ([]*models.KeyPair, error) {\n}\nfor _, key := range keyList {\n- result = append(result, &models.KeyPair{Name: key.Name})\n+ result = append(result, &models.KeyPair{Name: key.Name, PublicKey: key.PublicKey})\n}\nreturn result, nil\n"
},
{
"change_type": "MODIFY",
"old_path": "swagger.yml",
"new_path": "swagger.yml",
"diff": "@@ -235,6 +235,9 @@ definitions:\nproperties:\nname:\ntype: string\n+ publicKey:\n+ type: string\n+\nrouters:\ntype: array\nitems:\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
Adds Public-Key to GetMetadata Call
|
596,240 |
12.01.2018 15:07:35
| -3,600 |
d00cc7398dbde714fb7b2f9933a8f6ecc7364cbc
|
fixes stateful id logging
|
[
{
"change_type": "MODIFY",
"old_path": "pkg/util/log/middleware.go",
"new_path": "pkg/util/log/middleware.go",
"diff": "@@ -33,12 +33,14 @@ func LoggingHandler(logger kitlog.Logger, next http.Handler) http.Handler {\nreturn http.HandlerFunc(func(rw http.ResponseWriter, request *http.Request) {\nwrapper := makeWrapper(rw)\n+ inner_logger := kitlog.With(ingress_logger)\n+\nid := \"\"\nif reqId := request.Context().Value(KubernikusRequestID); reqId != nil {\nid = fmt.Sprintf(\"%s\", reqId)\n- logger = kitlog.With(logger, \"id\", id)\n+ inner_logger = kitlog.With(inner_logger, \"id\", id)\n}\n- request = request.WithContext(context.WithValue(request.Context(), \"logger\", logger))\n+ request = request.WithContext(context.WithValue(request.Context(), \"logger\", inner_logger))\ndefer func(begin time.Time) {\nvar keyvals = make([]interface{}, 0, 4)\n@@ -53,7 +55,7 @@ func LoggingHandler(logger kitlog.Logger, next http.Handler) http.Handler {\nkeyvals = append(keyvals, \"id\", id)\n}\n- log(ingress_logger, request, keyvals...)\n+ log(inner_logger, request, keyvals...)\n}(time.Now())\nnext.ServeHTTP(wrapper, request)\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
fixes stateful id logging
|
596,240 |
12.01.2018 16:18:20
| -3,600 |
3517d21d96274a4ccdae41cc22105e19ec5bce1d
|
removes dublicate id
|
[
{
"change_type": "MODIFY",
"old_path": "pkg/util/log/middleware.go",
"new_path": "pkg/util/log/middleware.go",
"diff": "@@ -51,10 +51,6 @@ func LoggingHandler(logger kitlog.Logger, next http.Handler) http.Handler {\n\"took\", time.Since(begin),\n)\n- if id != \"\" {\n- keyvals = append(keyvals, \"id\", id)\n- }\n-\nlog(inner_logger, request, keyvals...)\n}(time.Now())\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
removes dublicate id
|
596,240 |
08.01.2018 15:07:38
| -3,600 |
d1b75854ebf39913e3360cdbf9d578cf763d7fdf
|
makes GITHUB_TOKEN configurable
|
[
{
"change_type": "MODIFY",
"old_path": "Makefile",
"new_path": "Makefile",
"diff": "@@ -64,9 +64,13 @@ push:\ndocker push sapcc/kubernikus-kubectl:latest\nCHANGELOG.md:\n+ifndef GITHUB_TOKEN\n+ $(error you need to set KUBERNIKUS_TOKEN to a personal access token that has repo:read permission)\n+else\ndocker build $(BUILD_ARGS) -t sapcc/kubernikus-changelog-builder:$(VERSION) --cache-from=sapcc/kubernikus-changelog-builder:latest ./contrib/kubernikus-changelog-builder\ndocker tag sapcc/kubernikus-changelog-builder:$(VERSION) sapcc/kubernikus-changelog-builder:latest\n- docker run --name changelog -v $(PWD):/host sapcc/kubernikus-changelog-builder:latest\n+ docker run --rm -v $(PWD):/host -e GITHUB_TOKEN=$(GITHUB_TOKEN) sapcc/kubernikus-changelog-builder:latest\n+endif\ndocumentation:\ndocker build $(BUILD_ARGS) -t sapcc/kubernikus-docs-builder:$(VERSION) --cache-from=sapcc/kubernikus-docs-builder:latest ./contrib/kubernikus-docs-builder\n"
},
{
"change_type": "MODIFY",
"old_path": "contrib/kubernikus-changelog-builder/.github_changelog_generator",
"new_path": "contrib/kubernikus-changelog-builder/.github_changelog_generator",
"diff": "project=kubernikus\nuser=sapcc\nexclude_tags_regex=v1.0.0+.+\n-token=4ff4031cc4a09be557b020870e2111a6fe65c2d0\noutput=/host/CHANGELOG.md\n"
},
{
"change_type": "MODIFY",
"old_path": "contrib/kubernikus-changelog-builder/Dockerfile",
"new_path": "contrib/kubernikus-changelog-builder/Dockerfile",
"diff": "@@ -8,4 +8,4 @@ RUN gem install github_changelog_generator --version $GITHUB_CHANGELOG_GENERATOR\nCOPY .github_changelog_generator /\n-ENTRYPOINT [\"github_changelog_generator\"]\n+CMD github_changelog_generator --token $GITHUB_TOKEN\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
makes GITHUB_TOKEN configurable
|
596,240 |
08.01.2018 15:09:03
| -3,600 |
c448c673f93743983919e71d2e2e7a93c35e3e51
|
fixes wrong ENV variable in error message
|
[
{
"change_type": "MODIFY",
"old_path": "Makefile",
"new_path": "Makefile",
"diff": "@@ -65,7 +65,7 @@ push:\nCHANGELOG.md:\nifndef GITHUB_TOKEN\n- $(error you need to set KUBERNIKUS_TOKEN to a personal access token that has repo:read permission)\n+ $(error set GITHUB_TOKEN to a personal access token that has repo:read permission)\nelse\ndocker build $(BUILD_ARGS) -t sapcc/kubernikus-changelog-builder:$(VERSION) --cache-from=sapcc/kubernikus-changelog-builder:latest ./contrib/kubernikus-changelog-builder\ndocker tag sapcc/kubernikus-changelog-builder:$(VERSION) sapcc/kubernikus-changelog-builder:latest\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
fixes wrong ENV variable in error message
|
596,240 |
15.01.2018 16:11:03
| -3,600 |
9e36b4dbe19e4b7ef13c5e6eb57bdbd0a5b2cef3
|
adds at least 4 eyes as code owner
|
[
{
"change_type": "MODIFY",
"old_path": ".github/CODEOWNERS",
"new_path": ".github/CODEOWNERS",
"diff": "* @bugroger @databus23 @edda @auhlig @schwarzm @jonathanschwarze @vonneudeck\n/charts @databus23 @bugroger @vonneudeck @auhlig\n-/charts/k8sniff @auhlig\n+/charts/k8sniff @auhlig @databus23\n/charts/kubernikus-system @auhlig @bugroger\n/charts/kube-master @databus23 @bugroger\n/charts/kubernikus @databus23 @bugroger\n-/ci @databus23\n-/contrib @bugroger\n-/docs @bugroger\n-/pkg/api @databus23\n-ground @databus23\n-launch @bugroger\n-/pkg/templates @schwarzm\n-wormhole @databus23\n-/pkg/cmd/kubernikusctl @schwarzm\n-/pkg/cmd/kubernikusctl/auth @bugroger\n-swagger @databus23\n+/ci @databus23 @bugroger\n+/contrib @bugroger @databus23\n+/docs @bugroger @jonathanschwarze\n+/pkg @databus23 @bugroger\n+/pkg/templates @schwarzm @bugroger\n+/pkg/cmd/kubernikusctl @schwarzm @bugroger\n+/pkg/cmd/kubernikusctl/auth @bugroger @databus23\n+swagger @databus23 @bugroger\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
adds at least 4 eyes as code owner
|
596,233 |
18.01.2018 13:12:49
| -3,600 |
d2750c29eb55299ddecb09ca452ed7913abda682
|
scrape kubelets, cadvisor
scrapes cadvisor metrics from kubelets
|
[
{
"change_type": "MODIFY",
"old_path": "charts/kubernikus-system/charts/prometheus/templates/_prometheus.yaml.tpl",
"new_path": "charts/kubernikus-system/charts/prometheus/templates/_prometheus.yaml.tpl",
"diff": "@@ -202,8 +202,46 @@ scrape_configs:\naction: replace\nreplacement: ${1}\n+- job_name: 'kube-system/kubelet'\n+ kubernetes_sd_configs:\n+ - role: node\n+ relabel_configs:\n+ - action: labelmap\n+ regex: __meta_kubernetes_node_label_(.+)\n+ - target_label: component\n+ replacement: kubelet\n+ - action: replace\n+ source_labels: [__meta_kubernetes_node_name]\n+ target_label: instance\n+ - source_labels: [__address__]\n+ action: replace\n+ target_label: __address__\n+ regex: ([^:;]+):(\\d+)\n+ replacement: ${1}:10255\n+ - source_labels: [__scheme__]\n+ action: replace\n+ target_label: __scheme__\n+ regex: https\n+ replacement: http\n+\n+- job_name: 'kubernetes-cadvisors'\n+ scheme: https\n+ tls_config:\n+ ca_file: /var/run/secrets/kubernetes.io/serviceaccount/ca.crt\n+ bearer_token_file: /var/run/secrets/kubernetes.io/serviceaccount/token\n+ kubernetes_sd_configs:\n+ - role: node\n+ relabel_configs:\n+ - action: labelmap\n+ regex: __meta_kubernetes_node_label_(.+)\n+ - target_label: __address__\n+ replacement: kubernetes.default:443\n+ - source_labels: [__meta_kubernetes_node_name]\n+ regex: (.+)\n+ target_label: __metrics_path__\n+ replacement: /api/v1/nodes/${1}:4194/proxy/metrics\n+\n# Static Targets\n-#\n- job_name: 'kubernikus-prometheus'\nmetrics_path: /prometheus/metrics\nstatic_configs:\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
scrape kubelets, cadvisor (#180)
scrapes cadvisor metrics from kubelets
|
596,232 |
18.01.2018 13:51:40
| -3,600 |
26b5b6fee6d0574ea42aec63bf4ccc3dce7d25a1
|
Documentation files (new)
updates documentation
|
[
{
"change_type": "MODIFY",
"old_path": "README.md",
"new_path": "README.md",
"diff": "@@ -46,8 +46,8 @@ More documentation can be found at:\n## Contact\n-https://convergedcloud.slack.com\n+https://convergedcloud.slack.com/app_redirect?channel=kubernikus-users\n## License\n-This project is licensed under the Apache2 License - see the [LICENSE](blob/master/LICENSE) file for details\n+This project is licensed under the Apache2 License - see the [LICENSE](LICENSE) file for details\n"
},
{
"change_type": "ADD",
"old_path": "assets/images/docs/containers/kubernetes/credentials.png",
"new_path": "assets/images/docs/containers/kubernetes/credentials.png",
"diff": "Binary files /dev/null and b/assets/images/docs/containers/kubernetes/credentials.png differ\n"
},
{
"change_type": "ADD",
"old_path": "assets/images/docs/containers/kubernetes/loadbalancer0.png",
"new_path": "assets/images/docs/containers/kubernetes/loadbalancer0.png",
"diff": "Binary files /dev/null and b/assets/images/docs/containers/kubernetes/loadbalancer0.png differ\n"
},
{
"change_type": "ADD",
"old_path": "assets/images/docs/containers/kubernetes/loadbalancer1.png",
"new_path": "assets/images/docs/containers/kubernetes/loadbalancer1.png",
"diff": "Binary files /dev/null and b/assets/images/docs/containers/kubernetes/loadbalancer1.png differ\n"
},
{
"change_type": "ADD",
"old_path": "assets/images/docs/containers/kubernetes/loadbalancer2.png",
"new_path": "assets/images/docs/containers/kubernetes/loadbalancer2.png",
"diff": "Binary files /dev/null and b/assets/images/docs/containers/kubernetes/loadbalancer2.png differ\n"
},
{
"change_type": "ADD",
"old_path": "assets/images/docs/containers/kubernetes/roleassignment.png",
"new_path": "assets/images/docs/containers/kubernetes/roleassignment.png",
"diff": "Binary files /dev/null and b/assets/images/docs/containers/kubernetes/roleassignment.png differ\n"
},
{
"change_type": "ADD",
"old_path": "assets/images/docs/containers/kubernetes/setup.png",
"new_path": "assets/images/docs/containers/kubernetes/setup.png",
"diff": "Binary files /dev/null and b/assets/images/docs/containers/kubernetes/setup.png differ\n"
},
{
"change_type": "ADD",
"old_path": "assets/images/docs/containers/kubernetes/userroleassignments.png",
"new_path": "assets/images/docs/containers/kubernetes/userroleassignments.png",
"diff": "Binary files /dev/null and b/assets/images/docs/containers/kubernetes/userroleassignments.png differ\n"
},
{
"change_type": "MODIFY",
"old_path": "docs/guide/authentication.md",
"new_path": "docs/guide/authentication.md",
"diff": "title: Authentication\n---\n-## Authentication\n+## Intro\n+\n+In Kubernetes regular end-users are assumed to be managed by an outside,\n+independent service. In this regard, Kubernetes does not have objects which\n+represent normal user accounts. Regular users cannot be added to a cluster\n+through an API call.\n+\n+API requests are tied to either a normal user or a service account, or are\n+treated as anonymous requests. This means every process inside or outside the\n+cluster, from a human user typing kubectl on a workstation, to kubelets on\n+nodes, to members of the control plane, must authenticate when making requests\n+to the API server, or be treated as an anonymous user.\n+\n+### User-Management\n+\n+For Kubernikus clusters the user management is handled by OpenStack's Identity\n+Service (Keystone). Only users that have been given a `os:kubernikus_admin` or\n+`os:kubernikus_member` roles by an Keystone administrator are allowed to interact\n+with the service or clusters.\n+\n+### Authentication\n+\n+The authentication against Kubernikus clusters is based on x509 certificates.\n+Encoded into the certificate's `Common Name` field is the user name. The\n+certificate's `organization` fields indicate the user's OpenStack role\n+assignments.\n+\n+This effectively maps OpenStack roles to Kubernetes groups.\n+\n+These certificates are generated. They can be retrieved via UI or API. In order\n+to allow for revocation of authorization the certificates are short lived. They\n+automatically expire after 24h. Therefore they need to be periodically\n+refreshed.\n+\n+### Authorizations\n+\n+Using the `user` and `groups` provided by the authentication mechanism it is\n+then possible to use [Kubernetes\n+RBAC](https://kubernetes.io/docs/admin/authorization/rbac/) to define\n+authorizations within Kubernetes.\n+\n+By distinguishing between users as well as `kubernikus_admin` and `kubernikus_member`\n+roles/groups it is possible to assign different Kubernetes roles to groups or\n+individual users.\n+\n+## Manage Roles in OpenStack\n+\n+Users with `Keystone Administrator` role are allowed to change user role\n+assignments in a project.\n+\n+\n+\n+To add additional users to a cluster they need to be\n+given either `Kubernetes Admin` or `Kubernetes Member` roles.\n+\n+\n+/assets/images/docs/containers/kubernetes/userroleassignments.png\n+\n+## Authenticating with Kubernetes\n+\n+Kubernetes is usually remote controlled with the `kubectl` command line tool.\n+It is configured through a config file `.kubeconfig`. For installation\n+instructions please see the [official\n+documentation](https://kubernetes.io/docs/user-guide/kubectl-overview/).\n+\n+### Manual Download\n+A preconfigured `.kubeconfig` file can be downloaded from the UI or fetched via\n+API:\n+\n+\n+\n+### Automatic Refresh\n+\n+Since the certificates expire daily it becomes quite tedious to download new\n+`.kubeconfig` files every day. To help with this workflow there is a CLI tool\n+`kubernikusctl` for remote controlling Kubernikus clusters.\n+\n+The tool is available precompiled for Linux, MacOS and Windows. The latest\n+version can be downloaded from [Github](https://github.com/sapcc/kubernikus/releases/latest).\n+\n+Setting up an automatic refresh of the `.kubeconfig` file is a 2-step process:\n+\n+ 1. `kubernikusctl auth init --many --auth --options`\n+ 2. `kubernikusctl auth refresh`\n+\n+The initialisation only needs to be done once. Afterwards a `refresh` is\n+possible without repeating all authentication details.\n+\n+\n+\n+The UI provides the full `kubernikusctl auth init` initialisation command for\n+convenience.\n+\n+### Default Permissions\n+\n+By default any user with the `Kubernetes Admin` OpenStack role is assigned the\n+`cluster-admin` Kubernetes role. This is a super-admin that is allowed\n+everything.\n+\n+Otherwise, the default Kubernetes RBAC policies grant very restrictive\n+permissions. Users with the `Kubernetes Member` OpenStack role need to be\n+assigned further permissions.\n+\n+For example, to grant cluster-wide, full access:\n+\n+```\n+kubectl create clusterrolebinding cluster-admin\n+ --clusterrole=cluster-admin\n+ --group=kubernetes_member\n+```\n+\n+~> Note: This allows to perform any action against the API, including viewing secrets and modifying permissions. It is not a recommended policy.\n"
},
{
"change_type": "MODIFY",
"old_path": "docs/guide/best_practices.md",
"new_path": "docs/guide/best_practices.md",
"diff": "@@ -8,4 +8,4 @@ title: Best Practices\n## Automating Authentication Refresh\n-## Integration for CI Systems\n+## Integrationn for CI Systems\n"
},
{
"change_type": "MODIFY",
"old_path": "docs/guide/common_addons.md",
"new_path": "docs/guide/common_addons.md",
"diff": "@@ -3,3 +3,72 @@ title: Common Addons\n---\n## Kubernetes Dashboard\n+\n+[Kubernetes Dashboard](https://github.com/kubernetes/dashboard) is a general\n+purpose, web-based UI for Kubernetes clusters. It allows users to manage\n+applications running in the cluster and troubleshoot them, as well as manage\n+the cluster itself.\n+\n+\n+### Installation\n+\n+[Installation](https://github.com/kubernetes/dashboard) is straight forward:\n+\n+```\n+kubectl apply -f https://raw.githubusercontent.com/kubernetes/dashboard/master/src/deploy/recommended/kubernetes-dashboard.yaml\n+```\n+\n+### Granting Permissions\n+\n+You can grant full admin privileges to Dashboard's Service Account by creating\n+below `ClusterRoleBinding`.\n+\n+```\n+kubectl create clusterrolebinding kubernetes-dashboard --clusterrole=cluster-admin --serviceaccount=kube-system:kubernetes-dashboard\n+```\n+\n+### Accessing the Dashboard\n+\n+To access Dashboard from your local workstation you must create a secure\n+channel to your Kubernetes cluster. Run the following command:\n+\n+```\n+kubectl proxy\n+```\n+\n+Now access Dashboard at:\n+\n+[http://localhost:8001/api/v1/namespaces/kube-system/services/https:kubernetes-dashboard:/proxy/.](http://localhost:8001/api/v1/namespaces/kube-system/services/https:kubernetes-dashboard:/proxy/.)\n+\n+### Exposig the Dashboard\n+\n+In order to expose the Dashboard without the local proxy, we need to:\n+\n+ * Create a service of type `LoadBalancer`\n+ * Open the security groups for load-balancer to node communication\n+ * Assign a floating IP\n+\n+Let's create the service:\n+\n+```\n+kubectl expose deployment kubernetes-dashboard --namespace kube-system --type=LoadBalancer --name kubernete-dashboard-external --port=443\n+```\n+\n+This will create a Kubernetes service that exposes the dashboard on\n+a high-level service port on each node of the cluster. Additionally,\n+a load-balancer is created in OpenStack which points to each node.\n+\n+\n+\n+As the load-balancers are not in the default instance group, they need to be\n+added to the security group explicitly. Without this the health monitors won't\n+be able to reach the node port and the listener will be offline.\n+\n+\n+\n+Once the health monitors turn healthy, attaching a floating IP will make the\n+dashboard accessible from the outside via `https` on port `443`.\n+\n+\n+\n+~> This setup exposes a unauthenticated Dashboard with full access to the cluster. This is a security risk. See the [Access Control](https://github.com/kubernetes/dashboard/wiki/Access-control) documentation for more info.\n"
},
{
"change_type": "MODIFY",
"old_path": "docs/guide/getting_started.md",
"new_path": "docs/guide/getting_started.md",
"diff": "@@ -4,3 +4,53 @@ weight: -10\n---\n## Getting Started\n+\n+[Kubernetes](https://kubernetes.io/) is an open-source system for automating\n+deployment, scaling, and management of containerized applications.\n+\n+\"Kubernetes as a Service\" (Codename: Kubernikus) makes it easy to run Kubernetes clusters that are natively integrated with OpenStack. It is a managed service that takes care of installing, upgrading and operating the cluster.\n+\n+It provides an easy entry to deploy containerized payloads and getting started\n+with Kubernetes without the operational overhead of setting up Kubernetes.\n+Due to the tight and convenient integration with OpenStack it becomes easy to\n+combine VM and cloud-native workloads.\n+\n+### Key Features\n+\n+ * Masters are managed centrally\n+ * Workload nodes are located in customer's projects\n+ * Combine VM and containerised payloads\n+\n+### Enhanced Security\n+\n+ * Air-Gapped masters and nodes\n+ * Full TLS encryption between all components\n+ * Unified authorization policy between OpenStack and [Kubernetes RBAC](http://blog.kubernetes.io/2017/04/rbac-support-in-kubernetes.html)\n+ * Authentication tooling\n+ * Auto-Updating nodes based on [CoreOS Container Linux](https://coreos.com/why/)\n+\n+### Compliance\n+\n+ * 100% Vanilla Kubernetes\n+ * 100% Compatible OpenStack API\n+\n+## Demo\n+\n+[](http://www.youtube.com/watch?v=1dPxPU9fHTg)\n+\n+### Support\n+\n+To allow for direct, convenient feedback and support please join the\n+[#kubernikus-users](https://convergedcloud.slack.com/messages/kubernikus-users)\n+channel in the [Converged Cloud Slack](https://convergedcloud.slack.com)\n+workspace. Any SAP employee is allowed to sign up and access this workspace\n+using the SAP email address.\n+\n+There's also an open weekly meeting for all users and everyone interested in\n+Kubernikus. Next dates are being announced and pinned in\n+[#kubernikus-users](https://convergedcloud.slack.com/messages/kubernikus-users).\n+\n+There you will also find the\n+[#kubernetes](https://convergedcloud.slack.com/messages/kubernikus-users)\n+channel for general topics related to Kubernetes in SAP and specifially on\n+Converged Cloud.\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
Documentation files (new) (#181)
updates documentation
|
596,240 |
18.01.2018 14:02:03
| -3,600 |
d9a1a99b970edf4b23ef4ac4d6e15e0e67e605ab
|
Uses DeepCopy Instead of Backend Call
This refactors away a few backend calls that are unneeded now that we
have the generated DeepCopy helpers.
It helps us to save some traffic against the control plane kubernetes
API.
|
[
{
"change_type": "MODIFY",
"old_path": "pkg/controller/ground.go",
"new_path": "pkg/controller/ground.go",
"diff": "@@ -328,12 +328,9 @@ func (op *GroundControl) klusterUpdate(cur, old interface{}) {\n}\nfunc (op *GroundControl) updatePhase(kluster *v1.Kluster, phase models.KlusterPhase, message string) error {\n-\n//Never modify the cache, at least that's what I've been told\n- kluster, err := op.Clients.Kubernikus.Kubernikus().Klusters(kluster.Namespace).Get(kluster.Name, metav1.GetOptions{})\n- if err != nil {\n- return err\n- }\n+ kluster = kluster.DeepCopy()\n+\n//Do nothing is the phase is not changing\nif kluster.Status.Phase == phase {\nreturn nil\n@@ -342,7 +339,7 @@ func (op *GroundControl) updatePhase(kluster *v1.Kluster, phase models.KlusterPh\nkluster.Status.Message = message\nkluster.Status.Phase = phase\n- _, err = op.Clients.Kubernikus.Kubernikus().Klusters(kluster.Namespace).Update(kluster)\n+ _, err := op.Clients.Kubernikus.Kubernikus().Klusters(kluster.Namespace).Update(kluster)\nif err == nil {\n//Wait for up to 5 seconds for the local cache to reflect the phase change\nwaitutil.WaitForKluster(kluster, op.klusterInformer.GetIndexer(), func(k *v1.Kluster) (bool, error) {\n@@ -499,10 +496,7 @@ func (op *GroundControl) discoverKubernikusInfo(kluster *v1.Kluster) error {\n\"project\", kluster.Account(),\n\"v\", 5)\n- copy, err := op.Clients.Kubernikus.Kubernikus().Klusters(kluster.Namespace).Get(kluster.Name, metav1.GetOptions{})\n- if err != nil {\n- return err\n- }\n+ copy := kluster.DeepCopy()\nif copy.Status.Apiserver == \"\" {\ncopy.Status.Apiserver = fmt.Sprintf(\"https://%s.%s\", kluster.GetName(), op.Config.Kubernikus.Domain)\n@@ -522,7 +516,7 @@ func (op *GroundControl) discoverKubernikusInfo(kluster *v1.Kluster) error {\n\"project\", kluster.Account())\n}\n- _, err = op.Clients.Kubernikus.Kubernikus().Klusters(kluster.Namespace).Update(copy)\n+ _, err := op.Clients.Kubernikus.Kubernikus().Klusters(kluster.Namespace).Update(copy)\nreturn err\n}\n@@ -538,10 +532,7 @@ func (op *GroundControl) discoverOpenstackInfo(kluster *v1.Kluster) error {\nreturn err\n}\n- copy, err := op.Clients.Kubernikus.Kubernikus().Klusters(kluster.Namespace).Get(kluster.Name, metav1.GetOptions{})\n- if err != nil {\n- return err\n- }\n+ copy := kluster.DeepCopy()\nif copy.Spec.Openstack.ProjectID == \"\" {\ncopy.Spec.Openstack.ProjectID = kluster.Account()\n"
},
{
"change_type": "MODIFY",
"old_path": "pkg/controller/launch/pool_manager.go",
"new_path": "pkg/controller/launch/pool_manager.go",
"diff": "@@ -82,10 +82,7 @@ func (cpm *ConcretePoolManager) SetStatus(status *PoolStatus) error {\nSchedulable: int64(status.Running),\n}\n- copy, err := cpm.Clients.Kubernikus.Kubernikus().Klusters(cpm.Kluster.Namespace).Get(cpm.Kluster.Name, metav1.GetOptions{})\n- if err != nil {\n- return err\n- }\n+ copy := cpm.Kluster.DeepCopy()\nfor i, curInfo := range copy.Status.NodePools {\nif curInfo.Name == newInfo.Name {\n@@ -94,7 +91,7 @@ func (cpm *ConcretePoolManager) SetStatus(status *PoolStatus) error {\n}\ncopy.Status.NodePools[i] = newInfo\n- _, err = cpm.Clients.Kubernikus.Kubernikus().Klusters(copy.Namespace).Update(copy)\n+ _, err := cpm.Clients.Kubernikus.Kubernikus().Klusters(copy.Namespace).Update(copy)\nreturn err\n}\n}\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
Uses DeepCopy Instead of Backend Call (#176)
This refactors away a few backend calls that are unneeded now that we
have the generated DeepCopy helpers.
It helps us to save some traffic against the control plane kubernetes
API.
|
596,240 |
18.01.2018 14:29:28
| -3,600 |
3f8ce84a0775dab29ab9a180fd3a83f611038cb8
|
updates for v1.5.0
|
[
{
"change_type": "MODIFY",
"old_path": "CHANGELOG.md",
"new_path": "CHANGELOG.md",
"diff": "## [Unreleased](https://github.com/sapcc/kubernikus/tree/HEAD)\n-[Full Changelog](https://github.com/sapcc/kubernikus/compare/v1.4.0...HEAD)\n+[Full Changelog](https://github.com/sapcc/kubernikus/compare/rm...HEAD)\n**Implemented enhancements:**\n-- apiserver: Improve logs [\\#2](https://github.com/sapcc/kubernikus/issues/2)\n+- Allows to Specify an SSH Public Key for Login to Nodes [\\#82](https://github.com/sapcc/kubernikus/issues/82)\n+\n+**Fixed bugs:**\n+\n+- Allows to Specify an SSH Public Key for Login to Nodes [\\#82](https://github.com/sapcc/kubernikus/issues/82)\n+\n+## [rm](https://github.com/sapcc/kubernikus/tree/rm) (2018-01-18)\n+[Full Changelog](https://github.com/sapcc/kubernikus/compare/v1.5.0...rm)\n+\n+## [v1.5.0](https://github.com/sapcc/kubernikus/tree/v1.5.0) (2018-01-18)\n+[Full Changelog](https://github.com/sapcc/kubernikus/compare/v1.4.0...v1.5.0)\n+\n+**Implemented enhancements:**\n+\n+- Add Advanced Options to Web UI [\\#147](https://github.com/sapcc/kubernikus/issues/147)\n+- Metadata: Sort and filter flavors [\\#130](https://github.com/sapcc/kubernikus/issues/130)\n+- Preseed RBAC RoleBindings [\\#122](https://github.com/sapcc/kubernikus/issues/122)\n+- Don't offer flavours that don't work with our default image [\\#109](https://github.com/sapcc/kubernikus/issues/109)\n+- Run Conformance Tests on a Kluster [\\#66](https://github.com/sapcc/kubernikus/issues/66)\n+- Adds Automatic CHANGELOG.md Generation [\\#145](https://github.com/sapcc/kubernikus/pull/145) ([BugRoger](https://github.com/BugRoger))\n+- add preseed RBAC RoleBinding kubernikus member [\\#138](https://github.com/sapcc/kubernikus/pull/138) ([jonathanschwarze](https://github.com/jonathanschwarze))\n+\n+**Fixed bugs:**\n+\n+- Non-Default Security Group is Ignored [\\#152](https://github.com/sapcc/kubernikus/issues/152)\n+- Doc links broken using blob syntax [\\#140](https://github.com/sapcc/kubernikus/issues/140)\n**Closed issues:**\n+- Validate that the securitygroup id exists [\\#172](https://github.com/sapcc/kubernikus/issues/172)\n+- Migrate to FilteredSharedInformer [\\#169](https://github.com/sapcc/kubernikus/issues/169)\n+- Add Public-Key from Spec to Ignition Templates [\\#158](https://github.com/sapcc/kubernikus/issues/158)\n+- Add Public-Key to Kluster Spec [\\#157](https://github.com/sapcc/kubernikus/issues/157)\n+- Add Public-Key to GetMetadata API [\\#156](https://github.com/sapcc/kubernikus/issues/156)\n+- Upgrade Code-Generators [\\#154](https://github.com/sapcc/kubernikus/issues/154)\n- Discover all missing attributes via operator [\\#85](https://github.com/sapcc/kubernikus/issues/85)\n+- apiserver: Improve logs [\\#2](https://github.com/sapcc/kubernikus/issues/2)\n**Merged pull requests:**\n+- Documentation files \\(new\\) [\\#181](https://github.com/sapcc/kubernikus/pull/181) ([jonathanschwarze](https://github.com/jonathanschwarze))\n+- scrape kubelets, cadvisor [\\#180](https://github.com/sapcc/kubernikus/pull/180) ([auhlig](https://github.com/auhlig))\n+- remove one-time securitygroup name migration [\\#179](https://github.com/sapcc/kubernikus/pull/179) ([databus23](https://github.com/databus23))\n+- Uses DeepCopy Instead of Backend Call [\\#176](https://github.com/sapcc/kubernikus/pull/176) ([BugRoger](https://github.com/BugRoger))\n+- Use filtered informers scoped to given namespace [\\#171](https://github.com/sapcc/kubernikus/pull/171) ([databus23](https://github.com/databus23))\n+- Add user defined ssh public key to ignition config [\\#168](https://github.com/sapcc/kubernikus/pull/168) ([databus23](https://github.com/databus23))\n+- Specify security group by name [\\#167](https://github.com/sapcc/kubernikus/pull/167) ([databus23](https://github.com/databus23))\n+- Upgrades to client-go v6.0.0 [\\#163](https://github.com/sapcc/kubernikus/pull/163) ([BugRoger](https://github.com/BugRoger))\n+- Generate a random ssh password per kluster [\\#162](https://github.com/sapcc/kubernikus/pull/162) ([databus23](https://github.com/databus23))\n+- Add ssh public key to Kluster spec [\\#161](https://github.com/sapcc/kubernikus/pull/161) ([databus23](https://github.com/databus23))\n+- Surface nova errors message for bad request responses \\(400\\) [\\#160](https://github.com/sapcc/kubernikus/pull/160) ([databus23](https://github.com/databus23))\n+- Adds Public-Key to GetMetadata Call [\\#159](https://github.com/sapcc/kubernikus/pull/159) ([BugRoger](https://github.com/BugRoger))\n+- Use Finalizers for Orchestrating Kluster Deletion [\\#155](https://github.com/sapcc/kubernikus/pull/155) ([BugRoger](https://github.com/BugRoger))\n- Relax stalebot [\\#143](https://github.com/sapcc/kubernikus/pull/143) ([databus23](https://github.com/databus23))\n- Correct Test Flag Name [\\#142](https://github.com/sapcc/kubernikus/pull/142) ([notque](https://github.com/notque))\n- Typo fixes. [\\#141](https://github.com/sapcc/kubernikus/pull/141) ([notque](https://github.com/notque))\n**Implemented enhancements:**\n- Upgrade to 1.8 [\\#62](https://github.com/sapcc/kubernikus/issues/62)\n+- Sort \\(and filter\\) flavors in metadata response [\\#137](https://github.com/sapcc/kubernikus/pull/137) ([databus23](https://github.com/databus23))\n- Kubernetes v1.9.0 Support [\\#135](https://github.com/sapcc/kubernikus/pull/135) ([BugRoger](https://github.com/BugRoger))\n**Fixed bugs:**\n**Implemented enhancements:**\n- Setup CoreDNS for kubernikus-system [\\#70](https://github.com/sapcc/kubernikus/issues/70)\n-- Log/Tracing Utility [\\#24](https://github.com/sapcc/kubernikus/issues/24)\n-- Configurable Defaults [\\#21](https://github.com/sapcc/kubernikus/issues/21)\n**Fixed bugs:**\n- Sane Infrastructure Setup [\\#54](https://github.com/sapcc/kubernikus/issues/54)\n- Github Workflow [\\#50](https://github.com/sapcc/kubernikus/issues/50)\n- RKT Pods for Kubelet + Wormhole Client [\\#44](https://github.com/sapcc/kubernikus/issues/44)\n+- Log/Tracing Utility [\\#24](https://github.com/sapcc/kubernikus/issues/24)\n+- Configurable Defaults [\\#21](https://github.com/sapcc/kubernikus/issues/21)\n- Cleanup and enhance spec [\\#3](https://github.com/sapcc/kubernikus/issues/3)\n**Merged pull requests:**\n- Docker Options Dropin [\\#64](https://github.com/sapcc/kubernikus/issues/64)\n- Add Kube-Proxy to Nodes [\\#38](https://github.com/sapcc/kubernikus/issues/38)\n- Seed ClusterRoleBindings [\\#35](https://github.com/sapcc/kubernikus/issues/35)\n-- Add Kube-Proxy to Nodes [\\#34](https://github.com/sapcc/kubernikus/issues/34)\n- Expose NodePool CRUD via API [\\#31](https://github.com/sapcc/kubernikus/issues/31)\n**Fixed bugs:**\n- Improve NodeAPI [\\#49](https://github.com/sapcc/kubernikus/issues/49)\n- Remove Dependency OpenstackSeeder [\\#48](https://github.com/sapcc/kubernikus/issues/48)\n- Kube-Proxy br\\_netfilter Missing [\\#42](https://github.com/sapcc/kubernikus/issues/42)\n+- Add Kube-Proxy to Nodes [\\#34](https://github.com/sapcc/kubernikus/issues/34)\n- Cluster-State Aware LaunchController [\\#25](https://github.com/sapcc/kubernikus/issues/25)\n- Kluster persistence [\\#18](https://github.com/sapcc/kubernikus/issues/18)\n- Implement cluster edit [\\#17](https://github.com/sapcc/kubernikus/issues/17)\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
updates for v1.5.0
|
596,240 |
18.01.2018 14:40:38
| -3,600 |
2a512be0c0a152612b6f447f08b5339fcbdfaae2
|
updates for v1.5.0+
|
[
{
"change_type": "MODIFY",
"old_path": "CHANGELOG.md",
"new_path": "CHANGELOG.md",
"diff": "# Change Log\n-## [Unreleased](https://github.com/sapcc/kubernikus/tree/HEAD)\n-\n-[Full Changelog](https://github.com/sapcc/kubernikus/compare/rm...HEAD)\n-\n-**Implemented enhancements:**\n-\n-- Allows to Specify an SSH Public Key for Login to Nodes [\\#82](https://github.com/sapcc/kubernikus/issues/82)\n-\n-**Fixed bugs:**\n-\n-- Allows to Specify an SSH Public Key for Login to Nodes [\\#82](https://github.com/sapcc/kubernikus/issues/82)\n-\n-## [rm](https://github.com/sapcc/kubernikus/tree/rm) (2018-01-18)\n-[Full Changelog](https://github.com/sapcc/kubernikus/compare/v1.5.0...rm)\n-\n## [v1.5.0](https://github.com/sapcc/kubernikus/tree/v1.5.0) (2018-01-18)\n[Full Changelog](https://github.com/sapcc/kubernikus/compare/v1.4.0...v1.5.0)\n- Metadata: Sort and filter flavors [\\#130](https://github.com/sapcc/kubernikus/issues/130)\n- Preseed RBAC RoleBindings [\\#122](https://github.com/sapcc/kubernikus/issues/122)\n- Don't offer flavours that don't work with our default image [\\#109](https://github.com/sapcc/kubernikus/issues/109)\n+- Allows to Specify an SSH Public Key for Login to Nodes [\\#82](https://github.com/sapcc/kubernikus/issues/82)\n- Run Conformance Tests on a Kluster [\\#66](https://github.com/sapcc/kubernikus/issues/66)\n- Adds Automatic CHANGELOG.md Generation [\\#145](https://github.com/sapcc/kubernikus/pull/145) ([BugRoger](https://github.com/BugRoger))\n- add preseed RBAC RoleBinding kubernikus member [\\#138](https://github.com/sapcc/kubernikus/pull/138) ([jonathanschwarze](https://github.com/jonathanschwarze))\n- Non-Default Security Group is Ignored [\\#152](https://github.com/sapcc/kubernikus/issues/152)\n- Doc links broken using blob syntax [\\#140](https://github.com/sapcc/kubernikus/issues/140)\n+- Allows to Specify an SSH Public Key for Login to Nodes [\\#82](https://github.com/sapcc/kubernikus/issues/82)\n**Closed issues:**\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
updates for v1.5.0+
|
596,246 |
18.01.2018 08:33:22
| 25,200 |
3d210e89747c31ae5a9744a7b88986afc0e04d18
|
remove internal slack link from readme
remove internal slack link, github issue tracker is fine for contact.
|
[
{
"change_type": "MODIFY",
"old_path": "README.md",
"new_path": "README.md",
"diff": "@@ -44,10 +44,6 @@ More documentation can be found at:\n* [Kubernikus Docs](./docs/)\n-## Contact\n-\n-https://convergedcloud.slack.com/app_redirect?channel=kubernikus-users\n-\n## License\nThis project is licensed under the Apache2 License - see the [LICENSE](LICENSE) file for details\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
remove internal slack link from readme (#185)
remove internal slack link, github issue tracker is fine for contact.
|
596,240 |
18.01.2018 17:24:10
| -3,600 |
98f5fae1e200a0e31672401fce6fb9e46232f34a
|
reverts deep-copy change
|
[
{
"change_type": "MODIFY",
"old_path": "pkg/controller/ground.go",
"new_path": "pkg/controller/ground.go",
"diff": "@@ -329,7 +329,10 @@ func (op *GroundControl) klusterUpdate(cur, old interface{}) {\nfunc (op *GroundControl) updatePhase(kluster *v1.Kluster, phase models.KlusterPhase, message string) error {\n//Never modify the cache, at least that's what I've been told\n- kluster = kluster.DeepCopy()\n+ kluster, err := op.Clients.Kubernikus.Kubernikus().Klusters(kluster.Namespace).Get(kluster.Name, metav1.GetOptions{})\n+ if err != nil {\n+ return err\n+ }\n//Do nothing is the phase is not changing\nif kluster.Status.Phase == phase {\n@@ -339,7 +342,7 @@ func (op *GroundControl) updatePhase(kluster *v1.Kluster, phase models.KlusterPh\nkluster.Status.Message = message\nkluster.Status.Phase = phase\n- _, err := op.Clients.Kubernikus.Kubernikus().Klusters(kluster.Namespace).Update(kluster)\n+ _, err = op.Clients.Kubernikus.Kubernikus().Klusters(kluster.Namespace).Update(kluster)\nif err == nil {\n//Wait for up to 5 seconds for the local cache to reflect the phase change\nwaitutil.WaitForKluster(kluster, op.klusterInformer.GetIndexer(), func(k *v1.Kluster) (bool, error) {\n@@ -496,7 +499,10 @@ func (op *GroundControl) discoverKubernikusInfo(kluster *v1.Kluster) error {\n\"project\", kluster.Account(),\n\"v\", 5)\n- copy := kluster.DeepCopy()\n+ copy, err := op.Clients.Kubernikus.Kubernikus().Klusters(kluster.Namespace).Get(kluster.Name, metav1.GetOptions{})\n+ if err != nil {\n+ return err\n+ }\nif copy.Status.Apiserver == \"\" {\ncopy.Status.Apiserver = fmt.Sprintf(\"https://%s.%s\", kluster.GetName(), op.Config.Kubernikus.Domain)\n@@ -516,7 +522,7 @@ func (op *GroundControl) discoverKubernikusInfo(kluster *v1.Kluster) error {\n\"project\", kluster.Account())\n}\n- _, err := op.Clients.Kubernikus.Kubernikus().Klusters(kluster.Namespace).Update(copy)\n+ _, err = op.Clients.Kubernikus.Kubernikus().Klusters(kluster.Namespace).Update(copy)\nreturn err\n}\n"
},
{
"change_type": "MODIFY",
"old_path": "pkg/controller/launch/pool_manager.go",
"new_path": "pkg/controller/launch/pool_manager.go",
"diff": "@@ -82,7 +82,10 @@ func (cpm *ConcretePoolManager) SetStatus(status *PoolStatus) error {\nSchedulable: int64(status.Running),\n}\n- copy := cpm.Kluster.DeepCopy()\n+ copy, err := cpm.Clients.Kubernikus.Kubernikus().Klusters(cpm.Kluster.Namespace).Get(cpm.Kluster.Name, metav1.GetOptions{})\n+ if err != nil {\n+ return err\n+ }\nfor i, curInfo := range copy.Status.NodePools {\nif curInfo.Name == newInfo.Name {\n@@ -91,7 +94,7 @@ func (cpm *ConcretePoolManager) SetStatus(status *PoolStatus) error {\n}\ncopy.Status.NodePools[i] = newInfo\n- _, err := cpm.Clients.Kubernikus.Kubernikus().Klusters(copy.Namespace).Update(copy)\n+ _, err = cpm.Clients.Kubernikus.Kubernikus().Klusters(copy.Namespace).Update(copy)\nreturn err\n}\n}\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
reverts deep-copy change
|
596,232 |
18.01.2018 17:25:09
| -3,600 |
1b3b5ffd4927290f3ec40529b8dc5e5e0e9ade89
|
delete slack section
|
[
{
"change_type": "MODIFY",
"old_path": "docs/guide/getting_started.md",
"new_path": "docs/guide/getting_started.md",
"diff": "@@ -38,19 +38,3 @@ combine VM and cloud-native workloads.\n[](http://www.youtube.com/watch?v=1dPxPU9fHTg)\n-### Support\n-\n-To allow for direct, convenient feedback and support please join the\n-[#kubernikus-users](https://convergedcloud.slack.com/messages/kubernikus-users)\n-channel in the [Converged Cloud Slack](https://convergedcloud.slack.com)\n-workspace. Any SAP employee is allowed to sign up and access this workspace\n-using the SAP email address.\n-\n-There's also an open weekly meeting for all users and everyone interested in\n-Kubernikus. Next dates are being announced and pinned in\n-[#kubernikus-users](https://convergedcloud.slack.com/messages/kubernikus-users).\n-\n-There you will also find the\n-[#kubernetes](https://convergedcloud.slack.com/messages/kubernikus-users)\n-channel for general topics related to Kubernetes in SAP and specifially on\n-Converged Cloud.\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
delete slack section
|
596,232 |
22.01.2018 18:11:12
| -3,600 |
77eff82b12f26db05b6365417c0ca8945fb10011
|
general information for kubernikus CI system integration
|
[
{
"change_type": "MODIFY",
"old_path": "docs/guide/best_practices.md",
"new_path": "docs/guide/best_practices.md",
"diff": "@@ -8,4 +8,28 @@ title: Best Practices\n## Automating Authentication Refresh\n-## Integrationn for CI Systems\n+## Integration for CI Systems\n+\n+ * Add credentials to a build agent to communicate with a Kubernikus Kluster.\n+\n+### Set up kubectl an kubernikusctl\n+First you have to set up `kubectl` and `kubernikusctl` on your build agent.\n+Follow the [instructions](https://github.com/sapcc/kubernikus/blob/master/docs/guide/authentication.md#authenticating-with-kubernetes).\n+\n+### Create a technical user\n+To renew certs with `kubernikusctl auth` you have to store your `username` and `password` on the build agent. For privacy reasons you can create and add a technical user instead of your personal one. To create a technical user follow this [guide](https://documentation.global.cloud.sap/docs/support/specific-requests/technical-user-requests.html) (SAP only).\n+\n+### Add credentials to your environment\n+Add following variables to your environment on your build agent and fill it with your credentials:\n+\n+```\n+OS_AUTH_URL=https://identity-3.eu-nl-1.cloud.sap/v3\n+OS_USERNAME=T27F923CD2DC8D443\n+OS_PASSWORD=abcabc\n+OS_PROJECT_NAME=testproject\n+OS_PROJECT_DOMAIN_NAME=monsoon3\n+```\n+\n+### Use kubernikusctl into your build job\n+Add `kubernikusctl auth init` command to your build job. It will check your environment for the credentials and fetches certificates.\n+\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
general information for kubernikus CI system integration
|
596,240 |
02.02.2018 14:29:04
| -3,600 |
f2e74e78aa5a65d962012b84cf11464b4a579689
|
add helm repo for forked charts
|
[
{
"change_type": "MODIFY",
"old_path": "ci/task_helm_kubernikus-system.yaml",
"new_path": "ci/task_helm_kubernikus-system.yaml",
"diff": "@@ -25,6 +25,7 @@ run:\n#helm dep up --skip-refresh charts.git/charts/kubernikus-system/\n#helm upgrade kubernikus-system charts.git/charts/kubernikus-system/ --namespace=kubernikus-system --values secrets.git/$KUBERNIKUS_NAME/values/kubernikus-system.yaml --install\ncp -a charts.git charts-copy.git\n+ help repo add forked https://raw.githubusercontent.com/BugRoger/charts/repo\nhelm dep up --skip-refresh charts-copy.git/charts/kubernikus-system/\nhelm upgrade kubernikus-system charts-copy.git/charts/kubernikus-system/ --namespace=kubernikus-system --values secrets.git/$KUBERNIKUS_NAME/values/kubernikus-system.yaml --install\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
add helm repo for forked charts
|
596,240 |
06.02.2018 17:24:13
| -3,600 |
52b096db2b7077573291032a0bc8da2255a550d2
|
splits ingress specs. adds authentication via tls
|
[
{
"change_type": "MODIFY",
"old_path": "charts/kubernikus-system/charts/nginx-ingress/templates/controller-deployment.yaml",
"new_path": "charts/kubernikus-system/charts/nginx-ingress/templates/controller-deployment.yaml",
"diff": "@@ -39,7 +39,7 @@ spec:\n{{- if (contains \"0.9\" .Values.controller.image.tag) }}\n- --ingress-class={{ .Values.controller.ingressClass }}\n{{- end }}\n- {{- if (contains \"0.9\" .Values.controller.image.tag) }}\n+ {{- if (contains \"0.10\" .Values.controller.image.tag) }}\n- --configmap={{ .Release.Namespace }}/{{ template \"controller.fullname\" . }}\n{{- else }}\n- --nginx-configmap={{ .Release.Namespace }}/{{ template \"controller.fullname\" . }}\n"
},
{
"change_type": "MODIFY",
"old_path": "charts/kubernikus-system/charts/nginx-ingress/values.yaml",
"new_path": "charts/kubernikus-system/charts/nginx-ingress/values.yaml",
"diff": "@@ -5,10 +5,17 @@ controller:\nname: controller\nimage:\nrepository: sapcc/nginx-ingress-controller\n- tag: \"0.9.0-beta.17\"\n+ tag: \"0.10.2\"\npullPolicy: IfNotPresent\n- config: {}\n+ config:\n+ http-snippet: |\n+ map $ssl_client_s_dn $ssl_client_s_dn_cn {\n+ default \"anonymous\";\n+ ~CN=(?<CN>[^/,\\\"]+) $CN;\n+ }\n+ location-snippet: |\n+ proxy_set_header X-REMOTE-USER $ssl_client_s_dn_cn;\n# Required for use with CNI based kubernetes installations (such as ones set up by kubeadm),\n# since CNI and hostport don't mix yet. Can be deprecated once https://github.com/kubernetes/kubernetes/issues/23920\n"
},
{
"change_type": "MODIFY",
"old_path": "charts/kubernikus-system/charts/prometheus/templates/deployment.yaml",
"new_path": "charts/kubernikus-system/charts/prometheus/templates/deployment.yaml",
"diff": "@@ -29,8 +29,7 @@ spec:\n- --storage.tsdb.retention={{.Values.retention}}\n- --web.console.libraries=/etc/prometheus/console_libraries\n- --web.console.templates=/etc/prometheus/consoles\n- - --web.external-url=https://kubernikus.{{.Values.global.region}}.cloud.sap/prometheus\n- - --web.route-prefix=/\n+ - --web.external-url=https://prometheus.{{.Values.domain}}\n- --log.level={{ default \"info\" .Values.log_level }}\nports:\n- name: http\n"
},
{
"change_type": "ADD",
"old_path": null,
"new_path": "charts/kubernikus-system/templates/grafana-ingress.yaml",
"diff": "+apiVersion: extensions/v1beta1\n+kind: Ingress\n+\n+metadata:\n+ name: grafana\n+ annotations:\n+ vice-president: \"true\"\n+ prometheus.io/probe: \"true\"\n+ nginx.ingress.kubernetes.io/configuration-snippet: |\n+ rewrite ^/$ /dashboard/db/kubernikus?refresh=1m&orgId=1&kiosk=true redirect;\n+ {{- if .Values.authentication.enabled}}\n+ nginx.ingress.kubernetes.io/auth-tls-secret: \"kubernikus-system/ca-crt\"\n+ nginx.ingress.kubernetes.io/auth-tls-verify-depth: \"3\"\n+ nginx.ingress.kubernetes.io/auth-tls-verify-client: \"optional\"\n+ {{ end }}\n+\n+spec:\n+ tls:\n+ - secretName: grafana\n+ hosts: [grafana.{{ required \"domain missing\" .Values.domain }}]\n+ rules:\n+ - host: grafana.{{ required \"domain missing\" .Values.domain }}\n+ http:\n+ paths:\n+ - path: /\n+ backend:\n+ serviceName: kubernikus-system-grafana\n+ servicePort: 80\n"
},
{
"change_type": "ADD",
"old_path": null,
"new_path": "charts/kubernikus-system/templates/prometheus-ingress.yaml",
"diff": "+apiVersion: extensions/v1beta1\n+kind: Ingress\n+\n+metadata:\n+ name: prometheus\n+ annotations:\n+ vice-president: \"true\"\n+ prometheus.io/probe: \"true\"\n+ {{- if .Values.authentication.enabled}}\n+ nginx.ingress.kubernetes.io/auth-tls-secret: \"kubernikus-system/ca-crt\"\n+ nginx.ingress.kubernetes.io/auth-tls-verify-depth: \"3\"\n+ nginx.ingress.kubernetes.io/auth-tls-verify-client: \"on\"\n+ {{ end }}\n+\n+spec:\n+ tls:\n+ - secretName: prometheus\n+ hosts: [prometheus.{{ required \"domain missing\" .Values.domain }}]\n+ rules:\n+ - host: prometheus.{{ required \"domain missing\" .Values.domain }}\n+ http:\n+ paths:\n+ - path: /\n+ backend:\n+ serviceName: kubernikus-prometheus\n+ servicePort: 9090\n"
},
{
"change_type": "ADD",
"old_path": null,
"new_path": "charts/kubernikus-system/templates/secret.yaml",
"diff": "+{{- if .Values.authentication.enabled}}\n+apiVersion: v1\n+kind: Secret\n+type: Opaque\n+\n+metadata:\n+ name: ca-crt\n+\n+data:\n+ ca.crt: {{ default \"\" .Values.authentication.cacrt | b64enc | quote }}\n+{{- end }}\n"
},
{
"change_type": "MODIFY",
"old_path": "charts/kubernikus-system/values.yaml",
"new_path": "charts/kubernikus-system/values.yaml",
"diff": "+#authentication:\n+# enabled: true\n+# cacrt: |\n+# -----BEGIN CERTIFICATE-----\n+# MIIB+jCCAWOgAwIBAgIEAQAAADANBgkqhkiG9w0BAQUFADAvMQswCQYDVQQGEwJE\n+# ...\n+# vrEcgtBVIuzM+sVEp7RRM6Y+fL9u+69krtndZ8Ft\n+# -----END CERTIFICATE-----\n+\n+#domain: \"kubernikus.cloud.sap\"\n+\n#global:\n# region: eu-de-1\n@@ -11,8 +22,6 @@ grafana:\nimage: \"grafana/grafana:master\"\nimagePullPolicy: IfNotPresent\nname: grafana\n- adminUser: \"admin\"\n- adminPassword: \"admin\"\npersistentVolume:\nenabled: false\n@@ -21,7 +30,7 @@ grafana:\nhttpGet:\npath: /login\nport: 3000\n- initialDelaySeconds: 30\n+ initialDelaySeconds: 5\ntimeoutSeconds: 30\nservice:\n@@ -33,6 +42,79 @@ grafana:\nprovisioning:\nenabled: true\n+ provisioningDashboardsFiles:\n+ default.yaml: |\n+ - name: 'default'\n+ org_id: 1\n+ type: file\n+ options:\n+ path: /var/lib/grafana/dashboards\n+\n+ provisioningDatasourcesFiles:\n+ prometheus.yaml: |\n+ datasources:\n+ - name: Prometheus\n+ type: prometheus\n+ access: proxy\n+ org_id: 1\n+ url: http://kubernikus-prometheus:9090\n+ is_default: true\n+ version: 1\n+ editable: false\n+\n+ serverConfigFile:\n+ grafana.ini: |\n+ [paths]\n+ data = /var/lib/grafana/data\n+ logs = /var/log/grafana\n+ plugins = /var/lib/grafana/plugins\n+ provisioning = /var/lib/grafana/provisioning\n+\n+ [server]\n+ protocol = http\n+ http_addr =\n+ http_port = 3000\n+ domain = localhost\n+ enforce_domain = false\n+ root_url = %(protocol)s://%(domain)s:%(http_port)s\n+ router_logging = false\n+ static_root_path = public\n+ enable_gzip = false\n+ cert_file =\n+ cert_key =\n+ socket = /tmp/grafana.sock\n+\n+ [users]\n+ allow_sign_up = true\n+ allow_org_create = false\n+ auto_assign_org = true\n+ auto_assign_org_role = Admin\n+ default_theme = dark\n+\n+ [auth.anonymous]\n+ enabled = true\n+ org_name = Main Org.\n+ org_role = Viewer\n+\n+ [auth.proxy]\n+ enabled = true\n+ header_name = X-REMOTE-USER\n+ header_property = username\n+ auto_sign_up = true\n+\n+ [auth.basic]\n+ enabled = false\n+\n+ [smtp]\n+ enabled = false\n+\n+ [log]\n+ mode = console\n+ level = debug\n+\n+ [alerting]\n+ enabled = false\n+\nserverDashboardFiles:\nkubernikus.json: |\n{\n@@ -52,6 +134,7 @@ grafana:\n\"editable\": true,\n\"gnetId\": null,\n\"graphTooltip\": 0,\n+ \"iteration\": 1517825496293,\n\"links\": [],\n\"panels\": [\n{\n@@ -415,7 +498,6 @@ grafana:\n\"tableColumn\": \"\",\n\"targets\": [\n{\n- \"$$hashKey\": \"object:3742\",\n\"expr\": \"sum(kubernikus_kluster_status_phase{phase=\\\"Terminating\\\"})\",\n\"format\": \"time_series\",\n\"intervalFactor\": 1,\n@@ -504,7 +586,6 @@ grafana:\n\"renderer\": \"flot\",\n\"seriesOverrides\": [\n{\n- \"$$hashKey\": \"object:4033\",\n\"alias\": \"Avg\",\n\"fill\": 0,\n\"linewidth\": 3\n@@ -515,7 +596,6 @@ grafana:\n\"steppedLine\": false,\n\"targets\": [\n{\n- \"$$hashKey\": \"object:3461\",\n\"expr\": \"sum(probe_http_duration_seconds{path=\\\"/\\\"}) * 1000\",\n\"format\": \"time_series\",\n\"instant\": false,\n@@ -524,7 +604,6 @@ grafana:\n\"refId\": \"A\"\n},\n{\n- \"$$hashKey\": \"object:3922\",\n\"expr\": \"sum(avg_over_time(probe_http_duration_seconds{path=\\\"/\\\"}[1h])) * 1000\",\n\"format\": \"time_series\",\n\"intervalFactor\": 1,\n@@ -552,7 +631,6 @@ grafana:\n},\n\"yaxes\": [\n{\n- \"$$hashKey\": \"object:3486\",\n\"format\": \"ms\",\n\"label\": null,\n\"logBase\": 1,\n@@ -561,7 +639,6 @@ grafana:\n\"show\": true\n},\n{\n- \"$$hashKey\": \"object:3487\",\n\"format\": \"short\",\n\"label\": null,\n\"logBase\": 1,\n@@ -606,7 +683,6 @@ grafana:\n\"renderer\": \"flot\",\n\"seriesOverrides\": [\n{\n- \"$$hashKey\": \"object:4033\",\n\"alias\": \"Avg\",\n\"fill\": 0,\n\"linewidth\": 3\n@@ -617,7 +693,6 @@ grafana:\n\"steppedLine\": false,\n\"targets\": [\n{\n- \"$$hashKey\": \"object:3461\",\n\"expr\": \"sum(probe_http_duration_seconds{path=\\\"/\\\"}) * 1000\",\n\"format\": \"time_series\",\n\"instant\": false,\n@@ -626,7 +701,6 @@ grafana:\n\"refId\": \"A\"\n},\n{\n- \"$$hashKey\": \"object:3922\",\n\"expr\": \"sum(avg_over_time(probe_http_duration_seconds{path=\\\"/\\\"}[1h])) * 1000\",\n\"format\": \"time_series\",\n\"intervalFactor\": 1,\n@@ -654,7 +728,6 @@ grafana:\n},\n\"yaxes\": [\n{\n- \"$$hashKey\": \"object:3486\",\n\"format\": \"ms\",\n\"label\": null,\n\"logBase\": 1,\n@@ -663,7 +736,6 @@ grafana:\n\"show\": true\n},\n{\n- \"$$hashKey\": \"object:3487\",\n\"format\": \"short\",\n\"label\": null,\n\"logBase\": 1,\n@@ -677,17 +749,14 @@ grafana:\n\"backgroundColor\": \"rgb(60, 60, 60)\",\n\"colorMaps\": [\n{\n- \"$$hashKey\": \"object:6781\",\n\"color\": \"#e5ac0e\",\n\"text\": \"N/A\"\n},\n{\n- \"$$hashKey\": \"object:6782\",\n\"color\": \"#508642\",\n\"text\": \"up\"\n},\n{\n- \"$$hashKey\": \"object:6783\",\n\"color\": \"#bf1b00\",\n\"text\": \"down\"\n}\n@@ -720,7 +789,6 @@ grafana:\n\"metricNameColor\": \"#fff\",\n\"rangeMaps\": [\n{\n- \"$$hashKey\": \"object:6538\",\n\"from\": \"null\",\n\"text\": \"N/A\",\n\"to\": \"null\"\n@@ -737,7 +805,6 @@ grafana:\n\"showTransitionCount\": false,\n\"targets\": [\n{\n- \"$$hashKey\": \"object:6451\",\n\"expr\": \"probe_success{path=\\\"/\\\"}\",\n\"format\": \"time_series\",\n\"intervalFactor\": 1,\n@@ -745,7 +812,6 @@ grafana:\n\"refId\": \"A\"\n},\n{\n- \"$$hashKey\": \"object:6452\",\n\"expr\": \"probe_success{path=\\\"/grafana\\\"}\",\n\"format\": \"time_series\",\n\"intervalFactor\": 1,\n@@ -761,25 +827,21 @@ grafana:\n\"type\": \"natel-discrete-panel\",\n\"valueMaps\": [\n{\n- \"$$hashKey\": \"object:6530\",\n\"op\": \"=\",\n\"text\": \"N/A\",\n\"value\": \"null\"\n},\n{\n- \"$$hashKey\": \"object:6531\",\n\"op\": \"=\",\n\"text\": \"up\",\n\"value\": \"1\"\n},\n{\n- \"$$hashKey\": \"object:6532\",\n\"op\": \"=\",\n\"text\": \"down\",\n\"value\": \"0\"\n},\n{\n- \"$$hashKey\": \"object:6533\",\n\"op\": \"=\",\n\"text\": \"\",\n\"value\": \"\"\n@@ -833,6 +895,21 @@ grafana:\n\"transparent\": true,\n\"type\": \"text\"\n},\n+ {\n+ \"content\": \"<style>\\n table.info td {\\n border: none;\\n background: none;\\n } \\n table.info th {\\n border: none;\\n font-weight: bold;\\n background: none;\\n } \\n</style>\\n\\n<table class=\\\"info\\\">\\n<tr><th>CoreOS</th><td>$coreos_version</td></tr>\\n<tr><th>Kernel</th><td>$kernel_version</td></tr>\\n<tr><th>Kubernetes</th><td>$kubelet_version</td></tr>\\n<tr><th>Docker</th><td>$container_runtime_version</td></tr>\\n</table>\\n\",\n+ \"gridPos\": {\n+ \"h\": 6,\n+ \"w\": 2,\n+ \"x\": 3,\n+ \"y\": 6\n+ },\n+ \"id\": 67,\n+ \"links\": [],\n+ \"mode\": \"html\",\n+ \"title\": \"Control Plane\",\n+ \"transparent\": false,\n+ \"type\": \"text\"\n+ },\n{\n\"cacheTimeout\": null,\n\"colorBackground\": true,\n@@ -856,7 +933,7 @@ grafana:\n\"gridPos\": {\n\"h\": 3,\n\"w\": 2,\n- \"x\": 3,\n+ \"x\": 5,\n\"y\": 6\n},\n\"hideTimeOverride\": true,\n@@ -943,7 +1020,7 @@ grafana:\n\"gridPos\": {\n\"h\": 3,\n\"w\": 2,\n- \"x\": 5,\n+ \"x\": 7,\n\"y\": 6\n},\n\"hideTimeOverride\": true,\n@@ -1030,7 +1107,7 @@ grafana:\n\"gridPos\": {\n\"h\": 3,\n\"w\": 2,\n- \"x\": 7,\n+ \"x\": 9,\n\"y\": 6\n},\n\"hideTimeOverride\": true,\n@@ -1094,21 +1171,6 @@ grafana:\n],\n\"valueName\": \"current\"\n},\n- {\n- \"content\": \"\",\n- \"gridPos\": {\n- \"h\": 6,\n- \"w\": 2,\n- \"x\": 9,\n- \"y\": 6\n- },\n- \"id\": 52,\n- \"links\": [],\n- \"mode\": \"markdown\",\n- \"title\": \"\",\n- \"transparent\": true,\n- \"type\": \"text\"\n- },\n{\n\"content\": \"\\n\\n\\n \\n\",\n\"gridPos\": {\n@@ -1203,7 +1265,6 @@ grafana:\n\"tableColumn\": \"\",\n\"targets\": [\n{\n- \"$$hashKey\": \"object:4274\",\n\"expr\": \"limes_project_quota{project=\\\"kubernikus\\\",resource=\\\"volumes\\\",domain=\\\"ccadmin\\\"} - limes_project_usage{project=\\\"kubernikus\\\",resource=\\\"volumes\\\",domain=\\\"ccadmin\\\"}\",\n\"format\": \"time_series\",\n\"interval\": \"\",\n@@ -1291,7 +1352,6 @@ grafana:\n\"tableColumn\": \"\",\n\"targets\": [\n{\n- \"$$hashKey\": \"object:4344\",\n\"expr\": \"(limes_project_quota{project=\\\"kubernikus\\\",resource=\\\"capacity\\\",domain=\\\"ccadmin\\\",service=\\\"volumev2\\\"} - limes_project_usage{project=\\\"kubernikus\\\",resource=\\\"capacity\\\",domain=\\\"ccadmin\\\",service=\\\"volumev2\\\"}) / 10\",\n\"format\": \"time_series\",\n\"interval\": \"\",\n@@ -1300,7 +1360,6 @@ grafana:\n\"refId\": \"A\"\n},\n{\n- \"$$hashKey\": \"object:4365\",\n\"expr\": \"\",\n\"format\": \"time_series\",\n\"intervalFactor\": 1,\n@@ -1355,12 +1414,10 @@ grafana:\n\"mappingType\": 1,\n\"mappingTypes\": [\n{\n- \"$$hashKey\": \"object:7703\",\n\"name\": \"value to text\",\n\"value\": 1\n},\n{\n- \"$$hashKey\": \"object:7704\",\n\"name\": \"range to text\",\n\"value\": 2\n}\n@@ -1388,7 +1445,6 @@ grafana:\n\"tableColumn\": \"\",\n\"targets\": [\n{\n- \"$$hashKey\": \"object:4344\",\n\"expr\": \"limes_project_quota{project=\\\"kubernikus\\\",resource=\\\"instances\\\",domain=\\\"ccadmin\\\"} - limes_project_usage{project=\\\"kubernikus\\\",resource=\\\"instances\\\",domain=\\\"ccadmin\\\"}\",\n\"format\": \"time_series\",\n\"interval\": \"\",\n@@ -1405,7 +1461,6 @@ grafana:\n\"valueFontSize\": \"200%\",\n\"valueMaps\": [\n{\n- \"$$hashKey\": \"object:7706\",\n\"op\": \"=\",\n\"text\": \"N/A\",\n\"value\": \"null\"\n@@ -1446,12 +1501,10 @@ grafana:\n\"mappingType\": 1,\n\"mappingTypes\": [\n{\n- \"$$hashKey\": \"object:7703\",\n\"name\": \"value to text\",\n\"value\": 1\n},\n{\n- \"$$hashKey\": \"object:7704\",\n\"name\": \"range to text\",\n\"value\": 2\n}\n@@ -1479,7 +1532,6 @@ grafana:\n\"tableColumn\": \"\",\n\"targets\": [\n{\n- \"$$hashKey\": \"object:4344\",\n\"expr\": \"(limes_project_quota{project=\\\"kubernikus\\\",resource=\\\"cores\\\",domain=\\\"ccadmin\\\"} - limes_project_usage{project=\\\"kubernikus\\\",resource=\\\"cores\\\",domain=\\\"ccadmin\\\"}) / 8\",\n\"format\": \"time_series\",\n\"interval\": \"\",\n@@ -1496,7 +1548,6 @@ grafana:\n\"valueFontSize\": \"200%\",\n\"valueMaps\": [\n{\n- \"$$hashKey\": \"object:7706\",\n\"op\": \"=\",\n\"text\": \"N/A\",\n\"value\": \"null\"\n@@ -1537,12 +1588,10 @@ grafana:\n\"mappingType\": 1,\n\"mappingTypes\": [\n{\n- \"$$hashKey\": \"object:7703\",\n\"name\": \"value to text\",\n\"value\": 1\n},\n{\n- \"$$hashKey\": \"object:7704\",\n\"name\": \"range to text\",\n\"value\": 2\n}\n@@ -1570,7 +1619,6 @@ grafana:\n\"tableColumn\": \"\",\n\"targets\": [\n{\n- \"$$hashKey\": \"object:4344\",\n\"expr\": \"(limes_project_quota{project=\\\"kubernikus\\\",resource=\\\"ram\\\",domain=\\\"ccadmin\\\"} - limes_project_usage{project=\\\"kubernikus\\\",resource=\\\"ram\\\",domain=\\\"ccadmin\\\"}) / 16000\",\n\"format\": \"time_series\",\n\"interval\": \"\",\n@@ -1587,7 +1635,6 @@ grafana:\n\"valueFontSize\": \"200%\",\n\"valueMaps\": [\n{\n- \"$$hashKey\": \"object:7706\",\n\"op\": \"=\",\n\"text\": \"N/A\",\n\"value\": \"null\"\n@@ -1621,7 +1668,7 @@ grafana:\n\"gridPos\": {\n\"h\": 3,\n\"w\": 2,\n- \"x\": 3,\n+ \"x\": 5,\n\"y\": 9\n},\n\"id\": 41,\n@@ -1649,7 +1696,6 @@ grafana:\n\"steppedLine\": false,\n\"targets\": [\n{\n- \"$$hashKey\": \"object:260\",\n\"expr\": \"sum(irate(node_cpu{mode=~\\\"system|user|iowait\\\"}[5m])) by (instance) / sum(irate(node_cpu{}[5m])) by (instance)\",\n\"format\": \"time_series\",\n\"interval\": \"\",\n@@ -1677,7 +1723,6 @@ grafana:\n},\n\"yaxes\": [\n{\n- \"$$hashKey\": \"object:344\",\n\"decimals\": null,\n\"format\": \"percentunit\",\n\"label\": \"\",\n@@ -1687,7 +1732,6 @@ grafana:\n\"show\": false\n},\n{\n- \"$$hashKey\": \"object:345\",\n\"format\": \"short\",\n\"label\": null,\n\"logBase\": 1,\n@@ -1708,7 +1752,7 @@ grafana:\n\"gridPos\": {\n\"h\": 3,\n\"w\": 2,\n- \"x\": 5,\n+ \"x\": 7,\n\"y\": 9\n},\n\"id\": 42,\n@@ -1736,7 +1780,6 @@ grafana:\n\"steppedLine\": false,\n\"targets\": [\n{\n- \"$$hashKey\": \"object:260\",\n\"expr\": \"(sum(node_memory_MemTotal) by (instance) - sum(node_memory_MemFree+node_memory_Buffers+node_memory_Cached) by (instance)) / (sum(node_memory_MemTotal) by (instance)) * 100\",\n\"format\": \"time_series\",\n\"interval\": \"\",\n@@ -1745,7 +1788,6 @@ grafana:\n\"refId\": \"A\"\n},\n{\n- \"$$hashKey\": \"object:1694\",\n\"expr\": \"\",\n\"format\": \"time_series\",\n\"intervalFactor\": 1,\n@@ -1772,7 +1814,6 @@ grafana:\n},\n\"yaxes\": [\n{\n- \"$$hashKey\": \"object:344\",\n\"decimals\": null,\n\"format\": \"percent\",\n\"label\": \"\",\n@@ -1782,7 +1823,6 @@ grafana:\n\"show\": false\n},\n{\n- \"$$hashKey\": \"object:345\",\n\"format\": \"short\",\n\"label\": null,\n\"logBase\": 1,\n@@ -1803,7 +1843,7 @@ grafana:\n\"gridPos\": {\n\"h\": 3,\n\"w\": 2,\n- \"x\": 7,\n+ \"x\": 9,\n\"y\": 9\n},\n\"id\": 58,\n@@ -1831,7 +1871,6 @@ grafana:\n\"steppedLine\": false,\n\"targets\": [\n{\n- \"$$hashKey\": \"object:260\",\n\"expr\": \"sum(label_replace(kubelet_running_pod_count, \\\"node\\\", \\\"$1\\\", \\\"instance\\\", \\\"(.*)\\\")) by (node) / sum(kube_node_status_capacity_pods) by (node) * 100\",\n\"format\": \"time_series\",\n\"interval\": \"\",\n@@ -1840,7 +1879,6 @@ grafana:\n\"refId\": \"A\"\n},\n{\n- \"$$hashKey\": \"object:1694\",\n\"expr\": \"\",\n\"format\": \"time_series\",\n\"intervalFactor\": 1,\n@@ -1867,7 +1905,6 @@ grafana:\n},\n\"yaxes\": [\n{\n- \"$$hashKey\": \"object:344\",\n\"decimals\": null,\n\"format\": \"percent\",\n\"label\": \"\",\n@@ -1877,7 +1914,6 @@ grafana:\n\"show\": false\n},\n{\n- \"$$hashKey\": \"object:345\",\n\"format\": \"short\",\n\"label\": null,\n\"logBase\": 1,\n@@ -1965,7 +2001,6 @@ grafana:\n\"tableColumn\": \"\",\n\"targets\": [\n{\n- \"$$hashKey\": \"object:4935\",\n\"expr\": \"limes_project_usage{project=\\\"kubernikus\\\",resource=\\\"volumes\\\",domain=\\\"ccadmin\\\"} / limes_project_quota{project=\\\"kubernikus\\\",resource=\\\"volumes\\\",domain=\\\"ccadmin\\\"} * 100\",\n\"format\": \"time_series\",\n\"hide\": false,\n@@ -2053,7 +2088,6 @@ grafana:\n\"tableColumn\": \"\",\n\"targets\": [\n{\n- \"$$hashKey\": \"object:4935\",\n\"expr\": \"sum(container_fs_usage_bytes{pod_name=~\\\".*-etcd-.*\\\",image!~\\\".*pause.*\\\"}) / sum(kube_persistentvolumeclaim_resource_requests_storage_bytes) * 100\",\n\"format\": \"time_series\",\n\"hide\": false,\n@@ -2140,7 +2174,6 @@ grafana:\n\"tableColumn\": \"\",\n\"targets\": [\n{\n- \"$$hashKey\": \"object:7405\",\n\"expr\": \"limes_project_quota{project=\\\"kubernikus\\\",resource=\\\"instances\\\",domain=\\\"ccadmin\\\"} / limes_project_usage{project=\\\"kubernikus\\\",resource=\\\"instances\\\",domain=\\\"ccadmin\\\"}\",\n\"format\": \"time_series\",\n\"intervalFactor\": 1,\n@@ -2225,7 +2258,6 @@ grafana:\n\"tableColumn\": \"\",\n\"targets\": [\n{\n- \"$$hashKey\": \"object:7227\",\n\"expr\": \"limes_project_quota{project=\\\"kubernikus\\\",resource=\\\"cores\\\",domain=\\\"ccadmin\\\"} / limes_project_usage{project=\\\"kubernikus\\\",resource=\\\"cores\\\",domain=\\\"ccadmin\\\"}\",\n\"format\": \"time_series\",\n\"intervalFactor\": 1,\n@@ -2669,7 +2701,100 @@ grafana:\n\"style\": \"dark\",\n\"tags\": [],\n\"templating\": {\n- \"list\": []\n+ \"list\": [\n+ {\n+ \"allValue\": null,\n+ \"current\": {\n+ \"text\": \"1.12.6\",\n+ \"value\": \"1.12.6\"\n+ },\n+ \"datasource\": \"Prometheus\",\n+ \"hide\": 2,\n+ \"includeAll\": false,\n+ \"label\": null,\n+ \"multi\": false,\n+ \"name\": \"container_runtime_version\",\n+ \"options\": [],\n+ \"query\": \"label_values(kube_node_info,container_runtime_version)\",\n+ \"refresh\": 1,\n+ \"regex\": \"/[^0-9]*([0-9]+.[0-9]+.[0-9]+)[^0-9]*/\",\n+ \"sort\": 0,\n+ \"tagValuesQuery\": \"\",\n+ \"tags\": [],\n+ \"tagsQuery\": \"\",\n+ \"type\": \"query\",\n+ \"useTags\": false\n+ },\n+ {\n+ \"allValue\": null,\n+ \"current\": {\n+ \"text\": \"1.7.5\",\n+ \"value\": \"1.7.5\"\n+ },\n+ \"datasource\": \"Prometheus\",\n+ \"hide\": 2,\n+ \"includeAll\": false,\n+ \"label\": null,\n+ \"multi\": false,\n+ \"name\": \"kubelet_version\",\n+ \"options\": [],\n+ \"query\": \"label_values(kube_node_info, kubelet_version)\",\n+ \"refresh\": 1,\n+ \"regex\": \"/[^0-9]*([0-9]+.[0-9]+.[0-9]+)[^0-9]*/\",\n+ \"sort\": 0,\n+ \"tagValuesQuery\": \"\",\n+ \"tags\": [],\n+ \"tagsQuery\": \"\",\n+ \"type\": \"query\",\n+ \"useTags\": false\n+ },\n+ {\n+ \"allValue\": null,\n+ \"current\": {\n+ \"text\": \"1576.4.0\",\n+ \"value\": \"1576.4.0\"\n+ },\n+ \"datasource\": \"Prometheus\",\n+ \"hide\": 2,\n+ \"includeAll\": false,\n+ \"label\": null,\n+ \"multi\": false,\n+ \"name\": \"coreos_version\",\n+ \"options\": [],\n+ \"query\": \"label_values(kube_node_info, os_image)\",\n+ \"refresh\": 1,\n+ \"regex\": \"/[^0-9]*([0-9]+.[0-9]+.[0-9]+)[^0-9]*/\",\n+ \"sort\": 3,\n+ \"tagValuesQuery\": \"\",\n+ \"tags\": [],\n+ \"tagsQuery\": \"\",\n+ \"type\": \"query\",\n+ \"useTags\": false\n+ },\n+ {\n+ \"allValue\": null,\n+ \"current\": {\n+ \"text\": \"4.13.16\",\n+ \"value\": \"4.13.16\"\n+ },\n+ \"datasource\": \"Prometheus\",\n+ \"hide\": 2,\n+ \"includeAll\": false,\n+ \"label\": null,\n+ \"multi\": false,\n+ \"name\": \"kernel_version\",\n+ \"options\": [],\n+ \"query\": \"label_values(kube_node_info, kernel_version)\",\n+ \"refresh\": 1,\n+ \"regex\": \"/[^0-9]*([0-9]+.[0-9]+.[0-9]+)[^0-9]*/\",\n+ \"sort\": 3,\n+ \"tagValuesQuery\": \"\",\n+ \"tags\": [],\n+ \"tagsQuery\": \"\",\n+ \"type\": \"query\",\n+ \"useTags\": false\n+ }\n+ ]\n},\n\"time\": {\n\"from\": \"now-24h\",\n@@ -2702,179 +2827,5 @@ grafana:\n},\n\"timezone\": \"\",\n\"title\": \"Kubernikus\",\n- \"version\": 16\n+ \"version\": 5\n}\n-\n-\n-\n- provisioningDashboardsFiles:\n- default.yaml: |\n- - name: 'default'\n- org_id: 1\n- type: file\n- options:\n- path: /var/lib/grafana/dashboards\n-\n- provisioningDatasourcesFiles:\n- prometheus.yaml: |\n- datasources:\n- - name: Prometheus\n- type: prometheus\n- access: proxy\n- org_id: 1\n- url: http://kubernikus-prometheus:9090\n- is_default: true\n- version: 1\n- editable: false\n-\n- serverConfigFile:\n- grafana.ini: |\n- [paths]\n- data = /var/lib/grafana/data\n- logs = /var/log/grafana\n- plugins = /var/lib/grafana/plugins\n- provisioning = /var/lib/grafana/provisioning\n-\n- [server]\n- protocol = http\n- http_addr =\n- http_port = 3000\n- domain = localhost\n- enforce_domain = false\n- root_url = %(protocol)s://%(domain)s:%(http_port)s/grafana\n- router_logging = false\n- static_root_path = public\n- enable_gzip = false\n- cert_file =\n- cert_key =\n- socket = /tmp/grafana.sock\n-\n- [database]\n- ;type = sqlite3\n- ;host = 127.0.0.1:3306\n- ;name = grafana\n- ;user = root\n- ;password =\n- ;ssl_mode = disable\n- ;path = grafana.db\n-\n- [session]\n- ;provider = file\n- ;provider_config = sessions\n- ;cookie_name = grafana_sess\n- ;cookie_secure = false\n- ;session_life_time = 86400\n-\n- [analytics]\n- ;reporting_enabled = true\n- ;check_for_updates = true\n- ;google_analytics_ua_id =\n-\n- [security]\n- ;admin_user = admin\n- ;admin_password = admin\n- ;secret_key = SW2YcwTIb9zpOOhoPsMm\n- ;login_remember_days = 7\n- ;cookie_username = grafana_user\n- ;cookie_remember_name = grafana_remember\n- ;disable_gravatar = false\n- ;data_source_proxy_whitelist =\n-\n- [snapshots]\n- ;external_enabled = true\n- ;external_snapshot_url = https://snapshots-origin.raintank.io\n- ;external_snapshot_name = Publish to snapshot.raintank.io\n-\n- [users]\n- allow_sign_up = false\n- ;allow_org_create = true\n- ;auto_assign_org = true\n- ;auto_assign_org_role = Viewer\n- ;login_hint = email or username\n- ;default_theme = dark\n-\n- [auth.anonymous]\n- enabled = true\n- org_name = Main Org.\n- org_role = Viewer\n-\n- [auth.github]\n- ;enabled = false\n- ;allow_sign_up = false\n- ;client_id = some_id\n- ;client_secret = some_secret\n- ;scopes = user:email,read:org\n- ;auth_url = https://github.com/login/oauth/authorize\n- ;token_url = https://github.com/login/oauth/access_token\n- ;api_url = https://api.github.com/user\n- ;team_ids =\n- ;allowed_organizations =\n-\n- [auth.google]\n- ;enabled = false\n- ;allow_sign_up = false\n- ;client_id = some_client_id\n- ;client_secret = some_client_secret\n- ;scopes = https://www.googleapis.com/auth/userinfo.profile https://www.googleapis.com/auth/userinfo.email\n- ;auth_url = https://accounts.google.com/o/oauth2/auth\n- ;token_url = https://accounts.google.com/o/oauth2/token\n- ;api_url = https://www.googleapis.com/oauth2/v1/userinfo\n- ;allowed_domains =\n-\n- [auth.proxy]\n- ;enabled = false\n- ;header_name = X-WEBAUTH-USER\n- ;header_property = username\n- ;auto_sign_up = true\n-\n- [auth.basic]\n- enabled = true\n-\n- [auth.ldap]\n- ;enabled = false\n- ;config_file = /etc/grafana/ldap.toml\n-\n- [smtp]\n- enabled = false\n- ;host = localhost:25\n- ;user =\n- ;password =\n- ;cert_file =\n- ;key_file =\n- ;skip_verify = false\n- ;from_address = admin@grafana.localhost\n-\n- [emails]\n- ;welcome_email_on_sign_up = false\n-\n- [log]\n- mode = console\n- level = info\n-\n- [log.console]\n- ;level =\n- ;format = console\n-\n- [event_publisher]\n- ;enabled = false\n- ;rabbitmq_url = amqp://localhost/\n- ;exchange = grafana_events\n-\n- [dashboards.json]\n- enabled = true\n- path = /var/lib/grafana/dashboards\n-\n- [metrics]\n- ;enabled = true\n- ;interval_seconds = 10\n-\n- ; [metrics.graphite]\n- ; address = localhost:2003\n- ; prefix = prod.grafana.%(instance_name)s.\n-\n- [grafana_net]\n- url = https://grafana.net\n-\n- [alerting]\n- enabled = false\n-\n"
},
{
"change_type": "MODIFY",
"old_path": "charts/kubernikus/templates/ingress.yaml",
"new_path": "charts/kubernikus/templates/ingress.yaml",
"diff": "@@ -6,7 +6,7 @@ metadata:\nannotations:\nvice-president: \"true\"\nprometheus.io/probe: \"true\"\n- ingress.kubernetes.io/rewrite-target: /\n+\nspec:\ntls:\n- secretName: kubernikus-api\n@@ -19,11 +19,3 @@ spec:\nbackend:\nserviceName: kubernikus-api\nservicePort: {{ .Values.api.port }}\n- - path: /prometheus\n- backend:\n- serviceName: kubernikus-prometheus\n- servicePort: 9090\n- - path: /grafana\n- backend:\n- serviceName: kubernikus-system-grafana\n- servicePort: 80\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
splits ingress specs. adds authentication via tls
|
596,240 |
07.02.2018 16:18:20
| -3,600 |
f1735344d84e9fb71403d76ff5bab88d9a83ebb3
|
removes all usermgmt. it's bugged.
See:
|
[
{
"change_type": "MODIFY",
"old_path": "charts/kubernikus-system/values.yaml",
"new_path": "charts/kubernikus-system/values.yaml",
"diff": "@@ -85,7 +85,7 @@ grafana:\nsocket = /tmp/grafana.sock\n[users]\n- allow_sign_up = true\n+ allow_sign_up = false\nallow_org_create = false\nauto_assign_org = true\nauto_assign_org_role = Admin\n@@ -94,13 +94,13 @@ grafana:\n[auth.anonymous]\nenabled = true\norg_name = Main Org.\n- org_role = Viewer\n+ org_role = Admin\n[auth.proxy]\nenabled = true\nheader_name = X-REMOTE-USER\nheader_property = username\n- auto_sign_up = true\n+ auto_sign_up = false\n[auth.basic]\nenabled = false\n@@ -134,7 +134,7 @@ grafana:\n\"editable\": true,\n\"gnetId\": null,\n\"graphTooltip\": 0,\n- \"iteration\": 1517825496293,\n+ \"iteration\": 1517997489021,\n\"links\": [],\n\"panels\": [\n{\n@@ -596,16 +596,20 @@ grafana:\n\"steppedLine\": false,\n\"targets\": [\n{\n+ \"$$hashKey\": \"object:5818\",\n\"expr\": \"sum(probe_http_duration_seconds{path=\\\"/\\\"}) * 1000\",\n\"format\": \"time_series\",\n+ \"hide\": false,\n\"instant\": false,\n\"intervalFactor\": 1,\n\"legendFormat\": \"API\",\n\"refId\": \"A\"\n},\n{\n+ \"$$hashKey\": \"object:5819\",\n\"expr\": \"sum(avg_over_time(probe_http_duration_seconds{path=\\\"/\\\"}[1h])) * 1000\",\n\"format\": \"time_series\",\n+ \"hide\": false,\n\"intervalFactor\": 1,\n\"legendFormat\": \"Avg\",\n\"refId\": \"B\"\n@@ -2512,6 +2516,7 @@ grafana:\n\"steppedLine\": false,\n\"targets\": [\n{\n+ \"$$hashKey\": \"object:5395\",\n\"expr\": \"sum(300 * rate(kubernikus_launch_failed_operation_total{method!=\\\"Reconcile\\\"}[5m])) by (method)\",\n\"format\": \"time_series\",\n\"hide\": true,\n@@ -2521,6 +2526,7 @@ grafana:\n\"refId\": \"A\"\n},\n{\n+ \"$$hashKey\": \"object:5396\",\n\"expr\": \"sum(300 * rate(kubernikus_launch_successful_operation_total{method!=\\\"Reconcile\\\"}[5m])) by (method)\",\n\"format\": \"time_series\",\n\"hide\": true,\n@@ -2529,17 +2535,19 @@ grafana:\n\"refId\": \"B\"\n},\n{\n- \"expr\": \"60 * sum(rate(kubernikus_deorbit_successful_operation_total[5m]))\",\n+ \"$$hashKey\": \"object:5397\",\n+ \"expr\": \"60 * sum(rate(kubernikus_deorbit_successful_operation_total[11m]))\",\n\"format\": \"time_series\",\n\"interval\": \"\",\n- \"intervalFactor\": 10,\n+ \"intervalFactor\": 1,\n\"legendFormat\": \"Sucess\",\n\"refId\": \"C\"\n},\n{\n- \"expr\": \"60 * sum(rate(kubernikus_deorbit_failed_operation_total[5m]))\",\n+ \"$$hashKey\": \"object:5398\",\n+ \"expr\": \"60 * sum(rate(kubernikus_deorbit_failed_operation_total[11m]))\",\n\"format\": \"time_series\",\n- \"intervalFactor\": 10,\n+ \"intervalFactor\": 1,\n\"legendFormat\": \"Failed\",\n\"refId\": \"D\"\n}\n@@ -2628,6 +2636,7 @@ grafana:\n\"steppedLine\": false,\n\"targets\": [\n{\n+ \"$$hashKey\": \"object:4964\",\n\"expr\": \"sum(60 * rate(kubernikus_launch_failed_operation_total[5m])) by (method)\",\n\"format\": \"time_series\",\n\"hide\": true,\n@@ -2637,6 +2646,7 @@ grafana:\n\"refId\": \"A\"\n},\n{\n+ \"$$hashKey\": \"object:4965\",\n\"expr\": \"sum(60 * rate(kubernikus_launch_successful_operation_total{}[5m])) by (method)\",\n\"format\": \"time_series\",\n\"hide\": true,\n@@ -2645,17 +2655,19 @@ grafana:\n\"refId\": \"B\"\n},\n{\n- \"expr\": \"60 * sum(rate(kubernikus_launch_successful_operation_total[5m]))\",\n+ \"$$hashKey\": \"object:4966\",\n+ \"expr\": \"60 * sum(rate(kubernikus_launch_successful_operation_total[11m]))\",\n\"format\": \"time_series\",\n\"interval\": \"\",\n- \"intervalFactor\": 10,\n+ \"intervalFactor\": 1,\n\"legendFormat\": \"Sucess\",\n\"refId\": \"C\"\n},\n{\n- \"expr\": \"60 * sum(rate(kubernikus_launch_failed_operation_total[5m]))\",\n+ \"$$hashKey\": \"object:4967\",\n+ \"expr\": \"60 * sum(rate(kubernikus_launch_failed_operation_total[11m]))\",\n\"format\": \"time_series\",\n- \"intervalFactor\": 10,\n+ \"intervalFactor\": 1,\n\"legendFormat\": \"Failed\",\n\"refId\": \"D\"\n}\n@@ -2697,6 +2709,7 @@ grafana:\n]\n}\n],\n+ \"refresh\": \"5s\",\n\"schemaVersion\": 16,\n\"style\": \"dark\",\n\"tags\": [],\n@@ -2827,5 +2840,5 @@ grafana:\n},\n\"timezone\": \"\",\n\"title\": \"Kubernikus\",\n- \"version\": 5\n+ \"version\": 1\n}\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
removes all usermgmt. it's bugged.
See: https://github.com/grafana/grafana/issues/10276
|
596,240 |
08.02.2018 09:31:05
| -3,600 |
7c5feeaf2cd6a9254859d8f6ca746ba5f47d7569
|
read kubeconfig from tmpfile
|
[
{
"change_type": "MODIFY",
"old_path": "test/e2e/kubeCtlBuilder.go",
"new_path": "test/e2e/kubeCtlBuilder.go",
"diff": "@@ -3,7 +3,9 @@ package main\nimport (\n\"bytes\"\n\"fmt\"\n+ \"io/ioutil\"\n\"log\"\n+ \"os\"\n\"os/exec\"\n\"strings\"\n\"time\"\n@@ -20,11 +22,28 @@ func NewKubectlCommand(args ...string) *kubectlBuilder {\nreturn b\n}\n-func RunHostCmd(namespace, name, cmd string) (string, error) {\n- return RunKubectl(\"exec\", fmt.Sprintf(\"--namespace=%v\", namespace), name, \"--\", \"/bin/sh\", \"-c\", cmd)\n+func RunHostCmd(config, namespace, name, cmd string) (string, error) {\n+ return RunKubectl(config, \"exec\", fmt.Sprintf(\"--namespace=%v\", namespace), name, \"--\", \"/bin/sh\", \"-c\", cmd)\n}\n-func RunKubectl(args ...string) (string, error) {\n+func RunKubectl(config string, args ...string) (string, error) {\n+ tmpfile, err := ioutil.TempFile(\"\", \"kubeconfig\")\n+ if err != nil {\n+ return \"\", fmt.Errorf(\"Couldn't create temporary kubeconfig: %v\", err)\n+ }\n+ defer os.Remove(tmpfile.Name())\n+\n+ if _, err := tmpfile.Write([]byte(config)); err != nil {\n+ return \"\", fmt.Errorf(\"Couldn't write temporary kubeconfig: %v\", err)\n+ }\n+\n+ if err := tmpfile.Close(); err != nil {\n+ return \"\", fmt.Errorf(\"Couldn't close temporary kubeconfig: %v\", err)\n+ }\n+\n+ kubeConfigArg := fmt.Sprintf(\"--kubeconfig=%s\", tmpfile.Name())\n+ args = append([]string{kubeConfigArg}, args...)\n+\nreturn NewKubectlCommand(args...).Exec()\n}\n"
},
{
"change_type": "MODIFY",
"old_path": "test/e2e/smokeTests.go",
"new_path": "test/e2e/smokeTests.go",
"diff": "@@ -148,12 +148,12 @@ func (s *E2ETestSuite) dialServiceName(source *v1.Pod, target *v1.Service) {\nfunc (s *E2ETestSuite) dial(sourcePod *v1.Pod, targetIP string, targetPort int32) (string, error) {\ncmd := fmt.Sprintf(\"wget --timeout=%v -O - http://%v:%v\", TimeoutWGET, targetIP, targetPort)\n- return RunHostCmd(sourcePod.GetNamespace(), sourcePod.GetName(), cmd)\n+ return RunHostCmd(s.KubeConfig, sourcePod.GetNamespace(), sourcePod.GetName(), cmd)\n}\nfunc (s *E2ETestSuite) writeFileToMountedVolume() {\ncmd := fmt.Sprintf(\"echo hase > %v/myfile\", PVCMountPath)\n- _, err := RunHostCmd(Namespace, PVCName, cmd)\n+ _, err := RunHostCmd(s.KubeConfig, Namespace, PVCName, cmd)\nresult := \"success\"\nif err != nil {\nresult = \"failure\"\n@@ -169,7 +169,7 @@ func (s *E2ETestSuite) writeFileToMountedVolume() {\nfunc (s *E2ETestSuite) readFileFromMountedVolume() {\ncmd := fmt.Sprintf(\"cat %v/myfile\", PVCMountPath)\n- _, err := RunHostCmd(Namespace, PVCName, cmd)\n+ _, err := RunHostCmd(s.KubeConfig, Namespace, PVCName, cmd)\nresult := \"success\"\nif err != nil {\nresult = \"failure\"\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
read kubeconfig from tmpfile
|
596,233 |
08.02.2018 13:24:33
| -3,600 |
a4b14abbfa246541d43e4905b14680adb8be6490
|
Blackbox
Blackbox probing
|
[
{
"change_type": "MODIFY",
"old_path": "charts/kubernikus-system/charts/blackbox-exporter/templates/configmap.yaml",
"new_path": "charts/kubernikus-system/charts/blackbox-exporter/templates/configmap.yaml",
"diff": "@@ -11,40 +11,34 @@ data:\nmodules:\nhttp_2xx:\nprober: http\n- http_post_2xx:\n+ http:\n+ fail_if_not_ssl: true\n+ no_follow_redirects: false\n+ http_400:\n+ prober: http\n+ http:\n+ fail_if_not_ssl: true\n+ valid_status_codes: [400]\n+ http_401:\nprober: http\nhttp:\n+ fail_if_not_ssl: true\n+ valid_status_codes: [401]\n+ http_403:\n+ prober: http\n+ http:\n+ fail_if_not_ssl: true\n+ valid_status_codes: [403]\n+ http_post_4xx:\n+ prober: http\n+ http:\n+ fail_if_not_ssl: true\nmethod: POST\nicmp:\nprober: icmp\n- irc_banner:\n- prober: tcp\n- tcp:\n- query_response:\n- - send: NICK prober\n- - send: USER prober prober prober :prober\n- - expect: PING :([^ ]+)\n- send: PONG ${1}\n- - expect: ^:[^ ]+ 001\n- pop3s_banner:\n+ tcp_connect:\nprober: tcp\n+ timeout: 5s\ntcp:\n- query_response:\n- - expect: ^+OK\ntls: true\n- ssh_banner:\n- prober: tcp\n- tcp:\n- query_response:\n- - expect: ^SSH-2.0-\n- tcp_connect:\n- prober: tcp\n- http_4xx:\n- prober: http\n- http:\n- valid_status_codes: [401,403]\n- http_post_4xx:\n- prober: http\n- http:\n- method: POST\n"
},
{
"change_type": "MODIFY",
"old_path": "charts/kubernikus-system/charts/prometheus/templates/_prometheus.yaml.tpl",
"new_path": "charts/kubernikus-system/charts/prometheus/templates/_prometheus.yaml.tpl",
"diff": "@@ -267,10 +267,10 @@ scrape_configs:\n- source_labels: [__meta_kubernetes_ingress_annotation_prometheus_io_probe]\naction: keep\nregex: true\n- # consider prometheus.io/probe_code annotation. mind below regex.\n+ # consider prometheus.io/probe_code annotation.\n- source_labels: [__meta_kubernetes_ingress_annotation_prometheus_io_probe_code]\n- regex: ^(\\d).+\n- replacement: http_${1}xx\n+ regex: (\\b\\w{3,})\n+ replacement: http_${1}\ntarget_label: __param_module\n- source_labels: [__meta_kubernetes_ingress_scheme,__address__,__meta_kubernetes_ingress_path]\nregex: (.+);(.+);(.+)\n"
},
{
"change_type": "MODIFY",
"old_path": "charts/kubernikus-system/templates/grafana-ingress.yaml",
"new_path": "charts/kubernikus-system/templates/grafana-ingress.yaml",
"diff": "@@ -6,6 +6,7 @@ metadata:\nannotations:\nvice-president: \"true\"\nprometheus.io/probe: \"true\"\n+ prometheus.io/probe_code: \"400\"\nnginx.ingress.kubernetes.io/configuration-snippet: |\nrewrite ^/$ /dashboard/db/kubernikus?refresh=1m&orgId=1&kiosk=true redirect;\n{{- if .Values.authentication.enabled}}\n"
},
{
"change_type": "MODIFY",
"old_path": "charts/kubernikus-system/templates/prometheus-ingress.yaml",
"new_path": "charts/kubernikus-system/templates/prometheus-ingress.yaml",
"diff": "@@ -6,6 +6,7 @@ metadata:\nannotations:\nvice-president: \"true\"\nprometheus.io/probe: \"true\"\n+ prometheus.io/probe_code: \"400\"\n{{- if .Values.authentication.enabled}}\nnginx.ingress.kubernetes.io/auth-tls-secret: \"kubernikus-system/ca-crt\"\nnginx.ingress.kubernetes.io/auth-tls-verify-depth: \"3\"\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
Blackbox (#222)
Blackbox probing
|
596,240 |
08.02.2018 14:09:03
| -3,600 |
2b6519fa083a55f548e53313fa057c89ef0af83b
|
adds some extra debug
|
[
{
"change_type": "MODIFY",
"old_path": "test/e2e/setupSmokeTest.go",
"new_path": "test/e2e/setupSmokeTest.go",
"diff": "@@ -134,7 +134,7 @@ func (s *E2ETestSuite) getReadyNodes() {\nnodes, err := s.clientSet.CoreV1().Nodes().List(meta_v1.ListOptions{})\ns.handleError(err)\nfor _, node := range nodes.Items {\n- log.Printf(\"found node: %s\", node.Name)\n+ log.Printf(\"found node: %s podCidr: %s, IP: %s\", node.Name, node.Spec.PodCIDR, node.Status.Addresses[0].Address)\n}\ns.readyNodes = nodes.Items\n}\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
adds some extra debug
|
596,240 |
08.02.2018 17:21:16
| -3,600 |
e66c8ffe8d174828b8dd01daaf59500eaf9a389a
|
fixes ephemeral port ranges
|
[
{
"change_type": "MODIFY",
"old_path": "pkg/templates/node_1.7.go",
"new_path": "pkg/templates/node_1.7.go",
"diff": "@@ -234,8 +234,8 @@ storage:\n:INPUT ACCEPT [0:0]\n:OUTPUT ACCEPT [0:0]\n:POSTROUTING ACCEPT [0:0]\n- -A POSTROUTING -p tcp ! -d {{ .ClusterCIDR }} -m addrtype ! --dst-type LOCAL -j MASQUERADE --to-ports 32000-65000\n- -A POSTROUTING -p udp ! -d {{ .ClusterCIDR }} -m addrtype ! --dst-type LOCAL -j MASQUERADE --to-ports 32000-65000\n+ -A POSTROUTING -p tcp ! -d {{ .ClusterCIDR }} -m addrtype ! --dst-type LOCAL -j MASQUERADE --to-ports 32678-65535\n+ -A POSTROUTING -p udp ! -d {{ .ClusterCIDR }} -m addrtype ! --dst-type LOCAL -j MASQUERADE --to-ports 32678-65535\n-A POSTROUTING -p icmp ! -d {{ .ClusterCIDR }} -m addrtype ! --dst-type LOCAL -j MASQUERADE\nCOMMIT\n- path: /etc/sysctl.d/10-enable-icmp-redirects\n"
},
{
"change_type": "MODIFY",
"old_path": "pkg/templates/node_1.8.go",
"new_path": "pkg/templates/node_1.8.go",
"diff": "@@ -234,8 +234,8 @@ storage:\n:INPUT ACCEPT [0:0]\n:OUTPUT ACCEPT [0:0]\n:POSTROUTING ACCEPT [0:0]\n- -A POSTROUTING -p tcp ! -d {{ .ClusterCIDR }} -m addrtype ! --dst-type LOCAL -j MASQUERADE --to-ports 32000-65000\n- -A POSTROUTING -p udp ! -d {{ .ClusterCIDR }} -m addrtype ! --dst-type LOCAL -j MASQUERADE --to-ports 32000-65000\n+ -A POSTROUTING -p tcp ! -d {{ .ClusterCIDR }} -m addrtype ! --dst-type LOCAL -j MASQUERADE --to-ports 32678-65535\n+ -A POSTROUTING -p udp ! -d {{ .ClusterCIDR }} -m addrtype ! --dst-type LOCAL -j MASQUERADE --to-ports 32678-65535\n-A POSTROUTING -p icmp ! -d {{ .ClusterCIDR }} -m addrtype ! --dst-type LOCAL -j MASQUERADE\nCOMMIT\n- path: /etc/sysctl.d/10-enable-icmp-redirects\n"
},
{
"change_type": "MODIFY",
"old_path": "pkg/templates/node_1.9.go",
"new_path": "pkg/templates/node_1.9.go",
"diff": "@@ -234,8 +234,8 @@ storage:\n:INPUT ACCEPT [0:0]\n:OUTPUT ACCEPT [0:0]\n:POSTROUTING ACCEPT [0:0]\n- -A POSTROUTING -p tcp ! -d {{ .ClusterCIDR }} -m addrtype ! --dst-type LOCAL -j MASQUERADE --to-ports 32000-65000\n- -A POSTROUTING -p udp ! -d {{ .ClusterCIDR }} -m addrtype ! --dst-type LOCAL -j MASQUERADE --to-ports 32000-65000\n+ -A POSTROUTING -p tcp ! -d {{ .ClusterCIDR }} -m addrtype ! --dst-type LOCAL -j MASQUERADE --to-ports 32678-65535\n+ -A POSTROUTING -p udp ! -d {{ .ClusterCIDR }} -m addrtype ! --dst-type LOCAL -j MASQUERADE --to-ports 32678-65535\n-A POSTROUTING -p icmp ! -d {{ .ClusterCIDR }} -m addrtype ! --dst-type LOCAL -j MASQUERADE\nCOMMIT\n- path: /etc/sysctl.d/10-enable-icmp-redirects\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
fixes ephemeral port ranges
|
596,240 |
08.02.2018 19:34:26
| -3,600 |
c843c31992cb41f4610ff053aae3f611f2e053e8
|
try to coarse e2e tests to kubernikus.staging
|
[
{
"change_type": "MODIFY",
"old_path": "ci/task_e2e_tests.yaml",
"new_path": "ci/task_e2e_tests.yaml",
"diff": "@@ -16,7 +16,8 @@ run:\nargs:\n- -c\n- |\n- export KUBERNETES_VERSION=v1.7.7\n+ export KUBERNIKUS_API_SERVER=kubernikus.staging.cloud.sap\n+ export KUBERNETES_VERSION=v1.9.0\nexport GOPATH=$PWD/gopath\ncd gopath/src/github.com/sapcc/kubernikus\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
try to coarse e2e tests to kubernikus.staging
|
596,240 |
09.02.2018 09:15:22
| -3,600 |
5a865e1f7fe88a9d90b19ccdaeebc02fd67c1efe
|
use kubernikus staging for e2e tests
|
[
{
"change_type": "MODIFY",
"old_path": "ci/pipeline.yaml",
"new_path": "ci/pipeline.yaml",
"diff": "@@ -159,13 +159,13 @@ jobs:\n- task: e2e_tests\nfile: pipeline.git/ci/task_e2e_tests.yaml\nparams:\n- OS_AUTH_URL: https://identity-3.eu-nl-1.cloud.sap/v3\n- OS_USERNAME: {{kubernikus-eu-nl-1-username}}\n- OS_PASSWORD: {{kubernikus-eu-nl-1-password}}\n+ OS_AUTH_URL: https://identity-3.staging.cloud.sap/v3\n+ OS_USERNAME: {{kubernikus-staging-username}}\n+ OS_PASSWORD: {{kubernikus-staging-password}}\nOS_USER_DOMAIN_NAME: ccadmin\nOS_PROJECT_NAME: kubernikus-e2e\nOS_PROJECT_DOMAIN_NAME: ccadmin\n- OS_REGION_NAME: eu-nl-1\n+ OS_REGION_NAME: staging\non_failure:\n<<: *post_failure_to_slack\n"
},
{
"change_type": "MODIFY",
"old_path": "ci/task_e2e_tests.yaml",
"new_path": "ci/task_e2e_tests.yaml",
"diff": "@@ -16,7 +16,6 @@ run:\nargs:\n- -c\n- |\n- export KUBERNIKUS_API_SERVER=kubernikus.staging.cloud.sap\nexport KUBERNETES_VERSION=v1.9.0\nexport GOPATH=$PWD/gopath\ncd gopath/src/github.com/sapcc/kubernikus\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
use kubernikus staging for e2e tests
|
596,240 |
12.02.2018 14:46:07
| -3,600 |
4759e9582dec298fa9552375b5119f59388fab65
|
adds goroutine panel for operator
|
[
{
"change_type": "MODIFY",
"old_path": "charts/kubernikus-system/values.yaml",
"new_path": "charts/kubernikus-system/values.yaml",
"diff": "@@ -134,7 +134,8 @@ grafana:\n\"editable\": true,\n\"gnetId\": null,\n\"graphTooltip\": 0,\n- \"iteration\": 1517997489021,\n+ \"id\": 1,\n+ \"iteration\": 1518439197401,\n\"links\": [],\n\"panels\": [\n{\n@@ -596,7 +597,6 @@ grafana:\n\"steppedLine\": false,\n\"targets\": [\n{\n- \"$$hashKey\": \"object:5818\",\n\"expr\": \"sum(probe_http_duration_seconds{path=\\\"/\\\"}) * 1000\",\n\"format\": \"time_series\",\n\"hide\": false,\n@@ -606,7 +606,6 @@ grafana:\n\"refId\": \"A\"\n},\n{\n- \"$$hashKey\": \"object:5819\",\n\"expr\": \"sum(avg_over_time(probe_http_duration_seconds{path=\\\"/\\\"}[1h])) * 1000\",\n\"format\": \"time_series\",\n\"hide\": false,\n@@ -2516,7 +2515,6 @@ grafana:\n\"steppedLine\": false,\n\"targets\": [\n{\n- \"$$hashKey\": \"object:5395\",\n\"expr\": \"sum(300 * rate(kubernikus_launch_failed_operation_total{method!=\\\"Reconcile\\\"}[5m])) by (method)\",\n\"format\": \"time_series\",\n\"hide\": true,\n@@ -2526,7 +2524,6 @@ grafana:\n\"refId\": \"A\"\n},\n{\n- \"$$hashKey\": \"object:5396\",\n\"expr\": \"sum(300 * rate(kubernikus_launch_successful_operation_total{method!=\\\"Reconcile\\\"}[5m])) by (method)\",\n\"format\": \"time_series\",\n\"hide\": true,\n@@ -2535,7 +2532,6 @@ grafana:\n\"refId\": \"B\"\n},\n{\n- \"$$hashKey\": \"object:5397\",\n\"expr\": \"60 * sum(rate(kubernikus_deorbit_successful_operation_total[11m]))\",\n\"format\": \"time_series\",\n\"interval\": \"\",\n@@ -2544,7 +2540,6 @@ grafana:\n\"refId\": \"C\"\n},\n{\n- \"$$hashKey\": \"object:5398\",\n\"expr\": \"60 * sum(rate(kubernikus_deorbit_failed_operation_total[11m]))\",\n\"format\": \"time_series\",\n\"intervalFactor\": 1,\n@@ -2636,7 +2631,6 @@ grafana:\n\"steppedLine\": false,\n\"targets\": [\n{\n- \"$$hashKey\": \"object:4964\",\n\"expr\": \"sum(60 * rate(kubernikus_launch_failed_operation_total[5m])) by (method)\",\n\"format\": \"time_series\",\n\"hide\": true,\n@@ -2646,7 +2640,6 @@ grafana:\n\"refId\": \"A\"\n},\n{\n- \"$$hashKey\": \"object:4965\",\n\"expr\": \"sum(60 * rate(kubernikus_launch_successful_operation_total{}[5m])) by (method)\",\n\"format\": \"time_series\",\n\"hide\": true,\n@@ -2655,7 +2648,6 @@ grafana:\n\"refId\": \"B\"\n},\n{\n- \"$$hashKey\": \"object:4966\",\n\"expr\": \"60 * sum(rate(kubernikus_launch_successful_operation_total[11m]))\",\n\"format\": \"time_series\",\n\"interval\": \"\",\n@@ -2664,7 +2656,6 @@ grafana:\n\"refId\": \"C\"\n},\n{\n- \"$$hashKey\": \"object:4967\",\n\"expr\": \"60 * sum(rate(kubernikus_launch_failed_operation_total[11m]))\",\n\"format\": \"time_series\",\n\"intervalFactor\": 1,\n@@ -2707,9 +2698,108 @@ grafana:\n\"show\": true\n}\n]\n+ },\n+ {\n+ \"aliasColors\": {},\n+ \"bars\": false,\n+ \"dashLength\": 10,\n+ \"dashes\": false,\n+ \"datasource\": null,\n+ \"fill\": 0,\n+ \"gridPos\": {\n+ \"h\": 7,\n+ \"w\": 6,\n+ \"x\": 18,\n+ \"y\": 13\n+ },\n+ \"id\": 69,\n+ \"legend\": {\n+ \"avg\": false,\n+ \"current\": false,\n+ \"max\": false,\n+ \"min\": false,\n+ \"show\": false,\n+ \"total\": false,\n+ \"values\": false\n+ },\n+ \"lines\": true,\n+ \"linewidth\": 2,\n+ \"links\": [],\n+ \"nullPointMode\": \"connected\",\n+ \"percentage\": false,\n+ \"pointradius\": 5,\n+ \"points\": false,\n+ \"renderer\": \"flot\",\n+ \"seriesOverrides\": [\n+ {\n+ \"$$hashKey\": \"object:656\",\n+ \"alias\": \"24h Average\",\n+ \"dashes\": true,\n+ \"linewidth\": 5\n+ }\n+ ],\n+ \"spaceLength\": 10,\n+ \"stack\": false,\n+ \"steppedLine\": false,\n+ \"targets\": [\n+ {\n+ \"$$hashKey\": \"object:256\",\n+ \"expr\": \"rate(go_goroutines{app=\\\"kubernikus\\\"}[5m])\",\n+ \"format\": \"time_series\",\n+ \"intervalFactor\": 1,\n+ \"legendFormat\": \"Current\",\n+ \"refId\": \"D\"\n+ },\n+ {\n+ \"$$hashKey\": \"object:613\",\n+ \"expr\": \"rate(go_goroutines{app=\\\"kubernikus\\\"}[24h])\\n\",\n+ \"format\": \"time_series\",\n+ \"intervalFactor\": 1,\n+ \"legendFormat\": \"24h Average\",\n+ \"refId\": \"A\"\n+ }\n+ ],\n+ \"thresholds\": [],\n+ \"timeFrom\": null,\n+ \"timeShift\": null,\n+ \"title\": \"GoRoutines\",\n+ \"tooltip\": {\n+ \"shared\": true,\n+ \"sort\": 0,\n+ \"value_type\": \"individual\"\n+ },\n+ \"type\": \"graph\",\n+ \"xaxis\": {\n+ \"buckets\": null,\n+ \"mode\": \"time\",\n+ \"name\": null,\n+ \"show\": true,\n+ \"values\": []\n+ },\n+ \"yaxes\": [\n+ {\n+ \"$$hashKey\": \"object:318\",\n+ \"decimals\": null,\n+ \"format\": \"short\",\n+ \"label\": null,\n+ \"logBase\": 1,\n+ \"max\": null,\n+ \"min\": null,\n+ \"show\": true\n+ },\n+ {\n+ \"$$hashKey\": \"object:319\",\n+ \"format\": \"short\",\n+ \"label\": null,\n+ \"logBase\": 1,\n+ \"max\": null,\n+ \"min\": null,\n+ \"show\": true\n+ }\n+ ]\n}\n],\n- \"refresh\": \"5s\",\n+ \"refresh\": \"1m\",\n\"schemaVersion\": 16,\n\"style\": \"dark\",\n\"tags\": [],\n@@ -2840,5 +2930,5 @@ grafana:\n},\n\"timezone\": \"\",\n\"title\": \"Kubernikus\",\n- \"version\": 1\n+ \"version\": 3\n}\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
adds goroutine panel for operator
|
596,233 |
13.02.2018 14:03:29
| -3,600 |
c27424532131d488176aad8c3e1d15484185a5dc
|
grafana responds with 200
|
[
{
"change_type": "MODIFY",
"old_path": "charts/kubernikus-system/templates/grafana-ingress.yaml",
"new_path": "charts/kubernikus-system/templates/grafana-ingress.yaml",
"diff": "@@ -6,7 +6,6 @@ metadata:\nannotations:\nvice-president: \"true\"\nprometheus.io/probe: \"true\"\n- prometheus.io/probe_code: \"400\"\nnginx.ingress.kubernetes.io/configuration-snippet: |\nrewrite ^/$ /dashboard/db/kubernikus?refresh=1m&orgId=1&kiosk=true redirect;\n{{- if .Values.authentication.enabled}}\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
grafana responds with 200
|
596,240 |
15.02.2018 09:22:05
| -3,600 |
a13a769a50217a2d385506827a9b8d7144eda6f9
|
adds quantiles to kubernikus api response graph
|
[
{
"change_type": "MODIFY",
"old_path": "charts/kubernikus-system/values.yaml",
"new_path": "charts/kubernikus-system/values.yaml",
"diff": "@@ -121,6 +121,7 @@ grafana:\n\"annotations\": {\n\"list\": [\n{\n+ \"$$hashKey\": \"object:269\",\n\"builtIn\": 1,\n\"datasource\": \"default\",\n\"enable\": true,\n@@ -134,7 +135,7 @@ grafana:\n\"editable\": true,\n\"gnetId\": null,\n\"graphTooltip\": 0,\n- \"iteration\": 1518439197401,\n+ \"iteration\": 1518640266926,\n\"links\": [],\n\"panels\": [\n{\n@@ -552,7 +553,9 @@ grafana:\n\"type\": \"text\"\n},\n{\n- \"aliasColors\": {},\n+ \"aliasColors\": {\n+ \"{instance=\\\"https://kubernikus.eu-nl-1.cloud.sap/\\\",job=\\\"blackbox-ingress\\\",kubernetes_name=\\\"kubernikus-api\\\",kubernetes_namespace=\\\"kubernikus-system\\\",path=\\\"/\\\",phase=\\\"connect\\\"}\": \"#eab839\"\n+ },\n\"bars\": false,\n\"dashLength\": 10,\n\"dashes\": false,\n@@ -586,7 +589,9 @@ grafana:\n\"renderer\": \"flot\",\n\"seriesOverrides\": [\n{\n- \"alias\": \"Avg\",\n+ \"$$hashKey\": \"object:967\",\n+ \"alias\": \"0.95\",\n+ \"dashes\": true,\n\"fill\": 0,\n\"linewidth\": 3\n}\n@@ -596,20 +601,24 @@ grafana:\n\"steppedLine\": false,\n\"targets\": [\n{\n- \"expr\": \"sum(probe_http_duration_seconds{path=\\\"/\\\"}) * 1000\",\n+ \"$$hashKey\": \"object:497\",\n+ \"expr\": \"sum(probe_http_duration_seconds{instance=~\\\"https://kubernikus.*\\\",path=\\\"/\\\"}) * 1000\",\n\"format\": \"time_series\",\n\"hide\": false,\n\"instant\": false,\n+ \"interval\": \"\",\n\"intervalFactor\": 1,\n\"legendFormat\": \"API\",\n\"refId\": \"A\"\n},\n{\n- \"expr\": \"sum(avg_over_time(probe_http_duration_seconds{path=\\\"/\\\"}[1h])) * 1000\",\n+ \"$$hashKey\": \"object:498\",\n+ \"expr\": \"quantile_over_time(0.95, probe_duration_seconds{instance=~\\\"https://kubernikus.*\\\"}[24h]) * 1000\",\n\"format\": \"time_series\",\n\"hide\": false,\n+ \"interval\": \"\",\n\"intervalFactor\": 1,\n- \"legendFormat\": \"Avg\",\n+ \"legendFormat\": \"0.95\",\n\"refId\": \"B\"\n}\n],\n@@ -633,6 +642,7 @@ grafana:\n},\n\"yaxes\": [\n{\n+ \"$$hashKey\": \"object:543\",\n\"format\": \"ms\",\n\"label\": null,\n\"logBase\": 1,\n@@ -641,6 +651,7 @@ grafana:\n\"show\": true\n},\n{\n+ \"$$hashKey\": \"object:544\",\n\"format\": \"short\",\n\"label\": null,\n\"logBase\": 1,\n@@ -2798,7 +2809,7 @@ grafana:\n]\n}\n],\n- \"refresh\": \"1m\",\n+ \"refresh\": false,\n\"schemaVersion\": 16,\n\"style\": \"dark\",\n\"tags\": [],\n@@ -2899,8 +2910,8 @@ grafana:\n]\n},\n\"time\": {\n- \"from\": \"now-24h\",\n- \"to\": \"now\"\n+ \"from\": \"2018-02-14T15:51:41.744Z\",\n+ \"to\": \"2018-02-14T20:57:19.704Z\"\n},\n\"timepicker\": {\n\"refresh_intervals\": [\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
adds quantiles to kubernikus api response graph
|
596,240 |
15.02.2018 10:47:44
| -3,600 |
72ed2bacc6d8c75fc748441cfbe16ff14e96578b
|
makes e2e read Kubernikus options from ENV
|
[
{
"change_type": "MODIFY",
"old_path": "test/e2e/config.go",
"new_path": "test/e2e/config.go",
"diff": "@@ -32,7 +32,7 @@ func ReadConfig(filePath string) (Config, error) {\nfunc ReadFromEnv() Config {\nreturn Config{\n- APIURL: os.Getenv(\"KUBERNIKUS_API_SERVER\"),\n+ APIURL: os.Getenv(\"KUBERNIKUS_URL\"),\nAPIVersion: os.Getenv(\"KUBERNIKUS_API_VERSION\"),\n}\n}\n"
},
{
"change_type": "MODIFY",
"old_path": "test/e2e/e2e.go",
"new_path": "test/e2e/e2e.go",
"diff": "@@ -40,6 +40,7 @@ func NewE2ETestSuite(t *testing.T, options E2ETestSuiteOptions) *E2ETestSuite {\n}\nif err := options.Verify(); err != nil {\n+ options.Config = ReadFromEnv()\noptions.OpenStackCredentials = getOpenStackCredentialsFromENV()\nif err := options.Verify(); err != nil {\nlog.Fatalf(\"Checked config and env. Insufficient parameters for authentication : %v\", err)\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
makes e2e read Kubernikus options from ENV
|
596,240 |
15.02.2018 10:51:17
| -3,600 |
b7354ea74a085a5e0d67b39cf2e7958692104c3d
|
updates values location
|
[
{
"change_type": "MODIFY",
"old_path": "ci/task_helm_kubernikus-system.yaml",
"new_path": "ci/task_helm_kubernikus-system.yaml",
"diff": "@@ -22,12 +22,11 @@ run:\n# recent helm versions fails with `invalid cross-device link` when trying to rename charts.git\n# rename(2) for a directory is allowed only when both the source and the destination path are on the top layer\n# https://github.com/kubernetes/helm/issues/2998\n- #helm dep up --skip-refresh charts.git/charts/kubernikus-system/\n- #helm upgrade kubernikus-system charts.git/charts/kubernikus-system/ --namespace=kubernikus-system --values secrets.git/$KUBERNIKUS_NAME/values/kubernikus-system.yaml --install\ncp -a charts.git charts-copy.git\n+\nhelm repo add forked https://raw.githubusercontent.com/BugRoger/charts/repo\nhelm dep up --skip-refresh charts-copy.git/charts/kubernikus-system/\n- helm upgrade kubernikus-system charts-copy.git/charts/kubernikus-system/ --namespace=kubernikus-system --values secrets.git/$KUBERNIKUS_NAME/values/kubernikus-system.yaml --install\n+ helm upgrade kubernikus-system charts-copy.git/charts/kubernikus-system/ --namespace=kubernikus-system --values secrets.git/kubernikus/$KUBERNIKUS_NAME/values/kubernikus-system.yaml --install\nparams:\nOS_AUTH_URL:\n"
},
{
"change_type": "MODIFY",
"old_path": "ci/task_helm_kubernikus.yaml",
"new_path": "ci/task_helm_kubernikus.yaml",
"diff": "@@ -20,7 +20,7 @@ run:\nVERSION=$(cat kubernikus.builds/.git/HEAD)\nkubernikusctl auth init\nhelm dep up --skip-refresh kubernikus.builds/charts/kubernikus/\n- helm upgrade kubernikus kubernikus.builds/charts/kubernikus/ --namespace=kubernikus-system --values secrets.git/$KUBERNIKUS_NAME/values/kubernikus.yaml --set imageTag=$VERSION --install\n+ helm upgrade kubernikus kubernikus.builds/charts/kubernikus/ --namespace=kubernikus-system --values secrets.git/kubernikus/$KUBERNIKUS_NAME/values/kubernikus.yaml --set imageTag=$VERSION --install\nkubectl rollout status deployment/kubernikus-api --namespace=kubernikus-system\nkubectl rollout status deployment/kubernikus-operator --namespace=kubernikus-system\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
updates values location
|
596,240 |
15.02.2018 11:33:57
| -3,600 |
fb30ea3d419ef752693fc7505bdf15973d317296
|
renames charts repo
|
[
{
"change_type": "MODIFY",
"old_path": "ci/task_helm_kubernikus-system.yaml",
"new_path": "ci/task_helm_kubernikus-system.yaml",
"diff": "@@ -8,7 +8,7 @@ image_resource:\ntag: 'latest'\ninputs:\n- - name: charts.git\n+ - name: kubernikus.builds\n- name: secrets.git\nrun:\n@@ -22,11 +22,11 @@ run:\n# recent helm versions fails with `invalid cross-device link` when trying to rename charts.git\n# rename(2) for a directory is allowed only when both the source and the destination path are on the top layer\n# https://github.com/kubernetes/helm/issues/2998\n- cp -a charts.git charts-copy.git\n+ cp -a kubernikus.builds kubernikus-copy.builds\nhelm repo add forked https://raw.githubusercontent.com/BugRoger/charts/repo\n- helm dep up --skip-refresh charts-copy.git/charts/kubernikus-system/\n- helm upgrade kubernikus-system charts-copy.git/charts/kubernikus-system/ --namespace=kubernikus-system --values secrets.git/kubernikus/$KUBERNIKUS_NAME/values/kubernikus-system.yaml --install\n+ helm dep up --skip-refresh kubernikus-copy.builds/charts/kubernikus-system/\n+ helm upgrade kubernikus-system kubernikus-copy.builds/charts/kubernikus-system/ --namespace=kubernikus-system --values secrets.git/kubernikus/$KUBERNIKUS_NAME/values/kubernikus-system.yaml --install\nparams:\nOS_AUTH_URL:\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
renames charts repo
|
596,240 |
15.02.2018 11:39:35
| -3,600 |
bae64f68f728783cabe361a95ac1f2c8927e9875
|
builds feature branch as well
|
[
{
"change_type": "MODIFY",
"old_path": ".travis.yml",
"new_path": ".travis.yml",
"diff": "@@ -4,6 +4,7 @@ sudo: required\nbranches:\nonly:\n- master\n+ - feature\nenv:\n- VERSION=$TRAVIS_COMMIT\n@@ -16,7 +17,7 @@ script:\n- make pull build gh-pages\nafter_success:\n- - if [ \"$TRAVIS_BRANCH\" == \"master\" ]; then\n+ - if [ \"$TRAVIS_BRANCH\" == \"master\" ] || [ \"$TRAVIS_BRANCH\" == \"feature\" ]; then\necho $DOCKER_PASSWORD | docker login -u=\"$DOCKER_USERNAME\" --password-stdin;\nmake tag push;\nfi\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
builds feature branch as well
|
596,240 |
15.02.2018 14:10:13
| -3,600 |
009093263a521a20142b07320cc6dedfc0a2a03a
|
allows to pass Kubernikus URL
|
[
{
"change_type": "MODIFY",
"old_path": "test/e2e/auth.go",
"new_path": "test/e2e/auth.go",
"diff": "@@ -111,9 +111,6 @@ func (c *OpenStackCredentials) Verify() error {\nc.AuthURL += \"/\"\n}\n}\n- if c.RegionName == \"\" {\n- errorString += \"missing OS_REGION_NAME\\n\"\n- }\nif errorString != \"\" {\nreturn fmt.Errorf(errorString)\n}\n"
},
{
"change_type": "MODIFY",
"old_path": "test/e2e/config.go",
"new_path": "test/e2e/config.go",
"diff": "@@ -3,6 +3,8 @@ package main\nimport (\n\"fmt\"\n\"io/ioutil\"\n+ \"log\"\n+ \"net/url\"\n\"os\"\n@@ -31,8 +33,18 @@ func ReadConfig(filePath string) (Config, error) {\n}\nfunc ReadFromEnv() Config {\n+ env := os.Getenv(\"KUBERNIKUS_URL\")\n+ if env == \"\" {\n+ return Config{}\n+ }\n+\n+ kubernikus_url, err := url.Parse(env)\n+ if err != nil {\n+ log.Fatalf(\"Couldn't parse KUBERNIKUS_URL: %v\", err)\n+ }\n+\nreturn Config{\n- APIURL: os.Getenv(\"KUBERNIKUS_URL\"),\n+ APIURL: kubernikus_url.Host,\nAPIVersion: os.Getenv(\"KUBERNIKUS_API_VERSION\"),\n}\n}\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
allows to pass Kubernikus URL
|
596,240 |
15.02.2018 16:07:21
| -3,600 |
e0acdbce4581d4e16c0a024c9fbdaa14c8b9c648
|
force installation
|
[
{
"change_type": "MODIFY",
"old_path": "ci/task_helm_kubernikus-system.yaml",
"new_path": "ci/task_helm_kubernikus-system.yaml",
"diff": "@@ -26,7 +26,7 @@ run:\nhelm repo add forked https://raw.githubusercontent.com/BugRoger/charts/repo\nhelm dep up --skip-refresh kubernikus-copy.builds/charts/kubernikus-system/\n- helm upgrade kubernikus-system kubernikus-copy.builds/charts/kubernikus-system/ --namespace=kubernikus-system --values secrets.git/kubernikus/$KUBERNIKUS_NAME/values/kubernikus-system.yaml --install\n+ helm upgrade kubernikus-system kubernikus-copy.builds/charts/kubernikus-system/ --namespace=kubernikus-system --values secrets.git/kubernikus/$KUBERNIKUS_NAME/values/kubernikus-system.yaml --install --force\nparams:\nOS_AUTH_URL:\n"
},
{
"change_type": "MODIFY",
"old_path": "ci/task_helm_kubernikus.yaml",
"new_path": "ci/task_helm_kubernikus.yaml",
"diff": "@@ -20,7 +20,7 @@ run:\nVERSION=$(cat kubernikus.builds/.git/HEAD)\nkubernikusctl auth init\nhelm dep up --skip-refresh kubernikus.builds/charts/kubernikus/\n- helm upgrade kubernikus kubernikus.builds/charts/kubernikus/ --namespace=kubernikus-system --values secrets.git/kubernikus/$KUBERNIKUS_NAME/values/kubernikus.yaml --set imageTag=$VERSION --install\n+ helm upgrade kubernikus kubernikus.builds/charts/kubernikus/ --namespace=kubernikus-system --values secrets.git/kubernikus/$KUBERNIKUS_NAME/values/kubernikus.yaml --set imageTag=$VERSION --install --force\nkubectl rollout status deployment/kubernikus-api --namespace=kubernikus-system\nkubectl rollout status deployment/kubernikus-operator --namespace=kubernikus-system\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
force installation
|
596,240 |
16.02.2018 15:33:26
| -3,600 |
1d0d78d10ebc6402cfa8b38cb0bee79255cce0b6
|
reworked smoother pipeline
|
[
{
"change_type": "MODIFY",
"old_path": "ci/pipeline.yaml",
"new_path": "ci/pipeline.yaml",
"diff": "@@ -9,16 +9,69 @@ post_failure_to_slack: &post_failure_to_slack\n:boom: <$ATC_EXTERNAL_URL/teams/$BUILD_TEAM_NAME/pipelines/$BUILD_PIPELINE_NAME/jobs/$BUILD_JOB_NAME/builds/$BUILD_NAME|$BUILD_PIPELINE_NAME $BUILD_JOB_NAME job #$BUILD_NAME failed>\nTo debug run `fly hijack -j $BUILD_PIPELINE_NAME/$BUILD_JOB_NAME -b $BUILD_NAME`\n-resources:\n+auth_master: &auth_master\n+ OS_AUTH_URL: https://identity-3.eu-nl-1.cloud.sap/v3\n+ OS_USERNAME: {{kubernikus-eu-nl-1-username}}\n+ OS_PASSWORD: {{kubernikus-eu-nl-1-password}}\n+ OS_USER_DOMAIN_NAME: ccadmin\n+ OS_PROJECT_NAME: kubernikus-master\n+ OS_PROJECT_DOMAIN_NAME: ccadmin\n+ KUBERNIKUS_NAME: k-master\n+ KUBERNIKUS_URL: https://k-master.admin.cloud.sap\n- - name: kubernikusctl.release\n- type: github-release\n- source:\n- owner: sapcc\n- repository: kubernikus\n- access_token: ((github-com-access-token))\n+auth_feature: &auth_feature\n+ OS_AUTH_URL: https://identity-3.eu-nl-1.cloud.sap/v3\n+ OS_USERNAME: {{kubernikus-eu-nl-1-username}}\n+ OS_PASSWORD: {{kubernikus-eu-nl-1-password}}\n+ OS_USER_DOMAIN_NAME: ccadmin\n+ OS_PROJECT_NAME: kubernikus-feature\n+ OS_PROJECT_DOMAIN_NAME: ccadmin\n+ KUBERNIKUS_NAME: k-feature\n+ KUBERNIKUS_URL: https://k-feature.admin.cloud.sap\n- - name: kubernikus.builds\n+auth_e2e: &auth_e2e\n+ OS_AUTH_URL: https://identity-3.eu-nl-1.cloud.sap/v3\n+ OS_USERNAME: {{kubernikus-eu-nl-1-username}}\n+ OS_PASSWORD: {{kubernikus-eu-nl-1-password}}\n+ OS_USER_DOMAIN_NAME: ccadmin\n+ OS_PROJECT_NAME: kubernikus-e2e\n+ OS_PROJECT_DOMAIN_NAME: ccadmin\n+ KUBERNIKUS_NAME: e2e\n+ KUBERNIKUS_URL: https://kubernikus-master.eu-nl-1.cloud.sap\n+\n+auth_staging: &auth_staging\n+ OS_AUTH_URL: https://identity-3.staging.cloud.sap/v3\n+ OS_USERNAME: {{kubernikus-staging-username}}\n+ OS_PASSWORD: {{kubernikus-staging-password}}\n+ OS_USER_DOMAIN_NAME: ccadmin\n+ OS_PROJECT_NAME: kubernikus\n+ OS_PROJECT_DOMAIN_NAME: ccadmin\n+ KUBERNIKUS_NAME: k-staging\n+ KUBERNIKUS_URL: https://k-staging.admin.cloud.sap\n+\n+auth_eu-nl-1: &auth_eu-nl-1\n+ OS_AUTH_URL: https://identity-3.eu-nl-1.cloud.sap/v3\n+ OS_USERNAME: {{kubernikus-eu-nl-1-username}}\n+ OS_PASSWORD: {{kubernikus-eu-nl-1-password}}\n+ OS_USER_DOMAIN_NAME: ccadmin\n+ OS_PROJECT_NAME: kubernikus\n+ OS_PROJECT_DOMAIN_NAME: ccadmin\n+ KUBERNIKUS_NAME: k-eu-nl-1\n+ KUBERNIKUS_URL: https://k-eu-nl-1.admin.cloud.sap\n+\n+auth_na-us-1: &auth_na-us-1\n+ OS_AUTH_URL: https://identity-3.na-us-1.cloud.sap/v3\n+ OS_USERNAME: {{kubernikus-na-us-1-username}}\n+ OS_PASSWORD: {{kubernikus-na-us-1-password}}\n+ OS_USER_DOMAIN_NAME: ccadmin\n+ OS_PROJECT_NAME: kubernikus\n+ OS_PROJECT_DOMAIN_NAME: ccadmin\n+ KUBERNIKUS_NAME: k-na-us-1\n+ KUBERNIKUS_URL: https://k-na-us-1.admin.cloud.sap\n+\n+resources:\n+\n+ - name: master.builds\ntype: gh-status\nsource:\nusername: sapcc-bot\n@@ -28,27 +81,17 @@ resources:\naccess_token: ((github-com-access-token))\nbranch: master\ndepth: 1\n- ignore_paths: [\"charts\", \"ci\"]\n-\n- - name: charts.git\n- type: git\n- source:\n- username: sapcc-bot\n- password: ((github-com-access-token))\n- uri: https://github.com/sapcc/kubernikus.git\n- branch: master\n- depth: 1\n- paths: [\"charts\"]\n- - name: pipeline.git\n- type: git\n+ - name: feature.builds\n+ type: gh-status\nsource:\nusername: sapcc-bot\npassword: ((github-com-access-token))\n- uri: https://github.com/sapcc/kubernikus.git\n- branch: master\n+ owner: sapcc\n+ repo: kubernikus\n+ access_token: ((github-com-access-token))\n+ branch: feature\ndepth: 1\n- paths: [\"ci\"]\n- name: secrets.git\ntype: git\n@@ -58,10 +101,10 @@ resources:\nbranch: master\ndepth: 1\n- - name: slack-alert\n- type: slack-notification\n- source:\n- url: {{slack-webhook-url}}\n+ - name: daily\n+ type: time\n+ source: {interval: 24h}\n+\nresource_types:\n- name: time-version\n@@ -79,331 +122,191 @@ resource_types:\nsource:\nrepository: hub.global.cloud.sap/concourse/ghstatus-resource\ntag: latest\n-groups:\n- - name: kubernikus\n- jobs:\n- - cli\n- - staging\n- - e2e\n- - rollout\n- - eu-nl-1\n- - na-us-1\n- - name: admin\n- jobs:\n- - admin_staging\n- - rollout_admin\n- - admin_eu-nl-1\n- - admin_na-us-1\n- - name: kubernikus-system\n- jobs:\n- - staging_kubernikus-system\n- - rollout_kubernikus-system\n- - eu-nl-1_kubernikus-system\n- - na-us-1_kubernikus-system\njobs:\n-####################################################################################\n-#\n-# kubernikus\n-#\n-####################################################################################\n-\n- - name: cli\n+ - name: master\n+ serial: true\nplan:\n- aggregate:\n+ - get: secrets.git\n- get: kubernikus.builds\n+ resource: master.builds\ntrigger: true\n- - task: build\n- file: kubernikus.builds/ci/task_cli.yaml\n- - put: kubernikusctl.release\n+ - aggregate:\n+ - task: kubernikus\n+ file: kubernikus.builds/ci/task_helm_kubernikus.yaml\n+ params:\n+ <<: *auth_master\n+ - task: kubernikus-system\n+ file: kubernikus.builds/ci/task_helm_kubernikus-system.yaml\nparams:\n- name: binaries/tag\n- tag: binaries/tag\n- tag_prefix: v\n- commitish: binaries/commitish\n- globs:\n- - binaries/kubernikusctl*\n- get_params:\n- globs: []\n+ <<: *auth_master\n- - name: staging\n+ - name: feature\nserial: true\nplan:\n- aggregate:\n- get: secrets.git\n- - get: pipeline.git\n- get: kubernikus.builds\n+ resource: feature.builds\ntrigger: true\n- - task: deploy\n- file: pipeline.git/ci/task_helm_kubernikus.yaml\n+ - aggregate:\n+ - task: kubernikus\n+ file: kubernikus.builds/ci/task_helm_kubernikus.yaml\nparams:\n- OS_AUTH_URL: https://identity-3.eu-nl-1.cloud.sap/v3\n- OS_USERNAME: {{kubernikus-eu-nl-1-username}}\n- OS_PASSWORD: {{kubernikus-eu-nl-1-password}}\n- OS_USER_DOMAIN_NAME: ccadmin\n- OS_PROJECT_NAME: kubernikus-staging\n- OS_PROJECT_DOMAIN_NAME: ccadmin\n- KUBERNIKUS_NAME: k-staging\n- KUBERNIKUS_URL: https://k-staging.admin.cloud.sap\n- on_failure:\n- <<: *post_failure_to_slack\n+ <<: *auth_feature\n+ - task: kubernikus-system\n+ file: kubernikus.builds/ci/task_helm_kubernikus-system.yaml\n+ params:\n+ <<: *auth_feature\n- name: e2e\nserial: true\nplan:\n- aggregate:\n- - get: pipeline.git\n+ - get: secrets.git\n- get: kubernikus.builds\n- passed: [staging]\n+ resource: master.builds\n+ passed: [master]\ntrigger: true\n- task: e2e_tests\n- file: pipeline.git/ci/task_e2e_tests.yaml\n+ file: kubernikus.builds/ci/task_e2e_tests.yaml\nparams:\n- OS_AUTH_URL: https://identity-3.staging.cloud.sap/v3\n- OS_USERNAME: {{kubernikus-staging-username}}\n- OS_PASSWORD: {{kubernikus-staging-password}}\n- OS_USER_DOMAIN_NAME: ccadmin\n- OS_PROJECT_NAME: kubernikus-e2e\n- OS_PROJECT_DOMAIN_NAME: ccadmin\n- OS_REGION_NAME: staging\n- on_failure:\n- <<: *post_failure_to_slack\n+ <<: *auth_e2e\n- - name: rollout\n+ - name: conformance\nserial: true\nplan:\n+ - aggregate:\n- get: kubernikus.builds\n+ resource: master.builds\npassed: [e2e]\n+ - get: daily\ntrigger: true\n- - name: eu-nl-1\n+ - name: prod\nserial: true\nplan:\n- - aggregate:\n- - get: secrets.git\n- - get: pipeline.git\n- get: kubernikus.builds\n- passed: [rollout]\n- trigger: true\n- - task: deploy\n- file: pipeline.git/ci/task_helm_kubernikus.yaml\n- params:\n- OS_AUTH_URL: https://identity-3.eu-nl-1.cloud.sap/v3\n- OS_USERNAME: {{kubernikus-eu-nl-1-username}}\n- OS_PASSWORD: {{kubernikus-eu-nl-1-password}}\n- OS_USER_DOMAIN_NAME: ccadmin\n- OS_PROJECT_NAME: kubernikus\n- OS_PROJECT_DOMAIN_NAME: ccadmin\n- KUBERNIKUS_NAME: k-eu-nl-1\n- KUBERNIKUS_URL: https://k-eu-nl-1.admin.cloud.sap\n- on_failure:\n- <<: *post_failure_to_slack\n+ resource: master.builds\n+ passed: [e2e]\n- - name: na-us-1\n+ - name: emea\nserial: true\nplan:\n- aggregate:\n- get: secrets.git\n- - get: pipeline.git\n- get: kubernikus.builds\n- passed: [rollout]\n+ resource: master.builds\n+ passed: [prod]\ntrigger: true\n- - task: deploy\n- file: pipeline.git/ci/task_helm_kubernikus.yaml\n+ - aggregate:\n+ - task: kubernikus_eu-nl-1\n+ file: kubernikus.builds/ci/task_helm_kubernikus.yaml\nparams:\n- OS_AUTH_URL: https://identity-3.na-us-1.cloud.sap/v3\n- OS_USERNAME: {{kubernikus-na-us-1-username}}\n- OS_PASSWORD: {{kubernikus-na-us-1-password}}\n- OS_USER_DOMAIN_NAME: ccadmin\n- OS_PROJECT_NAME: kubernikus\n- OS_PROJECT_DOMAIN_NAME: ccadmin\n- KUBERNIKUS_NAME: k-na-us-1\n- KUBERNIKUS_URL: https://k-na-us-1.admin.cloud.sap\n- on_failure:\n- <<: *post_failure_to_slack\n+ <<: *auth_eu-nl-1\n+ - task: kubernikus-system_eu-nl-1\n+ file: kubernikus.builds/ci/task_helm_kubernikus-system.yaml\n+ params:\n+ <<: *auth_eu-nl-1\n-####################################################################################\n-#\n-# admin\n-#\n-####################################################################################\n-#\n- - name: admin_staging\n+ - name: asia\nserial: true\nplan:\n- aggregate:\n- get: secrets.git\n- - get: pipeline.git\n- get: kubernikus.builds\n+ resource: master.builds\n+ passed: [prod]\ntrigger: true\n- - task: deploy\n- file: kubernikus.builds/ci/task_helm-admin_kubernikus.yaml\n- params:\n- REGION: admin\n- KUBERNIKUS_NAME: k-staging\n- GITHUB_TOKEN: ((github-access-token))\n- on_failure:\n- <<: *post_failure_to_slack\n- - name: rollout_admin\n+ - name: americas\nserial: true\nplan:\n+ - aggregate:\n+ - get: secrets.git\n- get: kubernikus.builds\n- passed: [admin_staging]\n+ resource: master.builds\n+ passed: [prod]\ntrigger: true\n+ - aggregate:\n+ - task: kubernikus_na-us-1\n+ file: kubernikus.builds/ci/task_helm_kubernikus.yaml\n+ params:\n+ <<: *auth_na-us-1\n+ - task: kubernikus-system_na-us-1\n+ file: kubernikus.builds/ci/task_helm_kubernikus-system.yaml\n+ params:\n+ <<: *auth_na-us-1\n- - name: admin_eu-nl-1\n+ - name: admin\nserial: true\nplan:\n- aggregate:\n- get: secrets.git\n- - get: pipeline.git\n- get: kubernikus.builds\n- passed: [rollout_admin]\n+ resource: master.builds\n+ passed: [prod]\ntrigger: true\n- - task: deploy\n+ - aggregate:\n+ - task: k-master\n+ file: kubernikus.builds/ci/task_helm-admin_kubernikus.yaml\n+ params:\n+ REGION: admin\n+ KUBERNIKUS_NAME: k-master\n+ GITHUB_TOKEN: ((github-access-token))\n+ - task: k-feature\n+ file: kubernikus.builds/ci/task_helm-admin_kubernikus.yaml\n+ params:\n+ REGION: admin\n+ KUBERNIKUS_NAME: k-feature\n+ GITHUB_TOKEN: ((github-access-token))\n+ - task: k-eu-nl-1\nfile: kubernikus.builds/ci/task_helm-admin_kubernikus.yaml\nparams:\nREGION: admin\nKUBERNIKUS_NAME: k-eu-nl-1\nGITHUB_TOKEN: ((github-access-token))\n- on_failure:\n- <<: *post_failure_to_slack\n-\n- - name: admin_na-us-1\n- serial: true\n- plan:\n- - aggregate:\n- - get: secrets.git\n- - get: pipeline.git\n- - get: kubernikus.builds\n- passed: [rollout_admin]\n- trigger: true\n- - task: deploy\n+ - task: k-na-us-1\nfile: kubernikus.builds/ci/task_helm-admin_kubernikus.yaml\nparams:\nREGION: admin\nKUBERNIKUS_NAME: k-na-us-1\nGITHUB_TOKEN: ((github-access-token))\n- on_failure:\n- <<: *post_failure_to_slack\n-\n-####################################################################################\n-#\n-# kubernikus-system\n-#\n-####################################################################################\n- - name: staging_kubernikus-system\n+ - name: staging\nserial: true\nplan:\n- aggregate:\n- - get: charts.git\n- trigger: true\n- get: secrets.git\n- - get: pipeline.git\n- - task: deploy\n- file: pipeline.git/ci/task_helm_kubernikus-system.yaml\n- params:\n- OS_AUTH_URL: https://identity-3.eu-nl-1.cloud.sap/v3\n- OS_USERNAME: {{kubernikus-eu-nl-1-username}}\n- OS_PASSWORD: {{kubernikus-eu-nl-1-password}}\n- OS_USER_DOMAIN_NAME: ccadmin\n- OS_PROJECT_NAME: kubernikus-staging\n- OS_PROJECT_DOMAIN_NAME: ccadmin\n- KUBERNIKUS_NAME: k-staging\n- KUBERNIKUS_URL: https://k-staging.admin.cloud.sap\n- - task: reload prometheus configuration\n- file: pipeline.git/ci/task_reload_prometheus_config.yaml\n- params:\n- NAMESPACE: kubernikus-system\n- DELAY_SECONDS: 30\n- OS_AUTH_URL: https://identity-3.eu-nl-1.cloud.sap/v3\n- OS_USERNAME: {{kubernikus-eu-nl-1-username}}\n- OS_PASSWORD: {{kubernikus-eu-nl-1-password}}\n- OS_USER_DOMAIN_NAME: ccadmin\n- OS_PROJECT_NAME: kubernikus-staging\n- OS_PROJECT_DOMAIN_NAME: ccadmin\n- KUBERNIKUS_NAME: k-staging\n- KUBERNIKUS_URL: https://k-staging.admin.cloud.sap\n- on_failure:\n- <<: *post_failure_to_slack\n-\n- - name: rollout_kubernikus-system\n- serial: true\n- plan:\n- - get: charts.git\n- passed: [staging_kubernikus-system]\n+ - get: kubernikus.builds\n+ resource: feature.builds\n+ passed: [feature]\ntrigger: true\n-\n- - name: eu-nl-1_kubernikus-system\n- serial: true\n- plan:\n- aggregate:\n- - get: secrets.git\n- - get: pipeline.git\n- - get: charts.git\n- passed: [rollout_kubernikus-system]\n- trigger: true\n- - task: deploy\n- file: pipeline.git/ci/task_helm_kubernikus-system.yaml\n+ - task: k-staging\n+ file: kubernikus.builds/ci/task_helm-admin_kubernikus.yaml\nparams:\n- OS_AUTH_URL: https://identity-3.eu-nl-1.cloud.sap/v3\n- OS_USERNAME: {{kubernikus-eu-nl-1-username}}\n- OS_PASSWORD: {{kubernikus-eu-nl-1-password}}\n- OS_USER_DOMAIN_NAME: ccadmin\n- OS_PROJECT_NAME: kubernikus\n- OS_PROJECT_DOMAIN_NAME: ccadmin\n- KUBERNIKUS_NAME: k-eu-nl-1\n- KUBERNIKUS_URL: https://k-eu-nl-1.admin.cloud.sap\n- - task: reload prometheus configuration\n- file: pipeline.git/ci/task_reload_prometheus_config.yaml\n+ REGION: admin\n+ KUBERNIKUS_NAME: k-staging\n+ GITHUB_TOKEN: ((github-access-token))\n+ - task: kubernikus\n+ file: kubernikus.builds/ci/task_helm_kubernikus.yaml\nparams:\n- NAMESPACE: kubernikus-system\n- DELAY_SECONDS: 30\n- OS_AUTH_URL: https://identity-3.eu-nl-1.cloud.sap/v3\n- OS_USERNAME: {{kubernikus-eu-nl-1-username}}\n- OS_PASSWORD: {{kubernikus-eu-nl-1-password}}\n- OS_USER_DOMAIN_NAME: ccadmin\n- OS_PROJECT_NAME: kubernikus\n- OS_PROJECT_DOMAIN_NAME: ccadmin\n- KUBERNIKUS_NAME: k-eu-nl-1\n- KUBERNIKUS_URL: https://k-eu-nl-1.admin.cloud.sap\n- on_failure:\n- <<: *post_failure_to_slack\n+ <<: *auth_staging\n+ - task: kubernikus-system\n+ file: kubernikus.builds/ci/task_helm_kubernikus-system.yaml\n+ params:\n+ <<: *auth_staging\n- - name: na-us-1_kubernikus-system\n+ - name: qa\nserial: true\nplan:\n- aggregate:\n- get: secrets.git\n- - get: pipeline.git\n- - get: charts.git\n- passed: [rollout_kubernikus-system]\n+ - get: kubernikus.builds\n+ resource: feature.builds\n+ passed: [feature]\ntrigger: true\n- - task: deploy\n- file: pipeline.git/ci/task_helm_kubernikus-system.yaml\n- params:\n- OS_AUTH_URL: https://identity-3.na-us-1.cloud.sap/v3\n- OS_USERNAME: {{kubernikus-na-us-1-username}}\n- OS_PASSWORD: {{kubernikus-na-us-1-password}}\n- OS_USER_DOMAIN_NAME: ccadmin\n- OS_PROJECT_NAME: kubernikus\n- OS_PROJECT_DOMAIN_NAME: ccadmin\n- KUBERNIKUS_NAME: k-na-us-1\n- KUBERNIKUS_URL: https://k-na-us-1.admin.cloud.sap\n- - task: reload prometheus configuration\n- file: pipeline.git/ci/task_reload_prometheus_config.yaml\n- params:\n- NAMESPACE: kubernikus-system\n- DELAY_SECONDS: 30\n- OS_AUTH_URL: https://identity-3.na-us-1.cloud.sap/v3\n- OS_USERNAME: {{kubernikus-na-us-1-username}}\n- OS_PASSWORD: {{kubernikus-na-us-1-password}}\n- OS_USER_DOMAIN_NAME: ccadmin\n- OS_PROJECT_NAME: kubernikus\n- OS_PROJECT_DOMAIN_NAME: ccadmin\n- KUBERNIKUS_NAME: k-na-us-1\n- KUBERNIKUS_URL: https://k-na-us-1.admin.cloud.sap\n- on_failure:\n- <<: *post_failure_to_slack\n+\n+\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
reworked smoother pipeline
|
596,233 |
16.02.2018 17:37:14
| -3,600 |
13e34a3ed133c777bb75cbf527f7f2a4c41b0ce1
|
add alerts
* add alerts for operation error spikes
* add operator hanging alerts
* include all *.rules,*.alerts automatically
* detect goroutine leaks
fixes
|
[
{
"change_type": "MODIFY",
"old_path": "charts/kubernikus-system/charts/prometheus/kubernikus.alerts",
"new_path": "charts/kubernikus-system/charts/prometheus/kubernikus.alerts",
"diff": "@@ -12,3 +12,65 @@ groups:\nannotations:\ndescription: Kluster {{ $labels.kluster_id }} is stuck in {{ $labels.phase }} for 1h\nsummary: Kluster stuck in phase {{ $labels.phase }}\n+\n+- name: operator.alerts\n+ rules:\n+ - alert: KubernikusOperatorGoroutineLeak\n+ expr: sum(rate(go_goroutines{app=\"kubernikus\",type=\"operator\"}[5m])) by (app,type) > sum(avg_over_time(go_goroutines{app=\"kubernikus\",type=\"operator\"}[12h] offset 12h)) by (app,type)\n+ for: 10m\n+ labels:\n+ tier: kubernikus\n+ service: operator\n+ severity: warning\n+ context: operator\n+ annotations:\n+ description: High number of goroutines in kubernikus operator\n+ summary: Goroutine leak in kubernikus operator\n+\n+ - alert: KubernikusLaunchOperationErrorSpike\n+ expr: sum(irate(kubernikus_launch_failed_operation_total[5m])) by (method) > sum(avg_over_time(kubernikus_launch_failed_operation_total[12h] offset 12h)) by (method)\n+ for: 10m\n+ labels:\n+ tier: kubernikus\n+ service: launch\n+ severity: warning\n+ context: operator\n+ annotations:\n+ description: Unusually high amount of failed launchctl operations\n+ summary: Unusually many launchctl failures\n+\n+ - alert: KubernikusLaunchHanging\n+ expr: sum(kubernikus_launch_operation_total) == 0\n+ for: 15m\n+ labels:\n+ tier: kubernikus\n+ service: launchctl\n+ severity: critical\n+ context: operator\n+ annotations:\n+ description: Launchctl operations dropped to 0. The operator might be hanging.\n+ summary: Launchctl operations dropped to 0\n+\n+ - alert: KubernikusRouteGcOperationErrorSpike\n+ expr: sum(irate(kubernikus_routegc_failed_operation_total[5m])) by (method) > sum(avg_over_time(kubernikus_routegc_failed_operation_total[12h] offset 12h)) by (method)\n+ for: 10m\n+ labels:\n+ tier: kubernikus\n+ service: routegc\n+ severity: warning\n+ context: operator\n+ annotations:\n+ description: Unusually high amount of failed routegc operations\n+ summary: Unusually many routegc failures\n+\n+ - alert: KubernikusDeorbiterHanging\n+ expr: sum(kubernikus_deorbit_operation_total) == 0\n+ for: 10m\n+ labels:\n+ tier: kubernikus\n+ service: deorbit\n+ severity: critical\n+ context: operator\n+ annotations:\n+ description: Deorbiter operations dropped to 0. The operator might be hanging.\n+ summary: Deorbiter operations dropped to 0\n"
},
{
"change_type": "MODIFY",
"old_path": "charts/kubernikus-system/charts/prometheus/templates/config.yaml",
"new_path": "charts/kubernikus-system/charts/prometheus/templates/config.yaml",
"diff": "@@ -4,11 +4,18 @@ metadata:\nname: kubernikus-prometheus\ndata:\n- {{- $files := .Files }}\n- {{ range tuple \"aggregation.rules\" \"dns.alerts\" \"kubernetes.alerts\" \"kubernikus.alerts\" \"node.alerts\" }}\n- {{ . }}: |\n-{{ $files.Get . | indent 4 }}\n- {{- end }}\n+ # include all *.rules files\n+ {{- range $path, $bytes := .Files.Glob \"*.rules\" }}\n+ {{ printf \"%s\" $path }}: |\n+{{ printf \"%s\" $bytes | indent 4 }}\n+ {{ end }}\n+\n+ # include all *.alerts files\n+ {{- range $path, $bytes := .Files.Glob \"*.alerts\" }}\n+ {{ printf \"%s\" $path }}: |\n+{{ printf \"%s\" $bytes | indent 4 }}\n+ {{ end }}\n+\nprometheus.yaml: |\n{{ include \"kubernikus-system/charts/prometheus/templates/_prometheus.yaml.tpl\" . | indent 4 }}\n{{- if .Values.extra_scrape_config }}\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
add alerts (#227)
* add alerts for operation error spikes
* add operator hanging alerts
* include all *.rules,*.alerts automatically
* detect goroutine leaks
fixes #218 #219 #220
|
596,240 |
19.02.2018 09:29:24
| -3,600 |
6a1908e4b8abdc38a3b44bf9ce8a771873c2c674
|
retries wget connections - waiting for ASR sync
|
[
{
"change_type": "MODIFY",
"old_path": "test/e2e/constants.go",
"new_path": "test/e2e/constants.go",
"diff": "@@ -17,7 +17,8 @@ const (\nNginxPort = 80\nNamespace = \"default\"\n- TimeoutWGET = 10\n+ WGETRetries = 12\n+ WGETTimeout = 10\nPVCSize = \"1Gi\"\nPVCName = \"e2e-nginx-pvc\"\n"
},
{
"change_type": "MODIFY",
"old_path": "test/e2e/smokeTests.go",
"new_path": "test/e2e/smokeTests.go",
"diff": "@@ -234,7 +234,7 @@ func (s *E2ETestSuite) dialServiceName(source *v1.Pod, target *v1.Service) {\n}\nfunc (s *E2ETestSuite) dial(sourcePod *v1.Pod, targetIP string, targetPort int32) (string, error) {\n- cmd := fmt.Sprintf(\"wget --tries=1 --timeout=%v -O - http://%v:%v\", TimeoutWGET, targetIP, targetPort)\n+ cmd := fmt.Sprintf(\"wget --tries=%v --timeout=%v --retry-connrefused -O - http://%v:%v\", WGETRetries, WGETTimeout, targetIP, targetPort)\nreturn RunKubectlHostCmd(sourcePod.GetNamespace(), sourcePod.GetName(), cmd)\n}\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
retries wget connections - waiting for ASR sync
|
596,240 |
20.02.2018 10:03:01
| -3,600 |
f997cb8a3eff0cb64a19d721ab2074717c247475
|
refactors chart to use global values only
|
[
{
"change_type": "MODIFY",
"old_path": "charts/kubernikus-system/templates/grafana-ingress.yaml",
"new_path": "charts/kubernikus-system/templates/grafana-ingress.yaml",
"diff": "@@ -17,9 +17,9 @@ metadata:\nspec:\ntls:\n- secretName: grafana\n- hosts: [grafana.{{ required \"domain missing\" .Values.domain }}]\n+ hosts: [grafana.{{ required \"domain missing\" .Values.global.domain }}]\nrules:\n- - host: grafana.{{ required \"domain missing\" .Values.domain }}\n+ - host: grafana.{{ required \"domain missing\" .Values.global.domain }}\nhttp:\npaths:\n- path: /\n"
},
{
"change_type": "MODIFY",
"old_path": "charts/kubernikus-system/templates/prometheus-ingress.yaml",
"new_path": "charts/kubernikus-system/templates/prometheus-ingress.yaml",
"diff": "@@ -16,9 +16,9 @@ metadata:\nspec:\ntls:\n- secretName: prometheus\n- hosts: [prometheus.{{ required \"domain missing\" .Values.domain }}]\n+ hosts: [prometheus.{{ required \"domain missing\" .Values.global.domain }}]\nrules:\n- - host: prometheus.{{ required \"domain missing\" .Values.domain }}\n+ - host: prometheus.{{ required \"domain missing\" .Values.global.domain }}\nhttp:\npaths:\n- path: /\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
refactors chart to use global values only
|
596,240 |
20.02.2018 10:03:25
| -3,600 |
b7b76d530ea362a870d84a0d47049033689966de
|
dry up global values
|
[
{
"change_type": "MODIFY",
"old_path": "ci/task_helm-admin_kubernikus.yaml",
"new_path": "ci/task_helm-admin_kubernikus.yaml",
"diff": "@@ -19,7 +19,12 @@ run:\nset -exo pipefail\nVERSION=$(cat kubernikus.builds/.git/HEAD)\nhelm dep up --skip-refresh kubernikus.builds/charts/kubernikus/\n- helm upgrade $KUBERNIKUS_NAME kubernikus.builds/charts/kubernikus/ --namespace=$KUBERNIKUS_NAME --values secrets.git/admin/values/$KUBERNIKUS_NAME.yaml --set imageTag=$VERSION --install\n+ helm upgrade $KUBERNIKUS_NAME kubernikus.builds/charts/kubernikus/ \\\n+ --namespace=$KUBERNIKUS_NAME \\\n+ --values secrets.git/global/values/kubernikus.yaml \\\n+ --values secrets.git/admin/values/$KUBERNIKUS_NAME.yaml \\\n+ --set imageTag=$VERSION \\\n+ --install\nkubectl rollout status deployment/kubernikus-api --namespace=$KUBERNIKUS_NAME\nkubectl rollout status deployment/kubernikus-operator --namespace=$KUBERNIKUS_NAME\n"
},
{
"change_type": "MODIFY",
"old_path": "ci/task_helm_kubernikus-system.yaml",
"new_path": "ci/task_helm_kubernikus-system.yaml",
"diff": "@@ -26,7 +26,11 @@ run:\nhelm repo add forked https://raw.githubusercontent.com/BugRoger/charts/repo\nhelm dep up --skip-refresh kubernikus-copy.builds/charts/kubernikus-system/\n- helm upgrade kubernikus-system kubernikus-copy.builds/charts/kubernikus-system/ --namespace=kubernikus-system --values secrets.git/kubernikus/$KUBERNIKUS_NAME/values/kubernikus-system.yaml --install --force\n+ helm upgrade kubernikus-system kubernikus-copy.builds/charts/kubernikus-system/ \\\n+ --namespace=kubernikus-system\n+ --values secrets.git/global/values/kubernikus-system.yaml\n+ --values secrets.git/kubernikus/$KUBERNIKUS_NAME/values/kubernikus-system.yaml\n+ --install\nparams:\nOS_AUTH_URL:\n"
},
{
"change_type": "MODIFY",
"old_path": "ci/task_helm_kubernikus.yaml",
"new_path": "ci/task_helm_kubernikus.yaml",
"diff": "@@ -20,7 +20,13 @@ run:\nVERSION=$(cat kubernikus.builds/.git/HEAD)\nkubernikusctl auth init\nhelm dep up --skip-refresh kubernikus.builds/charts/kubernikus/\n- helm upgrade kubernikus kubernikus.builds/charts/kubernikus/ --namespace=kubernikus-system --values secrets.git/kubernikus/$KUBERNIKUS_NAME/values/kubernikus.yaml --set imageTag=$VERSION --install --force\n+ helm upgrade kubernikus kubernikus.builds/charts/kubernikus/ \\\n+ --namespace=kubernikus-system \\\n+ --values secrets.git/global/values/kubernikus.yaml \\\n+ --values secrets.git/kubernikus/$KUBERNIKUS_NAME/values/kubernikus.yaml \\\n+ --set imageTag=$VERSION \\\n+ --install \\\n+ --force\nkubectl rollout status deployment/kubernikus-api --namespace=kubernikus-system\nkubectl rollout status deployment/kubernikus-operator --namespace=kubernikus-system\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
dry up global values
|
596,240 |
20.02.2018 10:03:44
| -3,600 |
2f0e87cb6fc5bb653f58a6ccf32f8eb17131cc4c
|
keep e2e klusters in case of errors
|
[
{
"change_type": "MODIFY",
"old_path": "test/e2e/util.go",
"new_path": "test/e2e/util.go",
"diff": "@@ -216,9 +216,9 @@ func (s *E2ETestSuite) handleError(err error) {\n}\nlog.Print(err)\n// cleanup\n- if !s.IsNoTeardown {\n- s.tearDownCluster()\n- }\n+ //if !s.IsNoTeardown {\n+ // s.tearDownCluster()\n+ //}\nos.Exit(1)\n}\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
keep e2e klusters in case of errors
|
596,240 |
20.02.2018 13:31:31
| -3,600 |
11dd9b2f6ab58ae1315864ef8ba5efd1697487f8
|
waits for kube-dns to become ready
|
[
{
"change_type": "MODIFY",
"old_path": "test/e2e/setupSmokeTest.go",
"new_path": "test/e2e/setupSmokeTest.go",
"diff": "@@ -139,6 +139,19 @@ func (s *E2ETestSuite) getReadyNodes() {\ns.readyNodes = nodes.Items\n}\n+func (s *E2ETestSuite) waitForKubeDNS() {\n+ log.Printf(\"waiting for kube-system/kube-dns to become ready\")\n+ dnsPodWatch, err := s.clientSet.CoreV1().Pods(\"kube-system\").Watch(meta_v1.SingleObject(\n+ meta_v1.ObjectMeta{\n+ Name: \"kube-dns\",\n+ },\n+ ))\n+\n+ s.handleError(err)\n+ _, err = watch.Until(TimeoutPod, dnsPodWatch, isPodRunning)\n+ s.handleError(err)\n+}\n+\nfunc (s *E2ETestSuite) getReadyPods() {\npods, err := s.clientSet.CoreV1().Pods(Namespace).List(meta_v1.ListOptions{\nLabelSelector: fmt.Sprintf(\"app=%s\", NginxName),\n"
},
{
"change_type": "MODIFY",
"old_path": "test/e2e/smokeTests.go",
"new_path": "test/e2e/smokeTests.go",
"diff": "@@ -23,6 +23,7 @@ func (s *E2ETestSuite) SetupSmokeTest() {\ns.getReadyNodes()\ns.isClusterBigEnoughForSmokeTest()\ns.cleanUp()\n+ s.waitForKubeDNS()\ns.createPods()\ns.createServices()\n}\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
waits for kube-dns to become ready
|
596,240 |
20.02.2018 14:09:41
| -3,600 |
2c27b972dec9324fb1c427af345576019096ed18
|
wait for endpoints to be ready
|
[
{
"change_type": "MODIFY",
"old_path": "test/e2e/constants.go",
"new_path": "test/e2e/constants.go",
"diff": "@@ -7,6 +7,7 @@ const (\nCheckInterval = 10 * time.Second\nTimeoutPod = 5 * time.Minute\n+ TimeoutKubeDNS = 2 * time.Minute\nClusterName = \"e2e\"\nClusterSmallNodePoolSize = 2\n"
},
{
"change_type": "MODIFY",
"old_path": "test/e2e/setupSmokeTest.go",
"new_path": "test/e2e/setupSmokeTest.go",
"diff": "@@ -141,14 +141,25 @@ func (s *E2ETestSuite) getReadyNodes() {\nfunc (s *E2ETestSuite) waitForKubeDNS() {\nlog.Printf(\"waiting for kube-system/kube-dns to become ready\")\n- dnsPodWatch, err := s.clientSet.CoreV1().Pods(\"kube-system\").Watch(meta_v1.SingleObject(\n+ endpointWatch, err := s.clientSet.CoreV1().Endpoints(\"kube-system\").Watch(meta_v1.SingleObject(\nmeta_v1.ObjectMeta{\nName: \"kube-dns\",\n},\n))\ns.handleError(err)\n- _, err = watch.Until(TimeoutPod, dnsPodWatch, isPodRunning)\n+ _, err = watch.Until(TimeoutKubeDNS, endpointWatch, func(event watch.Event) (bool, error) {\n+ if endpoints, ok := event.Object.(*v1.Endpoints); ok {\n+ ready := 0\n+ for _, sub := range endpoints.Subsets {\n+ if len(sub.Addresses) > 0 {\n+ ready++\n+ }\n+ }\n+ return ready == len(endpoints.Subsets), nil\n+ }\n+ return false, nil\n+ })\ns.handleError(err)\n}\n"
},
{
"change_type": "MODIFY",
"old_path": "test/e2e/util.go",
"new_path": "test/e2e/util.go",
"diff": "@@ -173,6 +173,8 @@ func newE2ESmokeTestCluster(klusterName string) *models.Kluster {\n}\n}\n+\n+\nfunc isPodRunning(event watch.Event) (bool, error) {\nswitch event.Type {\ncase watch.Deleted:\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
wait for endpoints to be ready
|
596,240 |
20.02.2018 16:28:56
| -3,600 |
6c15c6ae20a9819edcdf0332a124657f2f64a6ff
|
adds ssh key to e2e cluster
|
[
{
"change_type": "MODIFY",
"old_path": "test/e2e/util.go",
"new_path": "test/e2e/util.go",
"diff": "@@ -130,6 +130,7 @@ func newE2ECluster(klusterName string) *models.Kluster {\nreturn &models.Kluster{\nName: klusterName,\nSpec: models.KlusterSpec{\n+ SSHPublicKey: \"ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAACAQCXIxVEUgtUVkvk2VM1hmIb8MxvxsmvYoiq9OBy3J8akTGNybqKsA2uhcwxSJX5Cn3si8kfMfka9EWiJT+e1ybvtsGILO5XRZPxyhYzexwb3TcALwc3LuzpF3Z/Dg2jYTRELTGhYmyca3mxzTlCjNXvYayLNedjJ8fIBzoCuSXNqDRToHru7h0Glz+wtuE74mNkOiXSvhtuJtJs7VCNVjobFQNfC1aeDsri2bPRHJJZJ0QF4LLYSayMEz3lVwIDyAviQR2Aa97WfuXiofiAemfGqiH47Kq6b8X7j3bOYGBvJKMUV7XeWhGsskAmTsvvnFxkc5PAD3Ct+liULjiQWlzDrmpTE8aMqLK4l0YQw7/8iRVz6gli42iEc2ZG56ob1ErpTLAKFWyCNOebZuGoygdEQaGTIIunAncXg5Rz07TdPl0Tf5ZZLpiAgR5ck0H1SETnjDTZ/S83CiVZWJgmCpu8YOKWyYRD4orWwdnA77L4+ixeojLIhEoNL8KlBgsP9Twx+fFMWLfxMmiuX+yksM6Hu+Lsm+Ao7Q284VPp36EB1rxP1JM7HCiEOEm50Jb6hNKjgN4aoLhG5yg+GnDhwCZqUwcRJo1bWtm3QvRA+rzrGZkId4EY3cyOK5QnYV5+24x93Ex0UspHMn7HGsHUESsVeV0fLqlfXyd2RbHTmDMP6w== Kubernikus Master Key\",\nNodePools: []models.NodePool{\n{\nName: \"small\",\n@@ -173,8 +174,6 @@ func newE2ESmokeTestCluster(klusterName string) *models.Kluster {\n}\n}\n-\n-\nfunc isPodRunning(event watch.Event) (bool, error) {\nswitch event.Type {\ncase watch.Deleted:\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
adds ssh key to e2e cluster
|
596,240 |
20.02.2018 16:29:05
| -3,600 |
eac8d792595916d1a3fdd5bbc89dd6c5dd526ee2
|
updates dashboard
|
[
{
"change_type": "MODIFY",
"old_path": "charts/kubernikus-system/values.yaml",
"new_path": "charts/kubernikus-system/values.yaml",
"diff": "@@ -134,7 +134,7 @@ grafana:\n\"editable\": true,\n\"gnetId\": null,\n\"graphTooltip\": 0,\n- \"iteration\": 1519055723950,\n+ \"iteration\": 1519059007733,\n\"links\": [],\n\"panels\": [\n{\n@@ -599,7 +599,6 @@ grafana:\n\"steppedLine\": false,\n\"targets\": [\n{\n- \"$$hashKey\": \"object:855\",\n\"expr\": \"sum(probe_http_duration_seconds{instance=~\\\"https://kubernikus.*\\\",path=\\\"/\\\"}) \",\n\"format\": \"time_series\",\n\"hide\": false,\n@@ -610,7 +609,6 @@ grafana:\n\"refId\": \"A\"\n},\n{\n- \"$$hashKey\": \"object:856\",\n\"expr\": \"quantile_over_time(0.95, probe_duration_seconds{instance=~\\\"https://kubernikus.*\\\"}[24h]) \",\n\"format\": \"time_series\",\n\"hide\": false,\n@@ -640,7 +638,6 @@ grafana:\n},\n\"yaxes\": [\n{\n- \"$$hashKey\": \"object:2124\",\n\"decimals\": null,\n\"format\": \"s\",\n\"label\": null,\n@@ -650,7 +647,6 @@ grafana:\n\"show\": true\n},\n{\n- \"$$hashKey\": \"object:2125\",\n\"format\": \"short\",\n\"label\": null,\n\"logBase\": 1,\n@@ -695,7 +691,6 @@ grafana:\n\"renderer\": \"flot\",\n\"seriesOverrides\": [\n{\n- \"$$hashKey\": \"object:1960\",\n\"alias\": \"Avg\",\n\"dashes\": true,\n\"fill\": 0,\n@@ -707,7 +702,6 @@ grafana:\n\"steppedLine\": false,\n\"targets\": [\n{\n- \"$$hashKey\": \"object:772\",\n\"expr\": \"avg(probe_http_duration_seconds{path=\\\"/\\\"})\",\n\"format\": \"time_series\",\n\"instant\": false,\n@@ -716,7 +710,6 @@ grafana:\n\"refId\": \"A\"\n},\n{\n- \"$$hashKey\": \"object:773\",\n\"expr\": \"avg(quantile_over_time(0.5, probe_duration_seconds{path=\\\"/\\\"}[24h]))\",\n\"format\": \"time_series\",\n\"interval\": \"\",\n@@ -745,7 +738,6 @@ grafana:\n},\n\"yaxes\": [\n{\n- \"$$hashKey\": \"object:1897\",\n\"decimals\": null,\n\"format\": \"s\",\n\"label\": null,\n@@ -755,7 +747,6 @@ grafana:\n\"show\": true\n},\n{\n- \"$$hashKey\": \"object:1898\",\n\"format\": \"short\",\n\"label\": null,\n\"logBase\": 1,\n@@ -769,17 +760,14 @@ grafana:\n\"backgroundColor\": \"rgb(60, 60, 60)\",\n\"colorMaps\": [\n{\n- \"$$hashKey\": \"object:1531\",\n\"color\": \"#e5ac0e\",\n\"text\": \"N/A\"\n},\n{\n- \"$$hashKey\": \"object:1532\",\n\"color\": \"#508642\",\n\"text\": \"up\"\n},\n{\n- \"$$hashKey\": \"object:1533\",\n\"color\": \"#bf1b00\",\n\"text\": \"down\"\n}\n@@ -812,7 +800,6 @@ grafana:\n\"metricNameColor\": \"#fff\",\n\"rangeMaps\": [\n{\n- \"$$hashKey\": \"object:1550\",\n\"from\": \"null\",\n\"text\": \"N/A\",\n\"to\": \"null\"\n@@ -829,7 +816,6 @@ grafana:\n\"showTransitionCount\": false,\n\"targets\": [\n{\n- \"$$hashKey\": \"object:967\",\n\"expr\": \"probe_success{path=\\\"/\\\"}\",\n\"format\": \"time_series\",\n\"hide\": false,\n@@ -847,25 +833,21 @@ grafana:\n\"type\": \"natel-discrete-panel\",\n\"valueMaps\": [\n{\n- \"$$hashKey\": \"object:1542\",\n\"op\": \"=\",\n\"text\": \"N/A\",\n\"value\": \"null\"\n},\n{\n- \"$$hashKey\": \"object:1543\",\n\"op\": \"=\",\n\"text\": \"up\",\n\"value\": \"1\"\n},\n{\n- \"$$hashKey\": \"object:1544\",\n\"op\": \"=\",\n\"text\": \"down\",\n\"value\": \"0\"\n},\n{\n- \"$$hashKey\": \"object:1545\",\n\"op\": \"=\",\n\"text\": \"\",\n\"value\": \"\"\n@@ -2432,10 +2414,12 @@ grafana:\n\"mappingType\": 1,\n\"mappingTypes\": [\n{\n+ \"$$hashKey\": \"object:684\",\n\"name\": \"value to text\",\n\"value\": 1\n},\n{\n+ \"$$hashKey\": \"object:685\",\n\"name\": \"range to text\",\n\"value\": 2\n}\n@@ -2463,9 +2447,10 @@ grafana:\n\"tableColumn\": \"\",\n\"targets\": [\n{\n- \"$$hashKey\": \"object:159\",\n+ \"$$hashKey\": \"object:661\",\n\"expr\": \"kubernikus_kluster_boot_duration_sum / kubernikus_kluster_boot_duration_count\",\n\"format\": \"time_series\",\n+ \"instant\": false,\n\"intervalFactor\": 1,\n\"legendFormat\": \"Boot Time\",\n\"refId\": \"A\"\n@@ -2477,6 +2462,7 @@ grafana:\n\"valueFontSize\": \"80%\",\n\"valueMaps\": [\n{\n+ \"$$hashKey\": \"object:687\",\n\"op\": \"=\",\n\"text\": \"N/A\",\n\"value\": \"null\"\n@@ -2735,12 +2721,13 @@ grafana:\n\"current\": false,\n\"max\": false,\n\"min\": false,\n+ \"rightSide\": false,\n\"show\": false,\n\"total\": false,\n\"values\": false\n},\n\"lines\": true,\n- \"linewidth\": 2,\n+ \"linewidth\": 3,\n\"links\": [],\n\"nullPointMode\": \"connected\",\n\"percentage\": false,\n@@ -2759,14 +2746,16 @@ grafana:\n\"steppedLine\": false,\n\"targets\": [\n{\n- \"expr\": \"sum(rate(go_goroutines{app=\\\"kubernikus\\\"}[5m]))\",\n+ \"$$hashKey\": \"object:444\",\n+ \"expr\": \"sum(go_goroutines{app=\\\"kubernikus\\\"})\",\n\"format\": \"time_series\",\n\"intervalFactor\": 1,\n- \"legendFormat\": \"Current\",\n+ \"legendFormat\": \"\",\n\"refId\": \"D\"\n},\n{\n- \"expr\": \"sum(rate(go_goroutines{app=\\\"kubernikus\\\"}[24h]))\\n\",\n+ \"$$hashKey\": \"object:445\",\n+ \"expr\": \"go_rout\",\n\"format\": \"time_series\",\n\"intervalFactor\": 1,\n\"legendFormat\": \"24h Average\",\n@@ -2792,6 +2781,7 @@ grafana:\n},\n\"yaxes\": [\n{\n+ \"$$hashKey\": \"object:517\",\n\"decimals\": null,\n\"format\": \"short\",\n\"label\": null,\n@@ -2801,6 +2791,7 @@ grafana:\n\"show\": true\n},\n{\n+ \"$$hashKey\": \"object:518\",\n\"format\": \"short\",\n\"label\": null,\n\"logBase\": 1,\n@@ -2943,5 +2934,5 @@ grafana:\n},\n\"timezone\": \"\",\n\"title\": \"Kubernikus\",\n- \"version\": 6\n+ \"version\": 7\n}\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
updates dashboard
|
596,240 |
20.02.2018 17:08:32
| -3,600 |
5a7f57de3499894e681348447a8c65f756c3699b
|
dries up admin values even more
|
[
{
"change_type": "MODIFY",
"old_path": "ci/task_helm-admin_kubernikus.yaml",
"new_path": "ci/task_helm-admin_kubernikus.yaml",
"diff": "@@ -22,6 +22,7 @@ run:\nhelm upgrade $KUBERNIKUS_NAME kubernikus.builds/charts/kubernikus/ \\\n--namespace=$KUBERNIKUS_NAME \\\n--values secrets.git/global/values/kubernikus.yaml \\\n+ --values secrets.git/kubernikus/$KUBERNIKUS_NAME/values/kubernikus.yaml \\\n--values secrets.git/admin/values/$KUBERNIKUS_NAME.yaml \\\n--set imageTag=$VERSION \\\n--install\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
dries up admin values even more
|
596,240 |
21.02.2018 09:37:35
| -3,600 |
f28969545da54c787c64d29c86cf0f1ab4d2f73a
|
updated pipeline with additional regions
|
[
{
"change_type": "MODIFY",
"old_path": "ci/pipeline.yaml",
"new_path": "ci/pipeline.yaml",
"diff": "@@ -11,8 +11,8 @@ post_failure_to_slack: &post_failure_to_slack\nauth_master: &auth_master\nOS_AUTH_URL: https://identity-3.eu-nl-1.cloud.sap/v3\n- OS_USERNAME: {{kubernikus-eu-nl-1-username}}\n- OS_PASSWORD: {{kubernikus-eu-nl-1-password}}\n+ OS_USERNAME: {{kubernikus-prod-username}}\n+ OS_PASSWORD: {{kubernikus-prod-password}}\nOS_USER_DOMAIN_NAME: ccadmin\nOS_PROJECT_NAME: kubernikus-master\nOS_PROJECT_DOMAIN_NAME: ccadmin\n@@ -21,8 +21,8 @@ auth_master: &auth_master\nauth_feature: &auth_feature\nOS_AUTH_URL: https://identity-3.eu-nl-1.cloud.sap/v3\n- OS_USERNAME: {{kubernikus-eu-nl-1-username}}\n- OS_PASSWORD: {{kubernikus-eu-nl-1-password}}\n+ OS_USERNAME: {{kubernikus-prod-username}}\n+ OS_PASSWORD: {{kubernikus-prod-password}}\nOS_USER_DOMAIN_NAME: ccadmin\nOS_PROJECT_NAME: kubernikus-feature\nOS_PROJECT_DOMAIN_NAME: ccadmin\n@@ -31,8 +31,8 @@ auth_feature: &auth_feature\nauth_e2e: &auth_e2e\nOS_AUTH_URL: https://identity-3.eu-nl-1.cloud.sap/v3\n- OS_USERNAME: {{kubernikus-eu-nl-1-username}}\n- OS_PASSWORD: {{kubernikus-eu-nl-1-password}}\n+ OS_USERNAME: {{kubernikus-prod-username}}\n+ OS_PASSWORD: {{kubernikus-prod-password}}\nOS_USER_DOMAIN_NAME: ccadmin\nOS_PROJECT_NAME: kubernikus-e2e\nOS_PROJECT_DOMAIN_NAME: ccadmin\n@@ -49,10 +49,40 @@ auth_staging: &auth_staging\nKUBERNIKUS_NAME: k-staging\nKUBERNIKUS_URL: https://k-staging.admin.cloud.sap\n+auth_qa-de-1: &auth_qa-de-1\n+ OS_AUTH_URL: https://identity-3.qa-de-1.cloud.sap/v3\n+ OS_USERNAME: {{kubernikus-qa-username}}\n+ OS_PASSWORD: {{kubernikus-qa-password}}\n+ OS_USER_DOMAIN_NAME: ccadmin\n+ OS_PROJECT_NAME: kubernikus\n+ OS_PROJECT_DOMAIN_NAME: ccadmin\n+ KUBERNIKUS_NAME: k-qa-de-1\n+ KUBERNIKUS_URL: https://k-qa-de-1.admin.cloud.sap\n+\n+auth_ap-au-1: &auth_ap-au-1\n+ OS_AUTH_URL: https://identity-3.ap-au-1.cloud.sap/v3\n+ OS_USERNAME: {{kubernikus-prod-username}}\n+ OS_PASSWORD: {{kubernikus-prod-password}}\n+ OS_USER_DOMAIN_NAME: ccadmin\n+ OS_PROJECT_NAME: kubernikus\n+ OS_PROJECT_DOMAIN_NAME: ccadmin\n+ KUBERNIKUS_NAME: k-ap-au-1\n+ KUBERNIKUS_URL: https://k-ap-au-1.admin.cloud.sap\n+\n+auth_eu-de-1: &auth_eu-de-1\n+ OS_AUTH_URL: https://identity-3.eu-de-1.cloud.sap/v3\n+ OS_USERNAME: {{kubernikus-prod-username}}\n+ OS_PASSWORD: {{kubernikus-prod-password}}\n+ OS_USER_DOMAIN_NAME: ccadmin\n+ OS_PROJECT_NAME: kubernikus\n+ OS_PROJECT_DOMAIN_NAME: ccadmin\n+ KUBERNIKUS_NAME: k-eu-de-1\n+ KUBERNIKUS_URL: https://k-eu-de-1.admin.cloud.sap\n+\nauth_eu-nl-1: &auth_eu-nl-1\nOS_AUTH_URL: https://identity-3.eu-nl-1.cloud.sap/v3\n- OS_USERNAME: {{kubernikus-eu-nl-1-username}}\n- OS_PASSWORD: {{kubernikus-eu-nl-1-password}}\n+ OS_USERNAME: {{kubernikus-prod-username}}\n+ OS_PASSWORD: {{kubernikus-prod-password}}\nOS_USER_DOMAIN_NAME: ccadmin\nOS_PROJECT_NAME: kubernikus\nOS_PROJECT_DOMAIN_NAME: ccadmin\n@@ -61,8 +91,8 @@ auth_eu-nl-1: &auth_eu-nl-1\nauth_na-us-1: &auth_na-us-1\nOS_AUTH_URL: https://identity-3.na-us-1.cloud.sap/v3\n- OS_USERNAME: {{kubernikus-na-us-1-username}}\n- OS_PASSWORD: {{kubernikus-na-us-1-password}}\n+ OS_USERNAME: {{kubernikus-prod-username}}\n+ OS_PASSWORD: {{kubernikus-prod-password}}\nOS_USER_DOMAIN_NAME: ccadmin\nOS_PROJECT_NAME: kubernikus\nOS_PROJECT_DOMAIN_NAME: ccadmin\n@@ -201,10 +231,18 @@ jobs:\npassed: [prod]\ntrigger: true\n- aggregate:\n+ - task: kubernikus_eu-de-1\n+ file: kubernikus.builds/ci/task_helm_kubernikus.yaml\n+ params:\n+ <<: *auth_eu-de-1\n- task: kubernikus_eu-nl-1\nfile: kubernikus.builds/ci/task_helm_kubernikus.yaml\nparams:\n<<: *auth_eu-nl-1\n+ - task: kubernikus-system_eu-de-1\n+ file: kubernikus.builds/ci/task_helm_kubernikus-system.yaml\n+ params:\n+ <<: *auth_eu-de-1\n- task: kubernikus-system_eu-nl-1\nfile: kubernikus.builds/ci/task_helm_kubernikus-system.yaml\nparams:\n@@ -219,6 +257,15 @@ jobs:\nresource: master.builds\npassed: [prod]\ntrigger: true\n+ - aggregate:\n+ - task: kubernikus_ap-au-1\n+ file: kubernikus.builds/ci/task_helm_kubernikus.yaml\n+ params:\n+ <<: *auth_ap-au-1\n+ - task: kubernikus-system_ap-au-1\n+ file: kubernikus.builds/ci/task_helm_kubernikus-system.yaml\n+ params:\n+ <<: *auth_ap-au-1\n- name: americas\nserial: true\n@@ -261,6 +308,18 @@ jobs:\nREGION: admin\nKUBERNIKUS_NAME: k-feature\nGITHUB_TOKEN: ((github-access-token))\n+ - task: k-ap-au-1\n+ file: kubernikus.builds/ci/task_helm-admin_kubernikus.yaml\n+ params:\n+ REGION: admin\n+ KUBERNIKUS_NAME: k-ap-au-1\n+ GITHUB_TOKEN: ((github-access-token))\n+ - task: k-eu-de-1\n+ file: kubernikus.builds/ci/task_helm-admin_kubernikus.yaml\n+ params:\n+ REGION: admin\n+ KUBERNIKUS_NAME: k-eu-de-1\n+ GITHUB_TOKEN: ((github-access-token))\n- task: k-eu-nl-1\nfile: kubernikus.builds/ci/task_helm-admin_kubernikus.yaml\nparams:\n@@ -308,5 +367,18 @@ jobs:\nresource: feature.builds\npassed: [feature]\ntrigger: true\n-\n-\n+ - aggregate:\n+ - task: k-qa-de-1\n+ file: kubernikus.builds/ci/task_helm-admin_kubernikus.yaml\n+ params:\n+ REGION: admin\n+ KUBERNIKUS_NAME: k-qa-de-1\n+ GITHUB_TOKEN: ((github-access-token))\n+ - task: kubernikus\n+ file: kubernikus.builds/ci/task_helm_kubernikus.yaml\n+ params:\n+ <<: *auth_qa-de-1\n+ - task: kubernikus-system\n+ file: kubernikus.builds/ci/task_helm_kubernikus-system.yaml\n+ params:\n+ <<: *auth_qa-de-1\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
updated pipeline with additional regions
|
596,240 |
21.02.2018 11:09:40
| -3,600 |
eb0b588aeba765eace72b93dae3c93f4b18643d9
|
rename to avoid conflicts
|
[
{
"change_type": "MODIFY",
"old_path": "charts/kubernikus/templates/clusterrolebinding.yaml",
"new_path": "charts/kubernikus/templates/clusterrolebinding.yaml",
"diff": "apiVersion: rbac.authorization.k8s.io/v1beta1\nkind: ClusterRoleBinding\nmetadata:\n- name: kubernikus-system-default\n+ name: kubernikus-default\nroleRef:\napiGroup: rbac.authorization.k8s.io\nkind: ClusterRole\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
rename to avoid conflicts
|
596,240 |
21.02.2018 11:11:11
| -3,600 |
8e0aae27e57647e2b257ca148e3f07a8e26a714a
|
removes forceful helm installation
|
[
{
"change_type": "MODIFY",
"old_path": "ci/task_helm_kubernikus.yaml",
"new_path": "ci/task_helm_kubernikus.yaml",
"diff": "@@ -25,8 +25,7 @@ run:\n--values secrets.git/global/values/kubernikus.yaml \\\n--values secrets.git/kubernikus/$KUBERNIKUS_NAME/values/kubernikus.yaml \\\n--set imageTag=$VERSION \\\n- --install \\\n- --force\n+ --install\nkubectl rollout status deployment/kubernikus-api --namespace=kubernikus-system\nkubectl rollout status deployment/kubernikus-operator --namespace=kubernikus-system\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
removes forceful helm installation
|
596,240 |
21.02.2018 15:52:50
| -3,600 |
3e03006b44cc250d75cfb9a6f07a061fb7aba094
|
adds usage plots
|
[
{
"change_type": "MODIFY",
"old_path": "charts/kubernikus-system/values.yaml",
"new_path": "charts/kubernikus-system/values.yaml",
"diff": "@@ -131,10 +131,10 @@ grafana:\n}\n]\n},\n- \"editable\": true,\n+ \"editable\": false,\n\"gnetId\": null,\n\"graphTooltip\": 0,\n- \"iteration\": 1519059007733,\n+ \"iteration\": 1519221161829,\n\"links\": [],\n\"panels\": [\n{\n@@ -1761,7 +1761,7 @@ grafana:\n\"x\": 7,\n\"y\": 9\n},\n- \"id\": 42,\n+ \"id\": 73,\n\"legend\": {\n\"alignAsTable\": false,\n\"avg\": false,\n@@ -1773,38 +1773,31 @@ grafana:\n\"values\": false\n},\n\"lines\": true,\n- \"linewidth\": 2,\n+ \"linewidth\": 3,\n\"links\": [],\n\"nullPointMode\": \"connected\",\n- \"percentage\": false,\n+ \"percentage\": true,\n\"pointradius\": 5,\n\"points\": false,\n\"renderer\": \"flot\",\n\"seriesOverrides\": [],\n\"spaceLength\": 10,\n- \"stack\": true,\n+ \"stack\": false,\n\"steppedLine\": false,\n\"targets\": [\n{\n- \"expr\": \"(sum(node_memory_MemTotal) by (instance) - sum(node_memory_MemFree+node_memory_Buffers+node_memory_Cached) by (instance)) / (sum(node_memory_MemTotal) by (instance)) * 100\",\n+ \"expr\": \"sum(irate(node_cpu{mode=~\\\"system|user|iowait\\\"}[5m])) by (instance) / sum(irate(node_cpu{}[5m])) by (instance)\",\n\"format\": \"time_series\",\n\"interval\": \"\",\n- \"intervalFactor\": 1,\n+ \"intervalFactor\": 3,\n\"legendFormat\": \"{{instance}}\",\n\"refId\": \"A\"\n- },\n- {\n- \"expr\": \"\",\n- \"format\": \"time_series\",\n- \"intervalFactor\": 1,\n- \"legendFormat\": \"\",\n- \"refId\": \"B\"\n}\n],\n\"thresholds\": [],\n\"timeFrom\": null,\n\"timeShift\": null,\n- \"title\": \"Memory\",\n+ \"title\": \"CPU\",\n\"tooltip\": {\n\"shared\": true,\n\"sort\": 0,\n@@ -1821,11 +1814,11 @@ grafana:\n\"yaxes\": [\n{\n\"decimals\": null,\n- \"format\": \"percent\",\n+ \"format\": \"percentunit\",\n\"label\": \"\",\n\"logBase\": 1,\n- \"max\": \"400\",\n- \"min\": \"0.0\",\n+ \"max\": \"1\",\n+ \"min\": \"0\",\n\"show\": false\n},\n{\n@@ -2404,7 +2397,7 @@ grafana:\n},\n\"gridPos\": {\n\"h\": 7,\n- \"w\": 6,\n+ \"w\": 3,\n\"x\": 0,\n\"y\": 13\n},\n@@ -2470,6 +2463,193 @@ grafana:\n],\n\"valueName\": \"avg\"\n},\n+ {\n+ \"aliasColors\": {},\n+ \"bars\": false,\n+ \"dashLength\": 10,\n+ \"dashes\": false,\n+ \"datasource\": null,\n+ \"decimals\": 0,\n+ \"fill\": 1,\n+ \"gridPos\": {\n+ \"h\": 7,\n+ \"w\": 4,\n+ \"x\": 3,\n+ \"y\": 13\n+ },\n+ \"id\": 72,\n+ \"legend\": {\n+ \"alignAsTable\": false,\n+ \"avg\": false,\n+ \"current\": false,\n+ \"max\": false,\n+ \"min\": false,\n+ \"rightSide\": false,\n+ \"show\": false,\n+ \"total\": false,\n+ \"values\": false\n+ },\n+ \"lines\": true,\n+ \"linewidth\": 3,\n+ \"links\": [],\n+ \"nullPointMode\": \"connected\",\n+ \"percentage\": false,\n+ \"pointradius\": 5,\n+ \"points\": false,\n+ \"renderer\": \"flot\",\n+ \"seriesOverrides\": [],\n+ \"spaceLength\": 10,\n+ \"stack\": false,\n+ \"steppedLine\": false,\n+ \"targets\": [\n+ {\n+ \"$$hashKey\": \"object:506\",\n+ \"expr\": \"sum(label_replace(rate(container_cpu_usage_seconds_total{namespace=\\\"kubernikus\\\",name=~\\\".*(etcd|manager|scheduler|apiserver).*\\\",image!~\\\".*pause.*\\\"}[1h]), \\\"container\\\", \\\"$1\\\", \\\"name\\\", \\\".*(etcd|manager|scheduler|apiserver).*\\\")) by (container) / sum(label_replace(kube_pod_container_resource_requests_cpu_cores{container=~\\\"apiserver|etcd|scheduler|controller-manager\\\"}, \\\"container\\\", \\\"$1\\\", \\\"container\\\", \\\".*(etcd|manager|scheduler|apiserver)\\\")) by (container)\",\n+ \"format\": \"time_series\",\n+ \"hide\": false,\n+ \"instant\": false,\n+ \"intervalFactor\": 10,\n+ \"legendFormat\": \"{{container}}\",\n+ \"refId\": \"B\"\n+ },\n+ {\n+ \"$$hashKey\": \"object:870\",\n+ \"expr\": \"\",\n+ \"format\": \"time_series\",\n+ \"hide\": false,\n+ \"intervalFactor\": 1,\n+ \"legendFormat\": \"\",\n+ \"refId\": \"C\"\n+ }\n+ ],\n+ \"thresholds\": [],\n+ \"timeFrom\": null,\n+ \"timeShift\": null,\n+ \"title\": \"CPU Usage Ratio\",\n+ \"tooltip\": {\n+ \"shared\": true,\n+ \"sort\": 2,\n+ \"value_type\": \"individual\"\n+ },\n+ \"type\": \"graph\",\n+ \"xaxis\": {\n+ \"buckets\": null,\n+ \"mode\": \"time\",\n+ \"name\": null,\n+ \"show\": false,\n+ \"values\": []\n+ },\n+ \"yaxes\": [\n+ {\n+ \"$$hashKey\": \"object:531\",\n+ \"decimals\": 1,\n+ \"format\": \"percentunit\",\n+ \"label\": \"\",\n+ \"logBase\": 1,\n+ \"max\": null,\n+ \"min\": \"0\",\n+ \"show\": true\n+ },\n+ {\n+ \"$$hashKey\": \"object:532\",\n+ \"format\": \"short\",\n+ \"label\": null,\n+ \"logBase\": 1,\n+ \"max\": null,\n+ \"min\": null,\n+ \"show\": false\n+ }\n+ ]\n+ },\n+ {\n+ \"aliasColors\": {},\n+ \"bars\": false,\n+ \"dashLength\": 10,\n+ \"dashes\": false,\n+ \"datasource\": null,\n+ \"decimals\": 0,\n+ \"fill\": 1,\n+ \"gridPos\": {\n+ \"h\": 7,\n+ \"w\": 4,\n+ \"x\": 7,\n+ \"y\": 13\n+ },\n+ \"id\": 42,\n+ \"legend\": {\n+ \"alignAsTable\": false,\n+ \"avg\": false,\n+ \"current\": false,\n+ \"max\": false,\n+ \"min\": false,\n+ \"rightSide\": false,\n+ \"show\": false,\n+ \"total\": false,\n+ \"values\": false\n+ },\n+ \"lines\": true,\n+ \"linewidth\": 2,\n+ \"links\": [],\n+ \"nullPointMode\": \"connected\",\n+ \"percentage\": false,\n+ \"pointradius\": 5,\n+ \"points\": false,\n+ \"renderer\": \"flot\",\n+ \"seriesOverrides\": [],\n+ \"spaceLength\": 10,\n+ \"stack\": false,\n+ \"steppedLine\": false,\n+ \"targets\": [\n+ {\n+ \"$$hashKey\": \"object:1526\",\n+ \"expr\": \"sum(label_replace(container_memory_usage_bytes{namespace=\\\"kubernikus\\\",name=~\\\".*(etcd|manager|scheduler|apiserver).*\\\",image!~\\\".*pause.*\\\"}, \\\"container\\\", \\\"$1\\\", \\\"name\\\", \\\".*(etcd|manager|scheduler|apiserver).*\\\")) by (container) / sum(label_replace(kube_pod_container_resource_requests_memory_bytes{container=~\\\"apiserver|etcd|scheduler|controller-manager\\\"}, \\\"container\\\", \\\"$1\\\", \\\"container\\\", \\\".*(etcd|manager|scheduler|apiserver)\\\")) by (container)\",\n+ \"format\": \"time_series\",\n+ \"hide\": false,\n+ \"interval\": \"\",\n+ \"intervalFactor\": 10,\n+ \"legendFormat\": \"{{container}}\",\n+ \"refId\": \"A\"\n+ }\n+ ],\n+ \"thresholds\": [],\n+ \"timeFrom\": null,\n+ \"timeShift\": null,\n+ \"title\": \"Memory Usage Ratio\",\n+ \"tooltip\": {\n+ \"shared\": true,\n+ \"sort\": 2,\n+ \"value_type\": \"individual\"\n+ },\n+ \"type\": \"graph\",\n+ \"xaxis\": {\n+ \"buckets\": null,\n+ \"mode\": \"time\",\n+ \"name\": null,\n+ \"show\": false,\n+ \"values\": []\n+ },\n+ \"yaxes\": [\n+ {\n+ \"$$hashKey\": \"object:1887\",\n+ \"decimals\": null,\n+ \"format\": \"percentunit\",\n+ \"label\": \"\",\n+ \"logBase\": 1,\n+ \"max\": null,\n+ \"min\": null,\n+ \"show\": true\n+ },\n+ {\n+ \"$$hashKey\": \"object:1888\",\n+ \"format\": \"short\",\n+ \"label\": null,\n+ \"logBase\": 1,\n+ \"max\": null,\n+ \"min\": null,\n+ \"show\": false\n+ }\n+ ]\n+ },\n{\n\"aliasColors\": {},\n\"bars\": false,\n@@ -2479,8 +2659,8 @@ grafana:\n\"fill\": 0,\n\"gridPos\": {\n\"h\": 7,\n- \"w\": 6,\n- \"x\": 6,\n+ \"w\": 4,\n+ \"x\": 11,\n\"y\": 13\n},\n\"id\": 11,\n@@ -2595,8 +2775,8 @@ grafana:\n\"fill\": 0,\n\"gridPos\": {\n\"h\": 7,\n- \"w\": 6,\n- \"x\": 12,\n+ \"w\": 4,\n+ \"x\": 15,\n\"y\": 13\n},\n\"id\": 4,\n@@ -2711,8 +2891,8 @@ grafana:\n\"fill\": 0,\n\"gridPos\": {\n\"h\": 7,\n- \"w\": 6,\n- \"x\": 18,\n+ \"w\": 5,\n+ \"x\": 19,\n\"y\": 13\n},\n\"id\": 69,\n@@ -2934,5 +3114,6 @@ grafana:\n},\n\"timezone\": \"\",\n\"title\": \"Kubernikus\",\n- \"version\": 7\n+ \"uid\": \"5XtTRCqkk\",\n+ \"version\": 5\n}\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
adds usage plots
|
596,240 |
21.02.2018 15:56:19
| -3,600 |
b51aa1ef170ae5261e438f65c01cafcaf27c64fc
|
tuning for cpu usage
|
[
{
"change_type": "MODIFY",
"old_path": "charts/kube-master/charts/etcd/values.yaml",
"new_path": "charts/kube-master/charts/etcd/values.yaml",
"diff": "@@ -14,7 +14,7 @@ persistence:\n# existingClaim: claimName\nresources:\nrequests:\n- cpu: 250m\n+ cpu: 100m\nmemory: 256Mi\nlimits:\ncpu: 500m\n"
},
{
"change_type": "MODIFY",
"old_path": "charts/kube-master/values.yaml",
"new_path": "charts/kube-master/values.yaml",
"diff": "@@ -44,7 +44,7 @@ api:\n# wormholeHost:\nresources:\nrequests:\n- cpu: 250m\n+ cpu: 100m\nmemory: 256Mi\nlimits:\ncpu: 500m\n@@ -54,7 +54,7 @@ controllerManager:\nreplicaCount: 1\nresources:\nrequests:\n- cpu: 250m\n+ cpu: 50m\nmemory: 256Mi\nlimits:\ncpu: 500m\n@@ -64,7 +64,7 @@ scheduler:\nreplicaCount: 1\nresources:\nrequests:\n- cpu: 250m\n+ cpu: 50m\nmemory: 256Mi\nlimits:\ncpu: 500m\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
tuning for cpu usage
|
596,240 |
21.02.2018 15:56:46
| -3,600 |
298bd562162a0946b3f0dcd3d99d9d51f20ed11d
|
adds sophistication to the buildup docs
|
[
{
"change_type": "ADD",
"old_path": null,
"new_path": "docs/operations/buildup.md",
"diff": "+---\n+title: ControlPlane\n+---\n+\n+## Prepare Values\n+\n+In the secret repository create values:\n+\n+ * `admin/values/$REGION.yaml`\n+ * `kubernikus/$REGION/values/kubernikus.yaml`\n+ * `kubernikus/$REGION/values/kubernikus-system.yaml`\n+\n+Create a random password for authentication. Everything else should be a simple\n+copy/search/replace job.\n+\n+## Setting up a new Region\n+\n+Project is being created using the seed chart:\n+[openstack/kubernikus](https://github.com/sapcc/helm-charts/tree/master/openstack/kubernikus)\n+\n+Install with:\n+\n+```\n+helm upgrade kubernikus openstack/kubernikus --namespace monsoon3 --install\n+```\n+\n+Complete the project by sharing the external Floating-IP network. Per\n+convention this is `FloatingIP-external-ccadmin` in the `ccadmin-net-infra`\n+project in the `ccadmin` domain.\n+\n+Scope yourself to `cloud_admin` in `ccadmin`:\n+\n+```\n+openstack project show -c id -f value --domain ccadmin kubernikus\n+d7df5ce6c37643e49b3c93528c29818b\n+\n+openstack network show -c id -f value FloatingIP-external-ccadmin\n+c2b999de-adb1-4125-ac3f-f74b9f3a1c63\n+\n+openstack network rbac create --target-project d7df5ce6c37643e49b3c93528c29818b --action access_as_shared --type network c2b999de-adb1-4125-ac3f-f74b9f3a1c63\n++-------------------+--------------------------------------+\n+| Field | Value |\n++-------------------+--------------------------------------+\n+| action | access_as_shared |\n+| id | 8643f406-6282-46b2-beee-aa6720cf11d5 |\n+| name | None |\n+| object_id | c2b999de-adb1-4125-ac3f-f74b9f3a1c63 |\n+| object_type | network |\n+| project_id | adc7f04e690a4357a59098c6b2a48db0 |\n+| target_project_id | d7df5ce6c37643e49b3c93528c29818b |\n++-------------------+--------------------------------------+\n+```\n+\n+## Admin Control Plane\n+\n+### Add Pipeline Service User\n+\n+In `ccadmin/cloud_admin` add the Pipeline service user to the\n+`ccadmin/kubernikus` project and give it `kubernetes_admin` permissions. While\n+already here also do the same for the group `CCADMIN_CLOUD_ADMINS`\n+\n+### Adapt Pipeline\n+\n+Add authentication blob and new tasks to the `admin` job in the\n+`ci/pipeline.yaml`. Run the installation of the admin klusters.\n+\n+### Add DNS Entries\n+\n+Use `na-us-1/ccadmin/master` to add the following DNS entries:\n+\n+ * `k-$REGION.admin.cloud.sap.` CNAME ingress.admin.cloud.sap. 1800\n+ * `*.k-$REGION.admin.cloud.sap.` CNAME kubernikus.admin.cloud.sap. 1800\n+\n+### Rewire Kubernikus Dashboard UI\n+\n+Scoped as `ccadmin/cloud_admin` create an additional service and endpoint in\n+the catalog:\n+\n+```\n+openstack service create --name kubernikus kubernikus-kubernikus\n+openstack endpoint create --region $REGION $SERVICEID public https://k-$REGION.admin.cloud.sap\n+```\n+\n+### Smoke Test\n+\n+ 1. Check https://k-$REGION.admin.cloud.sap. You should see the Kubernikus splash\n+page.\n+ 2. Go to `ccadmin/kubernikus` to the Kubernetes tab. It should show you\n+ a workin UI with no klusters.\n+\n+\n+## Create Regional Control Plane\n+\n+Use the UI to create a cluster with the `k-$region` naming scheme in the\n+`ccadmin/kubernikus` project. Create a `default` pool with 3 nodes in\n+`m2.xlarge`. Add and select the Kubernikus Master public key.\n+\n+You should end up with a running kluster and healthy nodes.\n+\n+### Security Group\n+\n+Add TCP/UDP Ingress for the source range `198.18.0.0/15`. Required for load\n+balancers and as a safeguard for DVS agent missed events.\n+\n+### Authenticating\n+\n+Done via UI in the Dashboard in the `ccadmin/kubernikus` project\n+\n+### Prepare Kluster\n+\n+```\n+kubectl -n kube-system create sa tiller\n+kubectl create clusterrolebinding tiller --clusterrole cluster-admin --serviceaccount=kube-system:tiller\n+helm init --service-account tiller --history-max 5\n+```\n+\n+### Adapt Pipeline\n+\n+Check the regional `$CONTINENT` jobs. Add tasks for `kubernikus` and\n+`kubernikus-system` using the authentication blob from earlier. Run the\n+installation of the continent.\n+\n+### Load Balancer Config\n+\n+Once the installation was succesfull, two loadbalancers will appear. Make sure\n+they have IPs from the `147.*.*.*` range assigned. If not detach, release and\n+reallocate correct IPs.\n+\n+Take note that there is one load-balancer with 2 Pools. That is the ingress.\n+The LB with one pool is the sniffer. Mental mark.\n+\n+### Add DNS\n+\n+ * kubernikus-k8sniff.qa-de-1.cloud.sap. A 147.204.35.169 1800\n+ * kubernikus-ingress.qa-de-1.cloud.sap. A 147.204.35.168 180\n+ * *.kubernikus.qa-de-1.cloud.sap. CNAME kubernikus-k8sniff.qa-de-1.cloud.sap. 1800\n+ * kubernikus.qa-de-1.cloud.sap. CNAME kubernikus-ingress.qa-de-1.cloud.sap. 1800\n+ * prometheus.kubernikus.qa-de-1.cloud.sap.CNAME kubernikus-ingress.qa-de-1.cloud.sap. 1800\n+ * grafana.kubernikus.qa-de-1.cloud.sap. CNAME kubernikus-ingress.qa-de-1.cloud.sap. 1800\n+\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
adds sophistication to the buildup docs
|
596,230 |
23.02.2018 10:00:42
| -3,600 |
a6bf2165d43e8f213c1bee7fc6dfe0f7afca5ce9
|
rename sysctl conf file to .conf
|
[
{
"change_type": "MODIFY",
"old_path": "pkg/templates/node_1.7.go",
"new_path": "pkg/templates/node_1.7.go",
"diff": "@@ -238,7 +238,7 @@ storage:\n-A POSTROUTING -p udp ! -d {{ .ClusterCIDR }} -m addrtype ! --dst-type LOCAL -j MASQUERADE --to-ports 32678-65535\n-A POSTROUTING -p icmp ! -d {{ .ClusterCIDR }} -m addrtype ! --dst-type LOCAL -j MASQUERADE\nCOMMIT\n- - path: /etc/sysctl.d/10-enable-icmp-redirects\n+ - path: /etc/sysctl.d/10-enable-icmp-redirects.conf\nfilesystem: root\nmode: 0644\ncontents:\n"
},
{
"change_type": "MODIFY",
"old_path": "pkg/templates/node_1.8.go",
"new_path": "pkg/templates/node_1.8.go",
"diff": "@@ -238,7 +238,7 @@ storage:\n-A POSTROUTING -p udp ! -d {{ .ClusterCIDR }} -m addrtype ! --dst-type LOCAL -j MASQUERADE --to-ports 32678-65535\n-A POSTROUTING -p icmp ! -d {{ .ClusterCIDR }} -m addrtype ! --dst-type LOCAL -j MASQUERADE\nCOMMIT\n- - path: /etc/sysctl.d/10-enable-icmp-redirects\n+ - path: /etc/sysctl.d/10-enable-icmp-redirects.conf\nfilesystem: root\nmode: 0644\ncontents:\n"
},
{
"change_type": "MODIFY",
"old_path": "pkg/templates/node_1.9.go",
"new_path": "pkg/templates/node_1.9.go",
"diff": "@@ -238,7 +238,7 @@ storage:\n-A POSTROUTING -p udp ! -d {{ .ClusterCIDR }} -m addrtype ! --dst-type LOCAL -j MASQUERADE --to-ports 32678-65535\n-A POSTROUTING -p icmp ! -d {{ .ClusterCIDR }} -m addrtype ! --dst-type LOCAL -j MASQUERADE\nCOMMIT\n- - path: /etc/sysctl.d/10-enable-icmp-redirects\n+ - path: /etc/sysctl.d/10-enable-icmp-redirects.conf\nfilesystem: root\nmode: 0644\ncontents:\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
-- rename sysctl conf file to .conf
|
596,232 |
23.02.2018 15:35:58
| -3,600 |
8b9379ff51606c41c21b486c54cbd529f11e891c
|
add target-port to expose command
|
[
{
"change_type": "MODIFY",
"old_path": "docs/guide/common_addons.md",
"new_path": "docs/guide/common_addons.md",
"diff": "@@ -51,7 +51,7 @@ In order to expose the Dashboard without the local proxy, we need to:\nLet's create the service:\n```\n-kubectl expose deployment kubernetes-dashboard --namespace kube-system --type=LoadBalancer --name kubernete-dashboard-external --port=443\n+kubectl expose deployment kubernetes-dashboard --namespace kube-system --type=LoadBalancer --name kubernete-dashboard-external --port=443 --target-port=8443\n```\nThis will create a Kubernetes service that exposes the dashboard on\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
add target-port to expose command
|
596,240 |
25.02.2018 10:13:27
| -3,600 |
09764824e395b7864e4f49db1a359b32c30ada96
|
adds chnaged servicecidr for control plane installations
|
[
{
"change_type": "MODIFY",
"old_path": "docs/operations/buildup.md",
"new_path": "docs/operations/buildup.md",
"diff": "@@ -95,6 +95,14 @@ Use the UI to create a cluster with the `k-$region` naming scheme in the\n`ccadmin/kubernikus` project. Create a `default` pool with 3 nodes in\n`m2.xlarge`. Add and select the Kubernikus Master public key.\n+Note: The `serviceCIDR` needs to be changed!!!!!!!\n+\n+Until supported via UI the cluster needs to be created using the API:\n+\n+```\n+echo '{\"name\":\"k-eu-de-1\",\"spec\":{\"serviceCIDR\": \"192.168.128.0/17\",\"sshPublicKey\": \"ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAACAQCXIxVEUgtUVkvk2VM1hmIb8MxvxsmvYoiq9OBy3J8akTGNybqKsA2uhcwxSJX5Cn3si8kfMfka9EWiJT+e1ybvtsGILO5XRZPxyhYzexwb3TcALwc3LuzpF3Z/Dg2jYTRELTGhYmyca3mxzTlCjNXvYayLNedjJ8fIBzoCuSXNqDRToHru7h0Glz+wtuE74mNkOiXSvhtuJtJs7VCNVjobFQNfC1aeDsri2bPRHJJZJ0QF4LLYSayMEz3lVwIDyAviQR2Aa97WfuXiofiAemfGqiH47Kq6b8X7j3bOYGBvJKMUV7XeWhGsskAmTsvvnFxkc5PAD3Ct+liULjiQWlzDrmpTE8aMqLK4l0YQw7/8iRVz6gli42iEc2ZG56ob1ErpTLAKFWyCNOebZuGoygdEQaGTIIunAncXg5Rz07TdPl0Tf5ZZLpiAgR5ck0H1SETnjDTZ/S83CiVZWJgmCpu8YOKWyYRD4orWwdnA77L4+ixeojLIhEoNL8KlBgsP9Twx+fFMWLfxMmiuX+yksM6Hu+Lsm+Ao7Q284VPp36EB1rxP1JM7HCiEOEm50Jb6hNKjgN4aoLhG5yg+GnDhwCZqUwcRJo1bWtm3QvRA+rzrGZkId4EY3cyOK5QnYV5+24x93Ex0UspHMn7HGsHUESsVeV0fLqlfXyd2RbHTmDMP6w==\", \"nodePools\":[{\"flavor\":\"m1.xlarge_cpu\",\"image\":\"coreos-stable-amd64\",\"name\":\"payload\",\"size\":3}]}}' | curl -K _scratch/.curlrc-kubernikus-eu-de-1 -d @- -XPOST https://k-eu-de-1.admin.cloud.sap/api/v1/clusters\n+```\n+\nYou should end up with a running kluster and healthy nodes.\n### Security Group\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
adds chnaged servicecidr for control plane installations
|
596,230 |
01.03.2018 15:58:47
| -3,600 |
81886f4e775ab4bedd94db0b8d2ea16213e262df
|
use configdrive to create nodes
|
[
{
"change_type": "MODIFY",
"old_path": "pkg/client/openstack/client.go",
"new_path": "pkg/client/openstack/client.go",
"diff": "@@ -613,6 +613,7 @@ func (c *client) CreateNode(kluster *kubernikus_v1.Kluster, pool *models.NodePoo\nname = SimpleNameGenerator.GenerateName(fmt.Sprintf(\"%v-%v-\", kluster.Spec.Name, pool.Name))\n+ configDrive := true\nserver, err := compute.Create(client, servers.CreateOpts{\nName: name,\nFlavorName: pool.Flavor,\n@@ -621,6 +622,7 @@ func (c *client) CreateNode(kluster *kubernikus_v1.Kluster, pool *models.NodePoo\nUserData: userData,\nServiceClient: client,\nSecurityGroups: []string{kluster.Spec.Openstack.SecurityGroupName},\n+ ConfigDrive: &configDrive,\n}).Extract()\nif err != nil {\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
-- use configdrive to create nodes
|
596,233 |
02.03.2018 13:26:46
| -3,600 |
71befdc9750ac1c668a3d32389fc231677a8ee53
|
add kubernikus_node_pool_status metric
|
[
{
"change_type": "MODIFY",
"old_path": "pkg/controller/launch/pool_manager.go",
"new_path": "pkg/controller/launch/pool_manager.go",
"diff": "@@ -97,6 +97,16 @@ func (cpm *ConcretePoolManager) SetStatus(status *PoolStatus) error {\nSchedulable: int64(schedulable),\n}\n+ metrics.SetMetricNodePoolStatus(\n+ cpm.Kluster.GetName(),\n+ cpm.Pool.Name,\n+ map[string]int64{\n+ \"running\": newInfo.Running,\n+ \"healthy\": newInfo.Healthy,\n+ \"schedulable\": newInfo.Schedulable,\n+ },\n+ )\n+\n//TODO: Use util.UpdateKlusterWithRetries here\ncopy, err := cpm.Clients.Kubernikus.Kubernikus().Klusters(cpm.Kluster.Namespace).Get(cpm.Kluster.Name, metav1.GetOptions{})\nif err != nil {\n"
},
{
"change_type": "MODIFY",
"old_path": "pkg/controller/metrics/metrics.go",
"new_path": "pkg/controller/metrics/metrics.go",
"diff": "@@ -124,11 +124,11 @@ func setMetricNodePoolSize(klusterID, nodePoolName, imageName, flavorName string\n/*\nkubernikus_node_pool_status{\"kluster_id\"=\"<id\", \"node_pool\"=\"<name>\", \"status\"=\"<status>\"} < number of nodes in that status >\n-kubernikus_node_pool_status{\"kluster_id\"=\"<id\", \"node_pool\"=\"<name>\", \"status\"=\"ready\"} 1\n+kubernikus_node_pool_status{\"kluster_id\"=\"<id\", \"node_pool\"=\"<name>\", \"status\"=\"schedulable\"} 1\nkubernikus_node_pool_status{\"kluster_id\"=\"<id\", \"node_pool\"=\"<name>\", \"status\"=\"running\"} 1\nkubernikus_node_pool_status{\"kluster_id\"=\"<id\", \"node_pool\"=\"<name>\", \"status\"=\"healthy\"} 1\n*/\n-func setMetricNodePoolStatus(klusterID, nodePoolName string, status map[string]int64) {\n+func SetMetricNodePoolStatus(klusterID, nodePoolName string, status map[string]int64) {\nif status != nil {\nfor s, v := range status {\nnodePoolStatus.With(prometheus.Labels{\n"
},
{
"change_type": "MODIFY",
"old_path": "pkg/controller/metrics/metrics_test.go",
"new_path": "pkg/controller/metrics/metrics_test.go",
"diff": "@@ -26,9 +26,9 @@ kubernikus_node_pool_size{flavor_name=\"flavorName\",image_name=\"imageName\",kluste\nnodePoolStatus: `\n# HELP kubernikus_node_pool_status status of the node pool and the number of nodes nodes in that status\n# TYPE kubernikus_node_pool_status gauge\n-kubernikus_node_pool_status{kluster_id=\"klusterID\",node_pool=\"nodePoolName\",status=\"ready\"} 2\n+kubernikus_node_pool_status{kluster_id=\"klusterID\",node_pool=\"nodePoolName\",status=\"schedulable\"} 2\nkubernikus_node_pool_status{kluster_id=\"klusterID\",node_pool=\"nodePoolName\",status=\"running\"} 2\n-kubernikus_node_pool_status{kluster_id=\"klusterID\",node_pool=\"nodePoolName\",status=\"starting\"} 1\n+kubernikus_node_pool_status{kluster_id=\"klusterID\",node_pool=\"nodePoolName\",status=\"healthy\"} 1\n`,\nklusterInfo: `\n# HELP kubernikus_kluster_info detailed information on a kluster\n@@ -47,7 +47,7 @@ kubernikus_kluster_status_phase{kluster_id=\"klusterID\",phase=\"Terminating\"} 0\n// call functions that update the metrics here\nsetMetricNodePoolSize(\"klusterID\", \"nodePoolName\", \"imageName\", \"flavorName\", 3)\n- setMetricNodePoolStatus(\"klusterID\", \"nodePoolName\", map[string]int64{\"running\": 2, \"starting\": 1, \"ready\": 2})\n+ SetMetricNodePoolStatus(\"klusterID\", \"nodePoolName\", map[string]int64{\"schedulable\": 2, \"healthy\": 1, \"running\": 2})\nSetMetricKlusterInfo(\"namespace\", \"klusterName\", \"version\", \"projectID\", map[string]string{\"creator\": \"D012345\"}, map[string]string{\"account\": \"account\"})\nSetMetricKlusterStatusPhase(\"klusterID\", models.KlusterPhaseRunning)\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
add kubernikus_node_pool_status metric (#250)
|
596,230 |
02.03.2018 14:54:34
| -3,600 |
247271b9e987327b7aba446eff1f9c040a09a443
|
update default clusterrole to allow csr deletions
|
[
{
"change_type": "MODIFY",
"old_path": "pkg/controller/ground/bootstrap.go",
"new_path": "pkg/controller/ground/bootstrap.go",
"diff": "@@ -29,6 +29,9 @@ func SeedKluster(client clientset.Interface, kluster *v1.Kluster) error {\nif err := SeedCinderStorageClass(client); err != nil {\nreturn err\n}\n+ if err := SeedAllowCertificateControllerToDeleteCSRs(client); err != nil {\n+ return err\n+ }\nif err := dns.SeedKubeDNS(client, \"\", \"\", kluster.Spec.DNSDomain, kluster.Spec.DNSAddress); err != nil {\nreturn err\n}\n@@ -118,6 +121,43 @@ func SeedAllowBootstrapTokensToPostCSRs(client clientset.Interface) error {\n})\n}\n+// addresses https://github.com/kubernetes/kubernetes/issues/59351\n+func SeedAllowCertificateControllerToDeleteCSRs(client clientset.Interface) error {\n+ return CreateOrUpdateClusterRole(client, &rbac.ClusterRole{\n+ ObjectMeta: metav1.ObjectMeta{\n+ Name: \"system:controller:certificate-controller\",\n+ Annotations: map[string]string{\n+ \"rbac.authorization.kubernetes.io/autoupdate\": \"true\",\n+ },\n+ Labels: map[string]string{\n+ \"kubernetes.io/bootstrapping\": \"rbac-defaults\",\n+ },\n+ },\n+ Rules: []rbac.PolicyRule{\n+ rbac.PolicyRule{\n+ Verbs: []string{\"delete\", \"get\", \"list\", \"watch\"},\n+ APIGroups: []string{\"certificates.k8s.io\"},\n+ Resources: []string{\"certificatesigningrequests\"},\n+ },\n+ rbac.PolicyRule{\n+ Verbs: []string{\"update\"},\n+ APIGroups: []string{\"certificates.k8s.io\"},\n+ Resources: []string{\"certificatesigningrequests/approval\", \"certificatesigningrequests/status\"},\n+ },\n+ rbac.PolicyRule{\n+ Verbs: []string{\"create\"},\n+ APIGroups: []string{\"authorization.k8s.io\"},\n+ Resources: []string{\"subjectaccessreviews\"},\n+ },\n+ rbac.PolicyRule{\n+ Verbs: []string{\"create\", \"patch\", \"update\"},\n+ APIGroups: []string{\"\"}, //looks funny but is in the default rule ...\n+ Resources: []string{\"events\"},\n+ },\n+ },\n+ })\n+}\n+\nfunc SeedAutoApproveNodeBootstrapTokens(client clientset.Interface) error {\nerr := CreateOrUpdateClusterRole(client, &rbac.ClusterRole{\nObjectMeta: metav1.ObjectMeta{\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
-- update default clusterrole to allow csr deletions
|
596,230 |
02.03.2018 16:47:25
| -3,600 |
b7467eaadb1e0620e13801c68a63fb889edea1b8
|
add liveness to controller manager
|
[
{
"change_type": "MODIFY",
"old_path": "charts/kube-master/templates/controller-manager.yaml",
"new_path": "charts/kube-master/templates/controller-manager.yaml",
"diff": "@@ -81,6 +81,13 @@ spec:\n- --service-account-private-key-file=/etc/kubernetes/certs/apiserver-clients-ca-key.pem\n- --service-cluster-ip-range={{ .Values.serviceCIDR }}\n- --use-service-account-credentials\n+ livenessProbe:\n+ httpGet:\n+ host: 127.0.0.1\n+ path: /healthz\n+ port: 10252\n+ initialDelaySeconds: 15\n+ timeoutSeconds: 1\nvolumeMounts:\n- mountPath: /etc/kubernetes/certs/\nname: certs\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
-- add liveness to controller manager
|
596,230 |
02.03.2018 16:53:56
| -3,600 |
f039c19c728ebc796bec2ddb235325511ab3df54
|
increase initialdelay
|
[
{
"change_type": "MODIFY",
"old_path": "charts/kube-master/templates/controller-manager.yaml",
"new_path": "charts/kube-master/templates/controller-manager.yaml",
"diff": "@@ -86,7 +86,7 @@ spec:\nhost: 127.0.0.1\npath: /healthz\nport: 10252\n- initialDelaySeconds: 15\n+ initialDelaySeconds: 120\ntimeoutSeconds: 1\nvolumeMounts:\n- mountPath: /etc/kubernetes/certs/\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
-- increase initialdelay
|
596,240 |
05.03.2018 09:12:59
| -3,600 |
88c123915a0d23f3404970cbf0daeaec7f08869b
|
removes obsolete project call
|
[
{
"change_type": "MODIFY",
"old_path": "pkg/client/openstack/client.go",
"new_path": "pkg/client/openstack/client.go",
"diff": "@@ -10,7 +10,6 @@ import (\n\"github.com/gophercloud/gophercloud/openstack\"\n\"github.com/gophercloud/gophercloud/openstack/compute/v2/servers\"\n\"github.com/gophercloud/gophercloud/openstack/identity/v3/endpoints\"\n- \"github.com/gophercloud/gophercloud/openstack/identity/v3/projects\"\n\"github.com/gophercloud/gophercloud/openstack/identity/v3/services\"\n\"github.com/gophercloud/gophercloud/openstack/identity/v3/tokens\"\n\"github.com/gophercloud/gophercloud/openstack/identity/v3/users\"\n@@ -58,7 +57,6 @@ type Client interface {\nDeleteNode(*kubernikus_v1.Kluster, string) error\nGetNodes(*kubernikus_v1.Kluster, *models.NodePool) ([]Node, error)\n- GetProject(id string) (*Project, error)\nGetRegion() (string, error)\nGetRouters(project_id string) ([]Router, error)\nDeleteUser(username, domainID string) error\n@@ -97,13 +95,6 @@ func (simpleNameGenerator) GenerateName(base string) string {\nreturn fmt.Sprintf(\"%s%s\", base, utilrand.String(randomLength))\n}\n-type Project struct {\n- ID string\n- Name string\n- Domain string\n- DomainID string\n-}\n-\ntype Router struct {\nID string\nExternalNetworkID string\n@@ -307,29 +298,6 @@ func (c *client) KlusterClientFor(kluster *kubernikus_v1.Kluster) (*gophercloud.\nreturn provider, nil\n}\n-func (c *client) GetProject(id string) (*Project, error) {\n- provider, err := c.adminClient()\n- if err != nil {\n- return nil, err\n- }\n-\n- identity, err := openstack.NewIdentityV3(provider, gophercloud.EndpointOpts{})\n- if err != nil {\n- return nil, err\n- }\n- project, err := projects.Get(identity, id).Extract()\n- if err != nil {\n- return nil, err\n- }\n-\n- domain, err := domains.Get(identity, project.DomainID).Extract()\n- if err != nil {\n- return nil, err\n- }\n-\n- return &Project{ID: id, Name: project.Name, DomainID: project.DomainID, Domain: domain.Name}, nil\n-}\n-\nfunc (c *client) GetRouters(project_id string) ([]Router, error) {\nprovider, err := c.adminClient()\nif err != nil {\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
removes obsolete project call
|
596,230 |
05.03.2018 14:35:19
| -3,600 |
a65f2189f3c50dd8a20014d3f8613ba508da5485
|
change SAPNET cert hash to ensure it has run
|
[
{
"change_type": "MODIFY",
"old_path": "pkg/templates/node_1.7.go",
"new_path": "pkg/templates/node_1.7.go",
"diff": "@@ -156,7 +156,7 @@ systemd:\ncontents: |\n[Unit]\nDescription=Update the certificates w/ self-signed root CAs\n- ConditionPathIsSymbolicLink=!/etc/ssl/certs/48b11003.0\n+ ConditionPathIsSymbolicLink=!/etc/ssl/certs/381107d7.0\nBefore=early-docker.service docker.service\n[Service]\nExecStart=/usr/sbin/update-ca-certificates\n"
},
{
"change_type": "MODIFY",
"old_path": "pkg/templates/node_1.8.go",
"new_path": "pkg/templates/node_1.8.go",
"diff": "@@ -156,7 +156,7 @@ systemd:\ncontents: |\n[Unit]\nDescription=Update the certificates w/ self-signed root CAs\n- ConditionPathIsSymbolicLink=!/etc/ssl/certs/48b11003.0\n+ ConditionPathIsSymbolicLink=!/etc/ssl/certs/381107d7.0\nBefore=early-docker.service docker.service\n[Service]\nExecStart=/usr/sbin/update-ca-certificates\n"
},
{
"change_type": "MODIFY",
"old_path": "pkg/templates/node_1.9.go",
"new_path": "pkg/templates/node_1.9.go",
"diff": "@@ -156,7 +156,7 @@ systemd:\ncontents: |\n[Unit]\nDescription=Update the certificates w/ self-signed root CAs\n- ConditionPathIsSymbolicLink=!/etc/ssl/certs/48b11003.0\n+ ConditionPathIsSymbolicLink=!/etc/ssl/certs/381107d7.0\nBefore=early-docker.service docker.service\n[Service]\nExecStart=/usr/sbin/update-ca-certificates\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
-- change SAPNET cert hash to ensure it has run (#253)
|
596,230 |
08.03.2018 10:57:46
| -3,600 |
affe3d49e864a84f408ae344afa5d5e9a977a378
|
add quotes to KS_PASSWORD to allow spaces
|
[
{
"change_type": "MODIFY",
"old_path": "Procfile",
"new_path": "Procfile",
"diff": "-operator: bin/darwin/kubernikus operator --auth-username=$KS_USERNAME --auth-domain=$KS_USER_DOMAIN_NAME --auth-password=$KS_PASSWORD --auth-project=$KS_PROJECT_NAME --auth-project-domain=$KS_PROJECT_DOMAIN_NAME --auth-url=$KS_AUTH_URL --namespace=$KS_NAMESPACE --context=$KS_CONTEXT --kubernikus-domain=$KS_DOMAIN --v=5\n+operator: bin/darwin/kubernikus operator --auth-username=$KS_USERNAME --auth-domain=$KS_USER_DOMAIN_NAME --auth-password=\"$KS_PASSWORD\" --auth-project=$KS_PROJECT_NAME --auth-project-domain=$KS_PROJECT_DOMAIN_NAME --auth-url=$KS_AUTH_URL --namespace=$KS_NAMESPACE --context=$KS_CONTEXT --kubernikus-domain=$KS_DOMAIN --v=5\napi: bin/darwin/apiserver --context=$KS_CONTEXT --namespace=$KS_NAMESPACE --auth-url=$KS_AUTH_URL --v=5\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
-- add quotes to KS_PASSWORD to allow spaces
|
596,240 |
06.03.2018 18:05:31
| -3,600 |
9ecb71e79253b94747ac3f9ea8b5ebbbaa2f905f
|
fixes nodepool update overwriting securitygroup
|
[
{
"change_type": "MODIFY",
"old_path": "pkg/api/handlers/update_cluster.go",
"new_path": "pkg/api/handlers/update_cluster.go",
"diff": "@@ -31,7 +31,10 @@ func (d *updateCluster) Handle(params operations.UpdateClusterParams, principal\n// Update nodepool\nkluster.Spec.NodePools = nodePools\nkluster.Spec.SSHPublicKey = params.Body.Spec.SSHPublicKey\n+\n+ if params.Body.Spec.Openstack.SecurityGroupName != \"\" {\nkluster.Spec.Openstack.SecurityGroupName = params.Body.Spec.Openstack.SecurityGroupName\n+ }\n})\nif err != nil {\nif apierrors.IsNotFound(err) {\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
fixes nodepool update overwriting securitygroup
|
596,240 |
15.03.2018 10:16:29
| -3,600 |
5a6a89defc16c85014e32f6b2206a6debc5e8500
|
fixes ephemeral port range
|
[
{
"change_type": "MODIFY",
"old_path": "pkg/templates/node_1.7.go",
"new_path": "pkg/templates/node_1.7.go",
"diff": "@@ -234,8 +234,8 @@ storage:\n:INPUT ACCEPT [0:0]\n:OUTPUT ACCEPT [0:0]\n:POSTROUTING ACCEPT [0:0]\n- -A POSTROUTING -p tcp ! -d {{ .ClusterCIDR }} -m addrtype ! --dst-type LOCAL -j MASQUERADE --to-ports 32678-65535\n- -A POSTROUTING -p udp ! -d {{ .ClusterCIDR }} -m addrtype ! --dst-type LOCAL -j MASQUERADE --to-ports 32678-65535\n+ -A POSTROUTING -p tcp ! -d {{ .ClusterCIDR }} -m addrtype ! --dst-type LOCAL -j MASQUERADE --to-ports 32768-65535\n+ -A POSTROUTING -p udp ! -d {{ .ClusterCIDR }} -m addrtype ! --dst-type LOCAL -j MASQUERADE --to-ports 32768-65535\n-A POSTROUTING -p icmp ! -d {{ .ClusterCIDR }} -m addrtype ! --dst-type LOCAL -j MASQUERADE\nCOMMIT\n- path: /etc/sysctl.d/10-enable-icmp-redirects.conf\n"
},
{
"change_type": "MODIFY",
"old_path": "pkg/templates/node_1.8.go",
"new_path": "pkg/templates/node_1.8.go",
"diff": "@@ -234,8 +234,8 @@ storage:\n:INPUT ACCEPT [0:0]\n:OUTPUT ACCEPT [0:0]\n:POSTROUTING ACCEPT [0:0]\n- -A POSTROUTING -p tcp ! -d {{ .ClusterCIDR }} -m addrtype ! --dst-type LOCAL -j MASQUERADE --to-ports 32678-65535\n- -A POSTROUTING -p udp ! -d {{ .ClusterCIDR }} -m addrtype ! --dst-type LOCAL -j MASQUERADE --to-ports 32678-65535\n+ -A POSTROUTING -p tcp ! -d {{ .ClusterCIDR }} -m addrtype ! --dst-type LOCAL -j MASQUERADE --to-ports 32768-65535\n+ -A POSTROUTING -p udp ! -d {{ .ClusterCIDR }} -m addrtype ! --dst-type LOCAL -j MASQUERADE --to-ports 32768-65535\n-A POSTROUTING -p icmp ! -d {{ .ClusterCIDR }} -m addrtype ! --dst-type LOCAL -j MASQUERADE\nCOMMIT\n- path: /etc/sysctl.d/10-enable-icmp-redirects.conf\n"
},
{
"change_type": "MODIFY",
"old_path": "pkg/templates/node_1.9.go",
"new_path": "pkg/templates/node_1.9.go",
"diff": "@@ -234,8 +234,8 @@ storage:\n:INPUT ACCEPT [0:0]\n:OUTPUT ACCEPT [0:0]\n:POSTROUTING ACCEPT [0:0]\n- -A POSTROUTING -p tcp ! -d {{ .ClusterCIDR }} -m addrtype ! --dst-type LOCAL -j MASQUERADE --to-ports 32678-65535\n- -A POSTROUTING -p udp ! -d {{ .ClusterCIDR }} -m addrtype ! --dst-type LOCAL -j MASQUERADE --to-ports 32678-65535\n+ -A POSTROUTING -p tcp ! -d {{ .ClusterCIDR }} -m addrtype ! --dst-type LOCAL -j MASQUERADE --to-ports 32768-65535\n+ -A POSTROUTING -p udp ! -d {{ .ClusterCIDR }} -m addrtype ! --dst-type LOCAL -j MASQUERADE --to-ports 32768-65535\n-A POSTROUTING -p icmp ! -d {{ .ClusterCIDR }} -m addrtype ! --dst-type LOCAL -j MASQUERADE\nCOMMIT\n- path: /etc/sysctl.d/10-enable-icmp-redirects.conf\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
fixes ephemeral port range
|
596,240 |
19.03.2018 16:04:52
| -3,600 |
5c1d18fb632ce675142af1cea1cc70e73212f35f
|
enables smoke test deletion again. adds asr soaktest
|
[
{
"change_type": "MODIFY",
"old_path": "ci/pipeline.yaml",
"new_path": "ci/pipeline.yaml",
"diff": "@@ -39,6 +39,16 @@ auth_e2e: &auth_e2e\nKUBERNIKUS_NAME: e2e\nKUBERNIKUS_URL: https://kubernikus-master.eu-nl-1.cloud.sap\n+auth_e2e_qa-de-1: &auth_e2e_qa-de-1\n+ OS_AUTH_URL: https://identity-3.qa-de-1.cloud.sap/v3\n+ OS_USERNAME: {{kubernikus-qa-username}}\n+ OS_PASSWORD: {{kubernikus-qa-password}}\n+ OS_USER_DOMAIN_NAME: ccadmin\n+ OS_PROJECT_NAME: kubernikus-e2e\n+ OS_PROJECT_DOMAIN_NAME: ccadmin\n+ KUBERNIKUS_NAME: e2e_qa-de-1\n+ KUBERNIKUS_URL: https://kubernikus.qa-de-1.cloud.sap\n+\nauth_staging: &auth_staging\nOS_AUTH_URL: https://identity-3.staging.cloud.sap/v3\nOS_USERNAME: {{kubernikus-staging-username}}\n@@ -135,6 +145,9 @@ resources:\ntype: time\nsource: {interval: 24h}\n+ - name: hourly\n+ type: time\n+ source: {interval: 1h}\nresource_types:\n- name: time-version\n@@ -382,3 +395,18 @@ jobs:\nfile: kubernikus.builds/ci/task_helm_kubernikus-system.yaml\nparams:\n<<: *auth_qa-de-1\n+\n+ - name: asrsoaking\n+ serial: true\n+ plan:\n+ - aggregate:\n+ - get: secrets.git\n+ - get: kubernikus.builds\n+ resource: feature.builds\n+ passed: [qa]\n+ - get: hourly\n+ trigger: true\n+ - task: e2e_tests\n+ file: kubernikus.builds/ci/task_e2e_tests.yaml\n+ params:\n+ <<: *auth_e2e_qa-de-1\n"
},
{
"change_type": "MODIFY",
"old_path": "test/e2e/util.go",
"new_path": "test/e2e/util.go",
"diff": "@@ -216,10 +216,11 @@ func (s *E2ETestSuite) handleError(err error) {\nreturn\n}\nlog.Print(err)\n- // cleanup\n- //if !s.IsNoTeardown {\n- // s.tearDownCluster()\n- //}\n+ // Uncomment this to disabled auto-cleanup\n+ cleanup\n+ if !s.IsNoTeardown {\n+ s.tearDownCluster()\n+ }\nos.Exit(1)\n}\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
enables smoke test deletion again. adds asr soaktest
|
596,240 |
19.03.2018 16:19:53
| -3,600 |
c1b16a08bcec63dfaeafe09ac3590143500421c1
|
fixes cleanup call
|
[
{
"change_type": "MODIFY",
"old_path": "ci/pipeline.yaml",
"new_path": "ci/pipeline.yaml",
"diff": "@@ -402,8 +402,8 @@ jobs:\n- aggregate:\n- get: secrets.git\n- get: kubernikus.builds\n- resource: feature.builds\n- passed: [qa]\n+ resource: master.builds\n+ passed: [master]\n- get: hourly\ntrigger: true\n- task: e2e_tests\n"
},
{
"change_type": "MODIFY",
"old_path": "test/e2e/util.go",
"new_path": "test/e2e/util.go",
"diff": "@@ -217,7 +217,6 @@ func (s *E2ETestSuite) handleError(err error) {\n}\nlog.Print(err)\n// Uncomment this to disabled auto-cleanup\n- cleanup\nif !s.IsNoTeardown {\ns.tearDownCluster()\n}\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
fixes cleanup call
|
596,240 |
19.03.2018 16:21:51
| -3,600 |
ca267ce6bb7dfff961859297371f2ffb480e5cd2
|
soak tests using e2e tested suite
|
[
{
"change_type": "MODIFY",
"old_path": "ci/pipeline.yaml",
"new_path": "ci/pipeline.yaml",
"diff": "@@ -403,7 +403,7 @@ jobs:\n- get: secrets.git\n- get: kubernikus.builds\nresource: master.builds\n- passed: [master]\n+ passed: [e2e]\n- get: hourly\ntrigger: true\n- task: e2e_tests\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
soak tests using e2e tested suite
|
596,240 |
20.03.2018 17:21:12
| -3,600 |
f6e69f6e361244d06d70e9e567b32c382734b474
|
soak all regions in nice e2e testing
|
[
{
"change_type": "MODIFY",
"old_path": "ci/pipeline.yaml",
"new_path": "ci/pipeline.yaml",
"diff": "@@ -29,26 +29,6 @@ auth_feature: &auth_feature\nKUBERNIKUS_NAME: k-feature\nKUBERNIKUS_URL: https://k-feature.admin.cloud.sap\n-auth_e2e: &auth_e2e\n- OS_AUTH_URL: https://identity-3.eu-nl-1.cloud.sap/v3\n- OS_USERNAME: {{kubernikus-prod-username}}\n- OS_PASSWORD: {{kubernikus-prod-password}}\n- OS_USER_DOMAIN_NAME: ccadmin\n- OS_PROJECT_NAME: kubernikus-e2e\n- OS_PROJECT_DOMAIN_NAME: ccadmin\n- KUBERNIKUS_NAME: e2e\n- KUBERNIKUS_URL: https://kubernikus-master.eu-nl-1.cloud.sap\n-\n-auth_e2e_qa-de-1: &auth_e2e_qa-de-1\n- OS_AUTH_URL: https://identity-3.qa-de-1.cloud.sap/v3\n- OS_USERNAME: {{kubernikus-qa-username}}\n- OS_PASSWORD: {{kubernikus-qa-password}}\n- OS_USER_DOMAIN_NAME: ccadmin\n- OS_PROJECT_NAME: kubernikus-e2e\n- OS_PROJECT_DOMAIN_NAME: ccadmin\n- KUBERNIKUS_NAME: e2e_qa-de-1\n- KUBERNIKUS_URL: https://kubernikus.qa-de-1.cloud.sap\n-\nauth_staging: &auth_staging\nOS_AUTH_URL: https://identity-3.staging.cloud.sap/v3\nOS_USERNAME: {{kubernikus-staging-username}}\n@@ -109,6 +89,56 @@ auth_na-us-1: &auth_na-us-1\nKUBERNIKUS_NAME: k-na-us-1\nKUBERNIKUS_URL: https://k-na-us-1.admin.cloud.sap\n+auth_e2e_eu-de-1: &auth_e2e_ap-au-1\n+ OS_AUTH_URL: https://identity-3.ap-au-1.cloud.sap/v3\n+ OS_USERNAME: {{kubernikus-prod-username}}\n+ OS_PASSWORD: {{kubernikus-prod-password}}\n+ OS_USER_DOMAIN_NAME: ccadmin\n+ OS_PROJECT_NAME: kubernikus-e2e\n+ OS_PROJECT_DOMAIN_NAME: ccadmin\n+ KUBERNIKUS_NAME: e2e\n+ KUBERNIKUS_URL: https://kubernikus.ap-au-1.cloud.sap\n+\n+auth_e2e_eu-nl-1: &auth_e2e_eu-nl-1\n+ OS_AUTH_URL: https://identity-3.eu-nl-1.cloud.sap/v3\n+ OS_USERNAME: {{kubernikus-prod-username}}\n+ OS_PASSWORD: {{kubernikus-prod-password}}\n+ OS_USER_DOMAIN_NAME: ccadmin\n+ OS_PROJECT_NAME: kubernikus-e2e\n+ OS_PROJECT_DOMAIN_NAME: ccadmin\n+ KUBERNIKUS_NAME: e2e\n+ KUBERNIKUS_URL: https://kubernikus-master.eu-nl-1.cloud.sap\n+\n+auth_e2e_eu-de-1: &auth_e2e_eu-de-1\n+ OS_AUTH_URL: https://identity-3.eu-de-1.cloud.sap/v3\n+ OS_USERNAME: {{kubernikus-prod-username}}\n+ OS_PASSWORD: {{kubernikus-prod-password}}\n+ OS_USER_DOMAIN_NAME: ccadmin\n+ OS_PROJECT_NAME: kubernikus-e2e\n+ OS_PROJECT_DOMAIN_NAME: ccadmin\n+ KUBERNIKUS_NAME: e2e\n+ KUBERNIKUS_URL: https://kubernikus.eu-de-1.cloud.sap\n+\n+auth_e2e_eu-nl-1: &auth_e2e_na-us-1\n+ OS_AUTH_URL: https://identity-3.na-us-1.cloud.sap/v3\n+ OS_USERNAME: {{kubernikus-prod-username}}\n+ OS_PASSWORD: {{kubernikus-prod-password}}\n+ OS_USER_DOMAIN_NAME: ccadmin\n+ OS_PROJECT_NAME: kubernikus-e2e\n+ OS_PROJECT_DOMAIN_NAME: ccadmin\n+ KUBERNIKUS_NAME: e2e\n+ KUBERNIKUS_URL: https://kubernikus.na-us-1.cloud.sap\n+\n+auth_e2e_qa-de-1: &auth_e2e_qa-de-1\n+ OS_AUTH_URL: https://identity-3.qa-de-1.cloud.sap/v3\n+ OS_USERNAME: {{kubernikus-qa-username}}\n+ OS_PASSWORD: {{kubernikus-qa-password}}\n+ OS_USER_DOMAIN_NAME: ccadmin\n+ OS_PROJECT_NAME: kubernikus-e2e\n+ OS_PROJECT_DOMAIN_NAME: ccadmin\n+ KUBERNIKUS_NAME: e2e_qa-de-1\n+ KUBERNIKUS_URL: https://kubernikus.qa-de-1.cloud.sap\n+\nresources:\n- name: master.builds\n@@ -215,7 +245,7 @@ jobs:\n- task: e2e_tests\nfile: kubernikus.builds/ci/task_e2e_tests.yaml\nparams:\n- <<: *auth_e2e\n+ <<: *auth_e2e_eu-nl-1\n- name: conformance\nserial: true\n@@ -396,7 +426,53 @@ jobs:\nparams:\n<<: *auth_qa-de-1\n- - name: asrsoaking\n+\n+ - name: soak_ap-au-1\n+ serial: true\n+ plan:\n+ - aggregate:\n+ - get: secrets.git\n+ - get: kubernikus.builds\n+ resource: master.builds\n+ passed: [e2e]\n+ - get: hourly\n+ trigger: true\n+ - task: e2e_tests\n+ file: kubernikus.builds/ci/task_e2e_tests.yaml\n+ params:\n+ <<: *auth_e2e_ap-au-1\n+\n+ - name: soak_eu-de-1\n+ serial: true\n+ plan:\n+ - aggregate:\n+ - get: secrets.git\n+ - get: kubernikus.builds\n+ resource: master.builds\n+ passed: [e2e]\n+ - get: hourly\n+ trigger: true\n+ - task: e2e_tests\n+ file: kubernikus.builds/ci/task_e2e_tests.yaml\n+ params:\n+ <<: *auth_e2e_eu-de-1\n+\n+ - name: soak_na-us-1\n+ serial: true\n+ plan:\n+ - aggregate:\n+ - get: secrets.git\n+ - get: kubernikus.builds\n+ resource: master.builds\n+ passed: [e2e]\n+ - get: hourly\n+ trigger: true\n+ - task: e2e_tests\n+ file: kubernikus.builds/ci/task_e2e_tests.yaml\n+ params:\n+ <<: *auth_e2e_na-us-1\n+\n+ - name: soak_qa-de-1\nserial: true\nplan:\n- aggregate:\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
soak all regions in nice e2e testing
|
596,240 |
21.03.2018 17:02:40
| -3,600 |
ecdf9ecbf78735203362b70910de1c62cb244c7d
|
emergency fix for emergency fix
|
[
{
"change_type": "MODIFY",
"old_path": "pkg/client/openstack/factory.go",
"new_path": "pkg/client/openstack/factory.go",
"diff": "@@ -47,7 +47,7 @@ func NewSharedOpenstackClientFactory(secrets core_v1.SecretInterface, klusters c\nlogger: logger,\n}\n- if klusrers != nil {\n+ if klusters != nil {\nklusters.AddEventHandler(cache.ResourceEventHandlerFuncs{\nDeleteFunc: func(obj interface{}) {\nif kluster, ok := obj.(*kubernikus_v1.Kluster); ok {\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
emergency fix for emergency fix
|
596,240 |
21.03.2018 17:29:39
| -3,600 |
44b76c53c51feb0f2bfd1daee7f9a31eb8548a5d
|
fixes service client foo
|
[
{
"change_type": "MODIFY",
"old_path": "pkg/client/openstack/factory.go",
"new_path": "pkg/client/openstack/factory.go",
"diff": "@@ -99,7 +99,7 @@ func (f *factory) KlusterClientFor(kluster *kubernikus_v1.Kluster) (openstack_kl\n}\nvar client openstack_kluster.KlusterClient\n- client = openstack_kluster.NewKlusterClient(identity, compute, network, kluster)\n+ client = openstack_kluster.NewKlusterClient(network, compute, identity, kluster)\nclient = &openstack_kluster.LoggingClient{client, log.With(f.logger, \"kluster\", kluster.GetName(), \"project\", kluster.Account())}\nf.klusterClients.Store(kluster.GetUID(), client)\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
fixes service client foo
|
596,240 |
22.03.2018 09:36:40
| -3,600 |
97927e757e8afaf190a1471d65917e10ec6ba756
|
removes premature termination of nodes
|
[
{
"change_type": "MODIFY",
"old_path": "test/e2e/APITests.go",
"new_path": "test/e2e/APITests.go",
"diff": "@@ -109,9 +109,6 @@ func (s *E2ETestSuite) TestGetClusterInfo() {\n// TestTerminateCluster tests kluster deletion and waits until the kluster is gone\nfunc (s *E2ETestSuite) TestTerminateCluster() {\n- // empty nodePools before terminating cluster\n- s.emptyNodePoolsOfKluster()\n-\nlog.Printf(\"Testing kluster termination\")\n_, err := s.kubernikusClient.Operations.TerminateCluster(\n"
},
{
"change_type": "MODIFY",
"old_path": "test/e2e/util.go",
"new_path": "test/e2e/util.go",
"diff": "@@ -80,52 +80,6 @@ func getNodePoolSizeFromSpec(nodePoolsSpec []models.NodePool, name string) int64\nreturn -1\n}\n-func (s *E2ETestSuite) emptyNodePoolsOfKluster() {\n-\n- log.Printf(\"stopping all nodes of cluster %v\", s.ClusterName)\n-\n- cluster, err := s.kubernikusClient.Operations.ShowCluster(\n- operations.NewShowClusterParams().WithName(s.ClusterName),\n- s.authFunc(),\n- )\n- s.handleError(err)\n-\n- nodePools := []models.NodePool{}\n- for _, nodePool := range cluster.Payload.Spec.NodePools {\n- nodePool.Size = 0\n- nodePools = append(nodePools, nodePool)\n- }\n- cluster.Payload.Spec.NodePools = nodePools\n-\n- // empty node pools\n- _, err = s.kubernikusClient.Operations.UpdateCluster(\n- operations.NewUpdateClusterParams().\n- WithName(s.ClusterName).\n- WithBody(cluster.Payload),\n- s.authFunc(),\n- )\n- s.handleError(err)\n-\n- err = s.waitForCluster(\n- s.ClusterName,\n- fmt.Sprintf(\"Not all nodes of cluster %v could be terminated in time\", s.ClusterName),\n- func(k *models.Kluster, err error) bool {\n- if err != nil {\n- log.Println(err)\n- return false\n- }\n- for _, node := range k.Status.NodePools {\n- if node.Running != 0 {\n- log.Printf(\"Cluster %v has nodes in state running\", k.Name)\n- return false\n- }\n- }\n- return true\n- },\n- )\n- s.handleError(err)\n-}\n-\nfunc newE2ECluster(klusterName string) *models.Kluster {\nreturn &models.Kluster{\nName: klusterName,\n@@ -224,7 +178,6 @@ func (s *E2ETestSuite) handleError(err error) {\n}\nfunc (s *E2ETestSuite) tearDownCluster() {\n- s.emptyNodePoolsOfKluster()\nlog.Printf(\"Deleting cluster %v\", s.ClusterName)\n_, err := s.kubernikusClient.Operations.TerminateCluster(\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
removes premature termination of nodes
|
596,240 |
23.03.2018 10:05:31
| -3,600 |
5655c734c6018ea0791edd5b499e89656c6f4223
|
fixes startup error while finding proxier mode for kube-proxy
|
[
{
"change_type": "MODIFY",
"old_path": "pkg/templates/node_1.7.go",
"new_path": "pkg/templates/node_1.7.go",
"diff": "@@ -350,7 +350,7 @@ storage:\nminSyncPeriod: 0s\nsyncPeriod: 30s\nmetricsBindAddress: 127.0.0.1:10249\n- mode: \"\"\n+ mode: \"iptables\"\noomScoreAdj: -999\nportRange: \"\"\nresourceContainer: /kube-proxy\n"
},
{
"change_type": "MODIFY",
"old_path": "pkg/templates/node_1.8.go",
"new_path": "pkg/templates/node_1.8.go",
"diff": "@@ -346,7 +346,7 @@ storage:\nminSyncPeriod: 0s\nsyncPeriod: 30s\nmetricsBindAddress: 127.0.0.1:10249\n- mode: \"\"\n+ mode: \"iptables\"\noomScoreAdj: -999\nportRange: \"\"\nresourceContainer: /kube-proxy\n"
},
{
"change_type": "MODIFY",
"old_path": "pkg/templates/node_1.9.go",
"new_path": "pkg/templates/node_1.9.go",
"diff": "@@ -346,7 +346,7 @@ storage:\nminSyncPeriod: 0s\nsyncPeriod: 30s\nmetricsBindAddress: 127.0.0.1:10249\n- mode: \"\"\n+ mode: \"iptables\"\noomScoreAdj: -999\nportRange: \"\"\nresourceContainer: /kube-proxy\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
fixes startup error while finding proxier mode for kube-proxy
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.