author
int64 658
755k
| date
stringlengths 19
19
| timezone
int64 -46,800
43.2k
| hash
stringlengths 40
40
| message
stringlengths 5
490
| mods
list | language
stringclasses 20
values | license
stringclasses 3
values | repo
stringlengths 5
68
| original_message
stringlengths 12
491
|
---|---|---|---|---|---|---|---|---|---|
596,226 |
28.06.2021 12:00:10
| -7,200 |
dd449b8282c1fade956db4d80d938b9fa991e807
|
Set update strategy in etcd deployment
|
[
{
"change_type": "MODIFY",
"old_path": "charts/kube-master/charts/etcd/templates/deployment.yaml",
"new_path": "charts/kube-master/charts/etcd/templates/deployment.yaml",
"diff": "@@ -70,6 +70,11 @@ metadata:\nrelease: {{ .Release.Name }}\nspec:\nreplicas: 1\n+ strategy:\n+ rollingUpdate:\n+ maxUnavailable: 1\n+ maxSurge: 1\n+ type: RollingUpdate\nselector:\nmatchLabels:\napp: {{ include \"fullname\" . }}\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
Set update strategy in etcd deployment (#580)
|
596,226 |
14.07.2021 16:46:16
| -7,200 |
28a4dfd11d4a5883a9329611876543bc98a3115f
|
Remove servicing controller test
|
[
{
"change_type": "DELETE",
"old_path": "pkg/controller/servicing/controller_test.go",
"new_path": null,
"diff": "-package servicing\n-\n-import (\n- \"testing\"\n- \"time\"\n-\n- \"github.com/go-kit/kit/log\"\n- \"github.com/stretchr/testify/assert\"\n-\n- \"github.com/sapcc/kubernikus/pkg/api/models\"\n- \"github.com/sapcc/kubernikus/pkg/controller/nodeobservatory\"\n- \"github.com/sapcc/kubernikus/pkg/controller/servicing/flatcar\"\n- kubernikusfake \"github.com/sapcc/kubernikus/pkg/generated/clientset/fake\"\n-)\n-\n-func TestServicingControllerReconcile(t *testing.T) {\n- Now = func() time.Time { return time.Date(2019, 2, 3, 4, 0, 0, 0, time.UTC) }\n-\n- rec := Now().Add(-1 * time.Minute)\n- pre := Now().Add(-1 * ServiceInterval).Add(-1 * time.Second)\n- now := Now()\n-\n- type test struct {\n- message string\n- options *FakeKlusterOptions\n- expectedDrain bool\n- expectedReboot bool\n- expectedReplace bool\n- }\n- for _, subject := range []test{\n- //\n- // Test Kluster Phase\n- //\n- {\n- message: \"Running klusters should be reconciled\",\n- options: &FakeKlusterOptions{\n- Phase: models.KlusterPhaseRunning,\n- LastService: nil,\n- NodePools: []FakeNodePoolOptions{\n- {\n- AllowReboot: true,\n- AllowReplace: true,\n- NodeHealthy: true,\n- NodeOSOutdated: true,\n- NodeKubeletOutdated: true,\n- Size: 1,\n- },\n- },\n- },\n- expectedDrain: true,\n- expectedReboot: false,\n- expectedReplace: true,\n- },\n- {\n- message: \"Creating klusters should not be reconciled\",\n- options: &FakeKlusterOptions{\n- Phase: models.KlusterPhaseCreating,\n- LastService: nil,\n- NodePools: []FakeNodePoolOptions{\n- {\n- AllowReboot: true,\n- AllowReplace: true,\n- NodeHealthy: true,\n- NodeOSOutdated: true,\n- NodeKubeletOutdated: true,\n- Size: 1,\n- },\n- },\n- },\n- expectedDrain: false,\n- expectedReboot: false,\n- expectedReplace: false,\n- },\n- {\n- message: \"Pending klusters should not be reconciled\",\n- options: &FakeKlusterOptions{\n- Phase: models.KlusterPhasePending,\n- LastService: nil,\n- NodePools: []FakeNodePoolOptions{\n- {\n- AllowReboot: true,\n- AllowReplace: true,\n- NodeHealthy: true,\n- NodeOSOutdated: true,\n- NodeKubeletOutdated: true,\n- Size: 1,\n- },\n- },\n- },\n- expectedDrain: false,\n- expectedReboot: false,\n- expectedReplace: false},\n- {\n- message: \"Terminating klusters should not be reconciled\",\n- options: &FakeKlusterOptions{\n- Phase: models.KlusterPhaseTerminating,\n- LastService: nil,\n- NodePools: []FakeNodePoolOptions{\n- {\n- AllowReboot: true,\n- AllowReplace: true,\n- NodeHealthy: true,\n- NodeOSOutdated: true,\n- NodeKubeletOutdated: true,\n- Size: 1,\n- },\n- },\n- },\n- expectedDrain: false,\n- expectedReboot: false,\n- expectedReplace: false,\n- },\n-\n- //\n- // Test Service Interval\n- //\n- {\n- message: \"Never serviced klusters should be reconciled\",\n- options: &FakeKlusterOptions{\n- Phase: models.KlusterPhaseRunning,\n- LastService: nil,\n- NodePools: []FakeNodePoolOptions{\n- {\n- AllowReboot: true,\n- AllowReplace: true,\n- NodeHealthy: true,\n- NodeOSOutdated: true,\n- NodeKubeletOutdated: true,\n- Size: 1,\n- },\n- },\n- },\n- expectedDrain: true,\n- expectedReboot: false,\n- expectedReplace: true},\n- {\n- message: \"Klusters serviced recently should not be reconciled\",\n- options: &FakeKlusterOptions{\n- Phase: models.KlusterPhaseRunning,\n- LastService: &rec,\n- NodePools: []FakeNodePoolOptions{\n- {\n- AllowReboot: true,\n- AllowReplace: true,\n- NodeHealthy: true,\n- NodeOSOutdated: true,\n- NodeKubeletOutdated: true,\n- Size: 1,\n- },\n- },\n- },\n- expectedDrain: false,\n- expectedReboot: false,\n- expectedReplace: false,\n- },\n- {\n- message: \"Klusters serviced longer than service interval ago should be reconciled\",\n- options: &FakeKlusterOptions{\n- Phase: models.KlusterPhaseRunning,\n- LastService: &pre,\n- NodePools: []FakeNodePoolOptions{\n- {\n- AllowReboot: true,\n- AllowReplace: true,\n- NodeHealthy: true,\n- NodeOSOutdated: true,\n- NodeKubeletOutdated: true,\n- Size: 1,\n- },\n- },\n- },\n- expectedDrain: true,\n- expectedReboot: false,\n- expectedReplace: true,\n- },\n- {\n- message: \"Klusters serviced twice in a row should not be reconciled\",\n- options: &FakeKlusterOptions{\n- Phase: models.KlusterPhaseRunning,\n- LastService: &now,\n- NodePools: []FakeNodePoolOptions{\n- {\n- AllowReboot: true,\n- AllowReplace: true,\n- NodeHealthy: true,\n- NodeOSOutdated: true,\n- NodeKubeletOutdated: true,\n- Size: 1,\n- },\n- },\n- },\n- expectedDrain: false,\n- expectedReboot: false,\n- expectedReplace: false,\n- },\n-\n- //\n- // Test Unhealthy Klusters\n- //\n- {\n- message: \"Unhealthy klusters should not be reconciled\",\n- options: &FakeKlusterOptions{\n- Phase: models.KlusterPhaseRunning,\n- LastService: nil,\n- NodePools: []FakeNodePoolOptions{\n- {\n- AllowReboot: true,\n- AllowReplace: true,\n- NodeHealthy: false,\n- NodeOSOutdated: true,\n- NodeKubeletOutdated: true,\n- Size: 1,\n- },\n- },\n- },\n- expectedDrain: false,\n- expectedReboot: false,\n- expectedReplace: false,\n- },\n-\n- //\n- // Test Replacement\n- //\n- {\n- message: \"Nodes with outdated kubelet and OS should be replaced\",\n- options: &FakeKlusterOptions{\n- Phase: models.KlusterPhaseRunning,\n- LastService: nil,\n- NodePools: []FakeNodePoolOptions{\n- {\n- AllowReboot: true,\n- AllowReplace: true,\n- NodeHealthy: true,\n- NodeOSOutdated: true,\n- NodeKubeletOutdated: true,\n- Size: 1,\n- },\n- },\n- },\n- expectedDrain: true,\n- expectedReboot: false,\n- expectedReplace: true,\n- },\n- {\n- message: \"Nodes with outdate OS should be rebooted\",\n- options: &FakeKlusterOptions{\n- Phase: models.KlusterPhaseRunning,\n- LastService: nil,\n- NodePools: []FakeNodePoolOptions{\n- {\n- AllowReboot: true,\n- AllowReplace: true,\n- NodeHealthy: true,\n- NodeOSOutdated: true,\n- NodeKubeletOutdated: false,\n- Size: 1,\n- },\n- },\n- },\n- expectedDrain: true,\n- expectedReboot: true,\n- expectedReplace: false,\n- },\n- {\n- message: \"Up-to-date Node should neither be rebooted nor be replaced\",\n- options: &FakeKlusterOptions{\n- Phase: models.KlusterPhaseRunning,\n- LastService: nil,\n- NodePools: []FakeNodePoolOptions{\n- {\n- AllowReboot: true,\n- AllowReplace: true,\n- NodeHealthy: true,\n- NodeOSOutdated: false,\n- NodeKubeletOutdated: false,\n- Size: 1,\n- },\n- },\n- },\n- expectedDrain: false,\n- expectedReboot: false,\n- expectedReplace: false,\n- },\n- } {\n- t.Run(string(subject.message), func(t *testing.T) {\n- kluster, nodes := NewFakeKluster(subject.options)\n- logger := log.With(TestLogger(), \"controller\", \"servicing\")\n-\n- mockCycler := &MockLifeCycler{}\n- mockCycler.On(\"Reboot\", nodes[0]).Return(nil).Times(0)\n- mockCycler.On(\"Drain\", nodes[0]).Return(nil).Times(0)\n- mockCycler.On(\"Replace\", nodes[0]).Return(nil).Times(0)\n- mockCycler.On(\"Uncordon\", nodes[0]).Return(nil).Times(0)\n-\n- var cycler LifeCycler = &LoggingLifeCycler{\n- Logger: log.With(logger, \"kluster\", kluster.Spec.Name, \"project\", kluster.Account()),\n- LifeCycler: mockCycler,\n- }\n-\n- lifecyclers := &MockLifeCyclerFactory{}\n- lifecyclers.On(\"Make\", kluster).Return(cycler, nil)\n-\n- listers := &NodeListerFactory{\n- Logger: logger,\n- NodeObservatory: nodeobservatory.NewFakeController(kluster, nodes...),\n- FlatcarVersion: flatcar.NewFakeVersion(t, \"2303.4.0\"),\n- FlatcarRelease: flatcar.NewFakeRelease(t, \"2303.4.0\"),\n- }\n-\n- reconcilers := &KlusterReconcilerFactory{\n- Logger: logger,\n- ListerFactory: listers,\n- LifeCyclerFactory: lifecyclers,\n- KlusterLister: NewFakeKlusterLister(kluster),\n- KubernikusClient: kubernikusfake.NewSimpleClientset(kluster).Kubernikus(),\n- }\n-\n- controller := &Controller{\n- Logger: logger,\n- Reconciler: reconcilers,\n- }\n-\n- _, err := controller.Reconcile(kluster)\n- if subject.expectedDrain {\n- mockCycler.AssertCalled(t, \"Drain\", nodes[0])\n- } else {\n- mockCycler.AssertNotCalled(t, \"Drain\")\n- }\n-\n- if subject.expectedReboot {\n- mockCycler.AssertCalled(t, \"Reboot\", nodes[0])\n- } else {\n- mockCycler.AssertNotCalled(t, \"Reboot\")\n- }\n-\n- if subject.expectedReplace {\n- mockCycler.AssertCalled(t, \"Replace\", nodes[0])\n- } else {\n- mockCycler.AssertNotCalled(t, \"Replace\")\n- }\n- assert.NoError(t, err)\n- })\n- }\n-}\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
Remove servicing controller test
|
596,240 |
21.07.2021 13:14:01
| -7,200 |
84c29bd26f2868ceb0e94022585f53c6996f54b8
|
just replaces all nodes < 2905 instead of reboots
|
[
{
"change_type": "MODIFY",
"old_path": "pkg/controller/servicing/controller_test.go",
"new_path": "pkg/controller/servicing/controller_test.go",
"diff": "@@ -283,7 +283,7 @@ func TestServicingControllerReconcile(t *testing.T) {\n},\n} {\nt.Run(string(subject.message), func(t *testing.T) {\n- kluster, nodes := NewFakeKluster(subject.options)\n+ kluster, nodes := NewFakeKluster(subject.options, true)\nlogger := log.With(TestLogger(), \"controller\", \"servicing\")\nmockCycler := &MockLifeCycler{}\n@@ -303,8 +303,8 @@ func TestServicingControllerReconcile(t *testing.T) {\nlisters := &NodeListerFactory{\nLogger: logger,\nNodeObservatory: nodeobservatory.NewFakeController(kluster, nodes...),\n- FlatcarVersion: flatcar.NewFakeVersion(t, \"2303.4.0\"),\n- FlatcarRelease: flatcar.NewFakeRelease(t, \"2303.4.0\"),\n+ FlatcarVersion: flatcar.NewFakeVersion(t, \"3000.0.0\"),\n+ FlatcarRelease: flatcar.NewFakeRelease(t, \"3000.0.0\"),\n}\nreconcilers := &KlusterReconcilerFactory{\n"
},
{
"change_type": "MODIFY",
"old_path": "pkg/controller/servicing/lister.go",
"new_path": "pkg/controller/servicing/lister.go",
"diff": "@@ -149,7 +149,7 @@ func (d *NodeLister) Reboot() []*core_v1.Node {\ncontinue\n}\nif len(node.GetName()) == len(prefix)+generator.RandomLength {\n- if util.IsFlatcarNodeWithRkt(node) && latestFlatcar.Major() >= 2905 {\n+ if util.IsFlatcarNodeWithRkt(node) {\ncontinue\n}\n@@ -224,7 +224,7 @@ func (d *NodeLister) Replace() []*core_v1.Node {\nif len(node.GetName()) == len(prefix)+generator.RandomLength {\nnodeNameToPool[node.GetName()] = &d.Kluster.Spec.NodePools[i]\n- if *pool.Config.AllowReplace == true || (util.IsFlatcarNodeWithRkt(node) && latestFlatcar.Major() >= 2905) {\n+ if *pool.Config.AllowReplace == true || util.IsFlatcarNodeWithRkt(node) {\nupgradable = append(upgradable, node)\n}\n}\n@@ -281,7 +281,7 @@ func (d *NodeLister) Replace() []*core_v1.Node {\ncontinue\n}\n- if util.IsFlatcarNodeWithRkt(node) && latestFlatcar.Major() >= 2905 {\n+ if util.IsFlatcarNodeWithRkt(node) {\nuptodate := true\nif strings.HasPrefix(node.Status.NodeInfo.OSImage, \"Flatcar Container Linux\") {\n"
},
{
"change_type": "MODIFY",
"old_path": "pkg/controller/servicing/lister_test.go",
"new_path": "pkg/controller/servicing/lister_test.go",
"diff": "@@ -43,7 +43,7 @@ func NewFakeNodeLister(t *testing.T, logger log.Logger, kluster *v1.Kluster, nod\nreturn lister\n}\n-func NewFakeKlusterForListerTests() (*v1.Kluster, []runtime.Object) {\n+func NewFakeKlusterForListerTests(afterFlatCarRktRemoval bool) (*v1.Kluster, []runtime.Object) {\nreturn NewFakeKluster(\n&FakeKlusterOptions{\nPhase: models.KlusterPhaseRunning,\n@@ -164,35 +164,42 @@ func NewFakeKlusterForListerTests() (*v1.Kluster, []runtime.Object) {\n},\n},\n},\n+ afterFlatCarRktRemoval,\n)\n}\nfunc TestServicingListertAll(t *testing.T) {\n- kluster, nodes := NewFakeKlusterForListerTests()\n+ kluster, nodes := NewFakeKlusterForListerTests(false)\nlister := NewFakeNodeLister(t, TestLogger(), kluster, nodes, \"2605.7.0\")\nassert.Len(t, lister.All(), 16)\n- lister = NewFakeNodeLister(t, TestLogger(), kluster, nodes, \"2905.2.1\")\n+\n+ kluster, nodes = NewFakeKlusterForListerTests(true)\n+ lister = NewFakeNodeLister(t, TestLogger(), kluster, nodes, \"3000.0.0\")\nassert.Len(t, lister.All(), 16)\n}\nfunc TestServicingListerRequiringReboot(t *testing.T) {\n- kluster, nodes := NewFakeKlusterForListerTests()\n+ kluster, nodes := NewFakeKlusterForListerTests(false)\nlister := NewFakeNodeLister(t, TestLogger(), kluster, nodes, \"2605.7.0\")\n- assert.Len(t, lister.Reboot(), 4)\n- lister = NewFakeNodeLister(t, TestLogger(), kluster, nodes, \"2905.2.1\")\nassert.Len(t, lister.Reboot(), 0)\n+\n+ kluster, nodes = NewFakeKlusterForListerTests(true)\n+ lister = NewFakeNodeLister(t, TestLogger(), kluster, nodes, \"3000.0.0\")\n+ assert.Len(t, lister.Reboot(), 4)\n}\nfunc TestServicingListerRequiringReplacement(t *testing.T) {\n- kluster, nodes := NewFakeKlusterForListerTests()\n+ kluster, nodes := NewFakeKlusterForListerTests(false)\nlister := NewFakeNodeLister(t, TestLogger(), kluster, nodes, \"2605.7.0\")\n- assert.Len(t, lister.Replace(), 4)\n- lister = NewFakeNodeLister(t, TestLogger(), kluster, nodes, \"2905.2.1\")\nassert.Len(t, lister.Replace(), 10)\n+\n+ kluster, nodes = NewFakeKlusterForListerTests(true)\n+ lister = NewFakeNodeLister(t, TestLogger(), kluster, nodes, \"3000.0.0\")\n+ assert.Len(t, lister.Replace(), 4)\n}\nfunc TestServicingListerNotReady(t *testing.T) {\n- kluster, nodes := NewFakeKlusterForListerTests()\n+ kluster, nodes := NewFakeKlusterForListerTests(false)\nlister := NewFakeNodeLister(t, TestLogger(), kluster, nodes, \"2605.7.0\")\nassert.Len(t, lister.NotReady(), 15)\n}\n@@ -223,6 +230,7 @@ func TestServicingListerUpdating(t *testing.T) {\n},\n},\n},\n+ false,\n)\nlister := NewFakeNodeLister(t, TestLogger(), kluster, nodes, \"2605.7.0\")\nassert.Len(t, lister.Updating(), 1)\n@@ -310,6 +318,7 @@ func TestServicingListerUpdateSuccessful(t *testing.T) {\n},\n},\n},\n+ false,\n)\nlister := NewFakeNodeLister(t, TestLogger(), kluster, nodes, \"2605.7.0\")\nassert.Len(t, lister.Successful(), 2)\n@@ -397,6 +406,7 @@ func TestServicingListerUpdateFailed(t *testing.T) {\n},\n},\n},\n+ false,\n)\nlister := NewFakeNodeLister(t, TestLogger(), kluster, nodes, \"2605.7.0\")\nassert.Len(t, lister.Failed(), 3)\n"
},
{
"change_type": "MODIFY",
"old_path": "pkg/controller/servicing/testing.go",
"new_path": "pkg/controller/servicing/testing.go",
"diff": "@@ -51,7 +51,7 @@ type FakeNodePoolOptions struct {\n}\n// NewFakeKluster creates a Kluster Object for tests\n-func NewFakeKluster(opts *FakeKlusterOptions) (*v1.Kluster, []runtime.Object) {\n+func NewFakeKluster(opts *FakeKlusterOptions, afterFlatCarRktRemoval bool) (*v1.Kluster, []runtime.Object) {\nkluster := &v1.Kluster{\nObjectMeta: meta_v1.ObjectMeta{\nNamespace: \"servicing\",\n@@ -115,9 +115,13 @@ func NewFakeKluster(opts *FakeKlusterOptions) (*v1.Kluster, []runtime.Object) {\n}\nif p.NodeOSOutdated {\n+ if afterFlatCarRktRemoval {\n+ node.Status.NodeInfo.OSImage = \"Flatcar Container Linux by Kinvolk 2999.2.6 (Oklo)\"\n+ } else {\nnode.Status.NodeInfo.OSImage = \"Flatcar Container Linux by Kinvolk 1000.0.0 (Oklo)\"\n+ }\n} else {\n- node.Status.NodeInfo.OSImage = \"Flatcar Container Linux by Kinvolk 3000.0.0 (Oklo)\"\n+ node.Status.NodeInfo.OSImage = \"Flatcar Container Linux by Kinvolk 3000.1.2 (Oklo)\"\n}\nif p.NodeKubeletOutdated {\n"
},
{
"change_type": "MODIFY",
"old_path": "pkg/templates/ignition.go",
"new_path": "pkg/templates/ignition.go",
"diff": "@@ -27,6 +27,8 @@ var Ignition = &ignition{}\nvar passwordHashRounds = 1000000\n+const TEMPLATE_VERSION = \"1\"\n+\nfunc (i *ignition) getIgnitionTemplate(kluster *kubernikusv1.Kluster) (string, error) {\nswitch {\ncase strings.HasPrefix(kluster.Spec.Version, \"1.20\"):\n@@ -104,6 +106,7 @@ func (i *ignition) GenerateNode(kluster *kubernikusv1.Kluster, pool *models.Node\nfor _, userLabel := range pool.Labels {\nnodeLabels = append(nodeLabels, userLabel)\n}\n+ nodeLabels = append(nodeLabels, \"kubernikus.cloud.sap/template-version=\"+TEMPLATE_VERSION)\nisFlatcar = !strings.Contains(strings.ToLower(pool.Image), \"coreos\")\n}\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
just replaces all nodes < 2905 instead of reboots (#592)
|
596,240 |
22.07.2021 11:57:35
| -7,200 |
7ca3438aab629dba614a9f77145530abc7c36ebd
|
disables eviction
|
[
{
"change_type": "MODIFY",
"old_path": "pkg/controller/servicing/lifecycler.go",
"new_path": "pkg/controller/servicing/lifecycler.go",
"diff": "@@ -149,6 +149,7 @@ func (lc *NodeLifeCycler) Drain(node *core_v1.Node) error {\nGracePeriodSeconds: -1,\nIgnoreAllDaemonSets: true,\nTimeout: EvictionTimeout,\n+ DisableEviction: true,\nDeleteLocalData: true,\nSelector: \"\",\nPodSelector: \"\",\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
disables eviction
|
596,226 |
10.09.2021 14:36:43
| -7,200 |
16c04889b23cab8bc45904bb6184a786c36548a3
|
Fix calico cni networking
|
[
{
"change_type": "MODIFY",
"old_path": "charts/images.yaml",
"new_path": "charts/images.yaml",
"diff": "@@ -18,7 +18,7 @@ imagesForVersion:\ntag: '0.5.2'\nkubelet:\nrepository: 'keppel.$REGION.cloud.sap/ccloud/kubelet'\n- tag: 'v1.20.8'\n+ tag: 'v1.20.8-sap.1'\nkubeProxy:\nrepository: 'keppel.$REGION.cloud.sap/ccloud/kube-proxy'\ntag: 'v1.20.8'\n@@ -154,7 +154,7 @@ imagesForVersion:\ntag: '0.5.2'\nkubelet:\nrepository: 'keppel.$REGION.cloud.sap/ccloud/kubelet'\n- tag: 'v1.19.11'\n+ tag: 'v1.19.11-sap.1'\nkubeProxy:\nrepository: 'keppel.$REGION.cloud.sap/ccloud/kube-proxy'\ntag: 'v1.19.11'\n"
},
{
"change_type": "MODIFY",
"old_path": "contrib/all/Dockerfile.kubelet",
"new_path": "contrib/all/Dockerfile.kubelet",
"diff": "@@ -19,3 +19,8 @@ RUN rm -rf /usr/local/bin/* /apiserver /cloud-controller-manager /controller-man\nCOPY --from=builder /tmp/kubelet /usr/local/bin/kubelet\nCOPY --from=builder /tmp/kubectl /usr/local/bin/kubectl\n+\n+RUN apt-get update && \\\n+ apt-get -y --no-install-recommends install iproute2 && \\\n+ apt-get clean && \\\n+ rm -rf /var/lib/apt/lists/*\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
Fix calico cni networking
|
596,226 |
17.09.2021 14:18:10
| -7,200 |
53519917eb7be43b872cae2a57388e70558586a0
|
Add annotation for node replacement
|
[
{
"change_type": "MODIFY",
"old_path": "pkg/controller/servicing/lister.go",
"new_path": "pkg/controller/servicing/lister.go",
"diff": "@@ -21,6 +21,11 @@ import (\n\"github.com/sapcc/kubernikus/pkg/util/version\"\n)\n+const (\n+ AnnotationNodeForceReplace = \"kubernikus.cloud.sap/forceReplace\"\n+ AnnotationNodeSkipReplace = \"kubernikus.cloud.sap/skipReplace\"\n+)\n+\ntype (\n// Lister enumerates Nodes in various states\nLister interface {\n@@ -213,7 +218,7 @@ func (d *NodeLister) Replace() []*core_v1.Node {\nif util.IsKubernikusNode(node.Name, d.Kluster.Spec.Name, pool.Name) {\nnodeNameToPool[node.GetName()] = &d.Kluster.Spec.NodePools[i]\n- if *pool.Config.AllowReplace == true || util.IsFlatcarNodeWithRkt(node) {\n+ if *pool.Config.AllowReplace == true || util.IsFlatcarNodeWithRkt(node) || util.EnabledValue(node.Annotations[AnnotationNodeForceReplace]) {\nupgradable = append(upgradable, node)\n}\n}\n@@ -226,6 +231,15 @@ func (d *NodeLister) Replace() []*core_v1.Node {\n}\nfor _, node := range upgradable {\n+ if util.EnabledValue(node.Annotations[AnnotationNodeSkipReplace]) {\n+ continue\n+ }\n+\n+ if util.EnabledValue(node.Annotations[AnnotationNodeForceReplace]) {\n+ found = append(found, node)\n+ continue\n+ }\n+\nif util.IsCoreOSNode(node) && util.IsFlatcarNodePool(nodeNameToPool[node.GetName()]) {\nfound = append(found, node)\ncontinue\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
Add annotation for node replacement (#616)
|
596,226 |
21.09.2021 14:38:06
| -7,200 |
b9ef721706c2c891328b2307038d9260b01f0309
|
Use 1.21 in e2e test
|
[
{
"change_type": "MODIFY",
"old_path": "test/e2e/setup_test.go",
"new_path": "test/e2e/setup_test.go",
"diff": "@@ -38,7 +38,7 @@ func (s *SetupTests) Run(t *testing.T) {\n}\nfunc (s *SetupTests) CreateCluster(t *testing.T) {\n- version := \"1.20.8\"\n+ version := \"1.21.5\"\nif v := os.Getenv(\"KLUSTER_VERSION\"); v != \"\" {\nversion = v\n}\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
Use 1.21 in e2e test
|
596,226 |
30.09.2021 10:45:41
| -7,200 |
79457d1eea52b0ad7a102c5f87574d4eee9a7c55
|
Seed cinder csi roles on upgrade
|
[
{
"change_type": "MODIFY",
"old_path": "pkg/controller/ground.go",
"new_path": "pkg/controller/ground.go",
"diff": "@@ -651,6 +651,17 @@ func (op *GroundControl) upgradeKluster(kluster *v1.Kluster, toVersion string) e\n}\n}\n+ if !kluster.Spec.NoCloud && strings.HasPrefix(toVersion, \"1.21\") && strings.HasPrefix(kluster.Status.ApiserverVersion, \"1.20\") {\n+ kubernetes, err := op.Clients.Satellites.ClientFor(kluster)\n+ if err != nil {\n+ return errors.Wrap(err, \"client\")\n+ }\n+\n+ if err := csi.SeedCinderCSIRoles(kubernetes); err != nil {\n+ return errors.Wrap(err, \"seed cinder CSI roles on upgrade\")\n+ }\n+ }\n+\naccessMode, err := util.PVAccessMode(op.Clients.Kubernetes, kluster)\nif err != nil {\nreturn fmt.Errorf(\"Couldn't determine access mode for pvc: %s\", err)\n"
},
{
"change_type": "MODIFY",
"old_path": "pkg/controller/ground/bootstrap/csi/csi.go",
"new_path": "pkg/controller/ground/bootstrap/csi/csi.go",
"diff": "@@ -185,6 +185,50 @@ rescan-on-resize = yes\nreturn nil\n}\n+func SeedCinderCSIRoles(client clientset.Interface) error {\n+ err := createRole(client, CSIRole)\n+ if err != nil {\n+ return errors.Wrap(err, \"CSIRole\")\n+ }\n+\n+ err = createClusterRole(client, CSISnapshotControllerClusterRole)\n+ if err != nil {\n+ return errors.Wrap(err, \"CSISnapshotControllerClusterRole\")\n+ }\n+\n+ err = createClusterRole(client, CSIClusterRoleAttacher)\n+ if err != nil {\n+ return errors.Wrap(err, \"CSIClusterRoleAttacher\")\n+ }\n+\n+ err = createClusterRole(client, CSIClusterRoleNodePlugin)\n+ if err != nil {\n+ return errors.Wrap(err, \"CSIClusterRoleNodePlugin\")\n+ }\n+\n+ err = createClusterRole(client, CSIClusterRoleProvisioner)\n+ if err != nil {\n+ return errors.Wrap(err, \"CSIClusterRoleProvisioner\")\n+ }\n+\n+ err = createClusterRole(client, CSIClusterRoleResizer)\n+ if err != nil {\n+ return errors.Wrap(err, \"CSIClusterRoleResizer\")\n+ }\n+\n+ err = createClusterRole(client, CSIClusterRoleSnapshotter)\n+ if err != nil {\n+ return errors.Wrap(err, \"CSIClusterRoleSnapshotter\")\n+ }\n+\n+ err = createRole(client, CSISnapshotControllerRole)\n+ if err != nil {\n+ return errors.Wrap(err, \"CSISnapshotControllerRole\")\n+ }\n+\n+ return nil\n+}\n+\nfunc createDynamicResource(dynamicClient dynamic.Interface, manifest string, gvr schema.GroupVersionResource) error {\nvar decUnstructured = yaml.NewDecodingSerializer(unstructured.UnstructuredJSONScheme)\nresource := &unstructured.Unstructured{}\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
Seed cinder csi roles on upgrade (#619)
|
596,226 |
30.09.2021 11:06:49
| -7,200 |
8f9df38210d5bc2678cc174d0b6092fccfc8c930
|
Disable flatcar release test
|
[
{
"change_type": "MODIFY",
"old_path": "test/e2e/node_test.go",
"new_path": "test/e2e/node_test.go",
"diff": "@@ -45,7 +45,7 @@ func (k *NodeTests) Run(t *testing.T) {\n_ = t.Run(\"Created\", k.StateRunning) &&\nt.Run(\"Tagged\", k.Tagged) &&\nt.Run(\"Registered\", k.Registered) &&\n- t.Run(\"LatestStableContainerLinux\", k.LatestStableContainerLinux) &&\n+ //t.Run(\"LatestStableContainerLinux\", k.LatestStableContainerLinux) &&\nt.Run(\"Schedulable\", k.StateSchedulable) &&\nt.Run(\"NetworkUnavailable\", k.ConditionNetworkUnavailable) &&\nt.Run(\"Healthy\", k.StateHealthy) &&\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
Disable flatcar release test
|
596,226 |
07.10.2021 21:16:26
| -7,200 |
dec482ade9557077667ae94c3efaf7a32980be71
|
Fix resolv.conf race condition
* Revert "Mount resolv.conf in rkt container (#608)"
This reverts commit
* Test for dns resolution before starting rkt containers
* Add nss-lookup.target to templates
* Compact node template, add unit test
|
[
{
"change_type": "MODIFY",
"old_path": "pkg/templates/ignition.go",
"new_path": "pkg/templates/ignition.go",
"diff": "@@ -27,7 +27,7 @@ var Ignition = &ignition{}\nvar passwordHashRounds = 1000000\n-const TEMPLATE_VERSION = \"2\"\n+const TEMPLATE_VERSION = \"3\"\nfunc (i *ignition) getIgnitionTemplate(kluster *kubernikusv1.Kluster) (string, error) {\nswitch {\n"
},
{
"change_type": "MODIFY",
"old_path": "pkg/templates/ignition_test.go",
"new_path": "pkg/templates/ignition_test.go",
"diff": "@@ -70,6 +70,7 @@ func init() {\nimageRegistry = version.ImageRegistry{\nVersions: map[string]version.KlusterVersion{\n+ \"1.21\": {Kubelet: version.ImageVersion{Repository: \"nase\", Tag: \"v1.21\"}},\n\"1.20\": {Kubelet: version.ImageVersion{Repository: \"nase\", Tag: \"v1.20\"}},\n\"1.19\": {Kubelet: version.ImageVersion{Repository: \"nase\", Tag: \"v1.19\"}},\n\"1.18\": {Hyperkube: version.ImageVersion{Repository: \"nase\", Tag: \"v1.18\"}},\n@@ -110,7 +111,7 @@ func TestGenerateNode(t *testing.T) {\nfunc TestNodeLabels(t *testing.T) {\nkluster := testKluster.DeepCopy()\n- kluster.Spec.Version = \"1.20\"\n+ kluster.Spec.Version = \"1.21\"\npool := &models.NodePool{Name: \"some-name\"}\n"
},
{
"change_type": "MODIFY",
"old_path": "pkg/templates/node_1.10.go",
"new_path": "pkg/templates/node_1.10.go",
"diff": "@@ -49,8 +49,9 @@ systemd:\n[Unit]\nDescription=flannel - Network fabric for containers (System Application Container)\nDocumentation=https://github.com/coreos/flannel\n- After=etcd.service etcd2.service etcd-member.service\n+ After=etcd.service etcd2.service etcd-member.service network-online.target nss-lookup.target\nRequires=flannel-docker-opts.service\n+ Wants=network-online.target nss-lookup.target\n[Service]\nType=notify\n@@ -65,6 +66,7 @@ systemd:\nEnvironment=\"RKT_RUN_ARGS=--uuid-file-save=/var/lib/flatcar/flannel-wrapper.uuid\"\nEnvironmentFile=-/run/flannel/options.env\n+ ExecStartPre=/usr/bin/host identity-3.{{ .OpenstackRegion }}.cloud.sap\nExecStartPre=/sbin/modprobe ip_tables\nExecStartPre=/usr/bin/mkdir --parents /var/lib/flatcar /run/flannel\nExecStartPre=-/opt/bin/rkt rm --uuid-file=/var/lib/flatcar/flannel-wrapper.uuid\n@@ -106,6 +108,8 @@ systemd:\ncontents: |\n[Unit]\nDescription=Kubelet via Hyperkube ACI\n+ After=network-online.target nss-lookup.target\n+ Wants=network-online.target nss-lookup.target\n[Service]\nEnvironment=\"RKT_RUN_ARGS=--uuid-file-save=/var/run/kubelet-pod.uuid \\\n@@ -132,6 +136,7 @@ systemd:\nEnvironment=\"KUBELET_IMAGE_TAG={{ .HyperkubeImageTag }}\"\nEnvironment=\"KUBELET_IMAGE_URL=docker://{{ .HyperkubeImage }}\"\nEnvironment=\"KUBELET_IMAGE_ARGS=--name=kubelet --exec=/kubelet\"\n+ ExecStartPre=/usr/bin/host identity-3.{{ .OpenstackRegion }}.cloud.sap\n{{- if .CalicoNetworking }}\nExecStartPre=/bin/mkdir -p /etc/cni /opt/cni /var/lib/calico\n{{- end }}\n@@ -174,10 +179,11 @@ systemd:\ncontents: |\n[Unit]\nDescription=Kubernikus Wormhole\n- Requires=network-online.target\n- After=network-online.target\n+ After=network-online.target nss-lookup.target\n+ Wants=network-online.target nss-lookup.target\n[Service]\nSlice=machine.slice\n+ ExecStartPre=/usr/bin/host identity-3.{{ .OpenstackRegion }}.cloud.sap\nExecStartPre=/opt/bin/rkt fetch --insecure-options=image --pull-policy=new docker://{{ .KubernikusImage }}:{{ .KubernikusImageTag }}\nExecStart=/opt/bin/rkt run \\\n--inherit-env \\\n@@ -207,10 +213,11 @@ systemd:\ncontents: |\n[Unit]\nDescription=Kube-Proxy\n- Requires=network-online.target\n- After=network-online.target\n+ After=network-online.target nss-lookup.target\n+ Wants=network-online.target nss-lookup.target\n[Service]\nSlice=machine.slice\n+ ExecStartPre=/usr/bin/host identity-3.{{ .OpenstackRegion }}.cloud.sap\nExecStart=/opt/bin/rkt run \\\n--trust-keys-from-https \\\n--inherit-env \\\n"
},
{
"change_type": "MODIFY",
"old_path": "pkg/templates/node_1.11.go",
"new_path": "pkg/templates/node_1.11.go",
"diff": "@@ -49,8 +49,9 @@ systemd:\n[Unit]\nDescription=flannel - Network fabric for containers (System Application Container)\nDocumentation=https://github.com/coreos/flannel\n- After=etcd.service etcd2.service etcd-member.service\n+ After=etcd.service etcd2.service etcd-member.service network-online.target nss-lookup.target\nRequires=flannel-docker-opts.service\n+ Wants=network-online.target nss-lookup.target\n[Service]\nType=notify\n@@ -65,6 +66,7 @@ systemd:\nEnvironment=\"RKT_RUN_ARGS=--uuid-file-save=/var/lib/flatcar/flannel-wrapper.uuid\"\nEnvironmentFile=-/run/flannel/options.env\n+ ExecStartPre=/usr/bin/host identity-3.{{ .OpenstackRegion }}.cloud.sap\nExecStartPre=/sbin/modprobe ip_tables\nExecStartPre=/usr/bin/mkdir --parents /var/lib/flatcar /run/flannel\nExecStartPre=-/opt/bin/rkt rm --uuid-file=/var/lib/flatcar/flannel-wrapper.uuid\n@@ -90,10 +92,7 @@ systemd:\n--volume etc-kubernetes-certs,kind=host,source=/etc/kubernetes/certs,readOnly=true \\\n--mount volume=etc-kubernetes-certs,target=/etc/kubernetes/certs \\\n--volume etc-kube-flannel,kind=host,source=/etc/kube-flannel,readOnly=true \\\n- --mount volume=etc-kube-flannel,target=/etc/kube-flannel \\\n- --dns=host \\\n- --volume dns,kind=host,source=/run/systemd/resolve/resolv.conf,readOnly=true \\\n- --mount volume=dns,target=/etc/resolv.conf\"\n+ --mount volume=etc-kube-flannel,target=/etc/kube-flannel\"\n- name: flannel-docker-opts.service\nenable: true\ncontents: |\n@@ -109,14 +108,14 @@ systemd:\ncontents: |\n[Unit]\nDescription=Kubelet via Hyperkube ACI\n+ After=network-online.target nss-lookup.target\n+ Wants=network-online.target nss-lookup.target\n[Service]\nEnvironment=\"RKT_RUN_ARGS=--uuid-file-save=/var/run/kubelet-pod.uuid \\\n--inherit-env \\\n- --net=host \\\n--dns=host \\\n- --volume dns,kind=host,source=/run/systemd/resolve/resolv.conf,readOnly=true \\\n- --mount volume=dns,target=/etc/resolv.conf \\\n+ --net=host \\\n--volume var-lib-cni,kind=host,source=/var/lib/cni \\\n--volume var-log,kind=host,source=/var/log \\\n--volume etc-machine-id,kind=host,source=/etc/machine-id,readOnly=true \\\n@@ -137,6 +136,7 @@ systemd:\nEnvironment=\"KUBELET_IMAGE_TAG={{ .HyperkubeImageTag }}\"\nEnvironment=\"KUBELET_IMAGE_URL=docker://{{ .HyperkubeImage }}\"\nEnvironment=\"KUBELET_IMAGE_ARGS=--name=kubelet --exec=/kubelet\"\n+ ExecStartPre=/usr/bin/host identity-3.{{ .OpenstackRegion }}.cloud.sap\n{{- if .CalicoNetworking }}\nExecStartPre=/bin/mkdir -p /etc/cni /opt/cni /var/lib/calico\n{{- end }}\n@@ -177,17 +177,16 @@ systemd:\ncontents: |\n[Unit]\nDescription=Kubernikus Wormhole\n- Requires=network-online.target\n- After=network-online.target\n+ After=network-online.target nss-lookup.target\n+ Wants=network-online.target nss-lookup.target\n[Service]\nSlice=machine.slice\n+ ExecStartPre=/usr/bin/host identity-3.{{ .OpenstackRegion }}.cloud.sap\nExecStartPre=/opt/bin/rkt fetch --insecure-options=image --pull-policy=new docker://{{ .KubernikusImage }}:{{ .KubernikusImageTag }}\nExecStart=/opt/bin/rkt run \\\n--inherit-env \\\n--net=host \\\n--dns=host \\\n- --volume dns,kind=host,source=/run/systemd/resolve/resolv.conf,readOnly=true \\\n- --mount volume=dns,target=/etc/resolv.conf \\\n--volume var-lib-kubelet,kind=host,source=/var/lib/kubelet,readOnly=true \\\n--mount volume=var-lib-kubelet,target=/var/lib/kubelet \\\n--volume etc-kubernetes-certs,kind=host,source=/etc/kubernetes/certs,readOnly=true \\\n@@ -212,17 +211,16 @@ systemd:\ncontents: |\n[Unit]\nDescription=Kube-Proxy\n- Requires=network-online.target\n- After=network-online.target\n+ After=network-online.target nss-lookup.target\n+ Wants=network-online.target nss-lookup.target\n[Service]\nSlice=machine.slice\n+ ExecStartPre=/usr/bin/host identity-3.{{ .OpenstackRegion }}.cloud.sap\nExecStart=/opt/bin/rkt run \\\n--trust-keys-from-https \\\n--inherit-env \\\n--net=host \\\n--dns=host \\\n- --volume dns,kind=host,source=/run/systemd/resolve/resolv.conf,readOnly=true \\\n- --mount volume=dns,target=/etc/resolv.conf \\\n--volume etc-kubernetes,kind=host,source=/etc/kubernetes,readOnly=true \\\n--mount volume=etc-kubernetes,target=/etc/kubernetes \\\n--volume lib-modules,kind=host,source=/lib/modules,readOnly=true \\\n"
},
{
"change_type": "MODIFY",
"old_path": "pkg/templates/node_1.12.go",
"new_path": "pkg/templates/node_1.12.go",
"diff": "@@ -49,8 +49,9 @@ systemd:\n[Unit]\nDescription=flannel - Network fabric for containers (System Application Container)\nDocumentation=https://github.com/coreos/flannel\n- After=etcd.service etcd2.service etcd-member.service\n+ After=etcd.service etcd2.service etcd-member.service network-online.target nss-lookup.target\nRequires=flannel-docker-opts.service\n+ Wants=network-online.target nss-lookup.target\n[Service]\nType=notify\n@@ -65,6 +66,7 @@ systemd:\nEnvironment=\"RKT_RUN_ARGS=--uuid-file-save=/var/lib/flatcar/flannel-wrapper.uuid\"\nEnvironmentFile=-/run/flannel/options.env\n+ ExecStartPre=/usr/bin/host identity-3.{{ .OpenstackRegion }}.cloud.sap\nExecStartPre=/sbin/modprobe ip_tables\nExecStartPre=/usr/bin/mkdir --parents /var/lib/flatcar /run/flannel\nExecStartPre=-/opt/bin/rkt rm --uuid-file=/var/lib/flatcar/flannel-wrapper.uuid\n@@ -90,10 +92,7 @@ systemd:\n--volume etc-kubernetes-certs,kind=host,source=/etc/kubernetes/certs,readOnly=true \\\n--mount volume=etc-kubernetes-certs,target=/etc/kubernetes/certs \\\n--volume etc-kube-flannel,kind=host,source=/etc/kube-flannel,readOnly=true \\\n- --mount volume=etc-kube-flannel,target=/etc/kube-flannel \\\n- --dns=host \\\n- --volume dns,kind=host,source=/run/systemd/resolve/resolv.conf,readOnly=true \\\n- --mount volume=dns,target=/etc/resolv.conf\"\n+ --mount volume=etc-kube-flannel,target=/etc/kube-flannel\"\n- name: flannel-docker-opts.service\nenable: true\ncontents: |\n@@ -109,14 +108,14 @@ systemd:\ncontents: |\n[Unit]\nDescription=Kubelet via Hyperkube ACI\n+ After=network-online.target nss-lookup.target\n+ Wants=network-online.target nss-lookup.target\n[Service]\nEnvironment=\"RKT_RUN_ARGS=--uuid-file-save=/var/run/kubelet-pod.uuid \\\n--inherit-env \\\n- --net=host \\\n--dns=host \\\n- --volume dns,kind=host,source=/run/systemd/resolve/resolv.conf,readOnly=true \\\n- --mount volume=dns,target=/etc/resolv.conf \\\n+ --net=host \\\n--volume var-lib-cni,kind=host,source=/var/lib/cni \\\n--volume var-log,kind=host,source=/var/log \\\n--volume etc-machine-id,kind=host,source=/etc/machine-id,readOnly=true \\\n@@ -137,6 +136,7 @@ systemd:\nEnvironment=\"KUBELET_IMAGE_TAG={{ .HyperkubeImageTag }}\"\nEnvironment=\"KUBELET_IMAGE_URL=docker://{{ .HyperkubeImage }}\"\nEnvironment=\"KUBELET_IMAGE_ARGS=--name=kubelet --exec=/kubelet\"\n+ ExecStartPre=/usr/bin/host identity-3.{{ .OpenstackRegion }}.cloud.sap\n{{- if .CalicoNetworking }}\nExecStartPre=/bin/mkdir -p /etc/cni /opt/cni /var/lib/calico\n{{- end }}\n@@ -179,17 +179,16 @@ systemd:\ncontents: |\n[Unit]\nDescription=Kubernikus Wormhole\n- Requires=network-online.target\n- After=network-online.target\n+ After=network-online.target nss-lookup.target\n+ Wants=network-online.target nss-lookup.target\n[Service]\nSlice=machine.slice\n+ ExecStartPre=/usr/bin/host identity-3.{{ .OpenstackRegion }}.cloud.sap\nExecStartPre=/opt/bin/rkt fetch --insecure-options=image --pull-policy=new docker://{{ .KubernikusImage }}:{{ .KubernikusImageTag }}\nExecStart=/opt/bin/rkt run \\\n--inherit-env \\\n--net=host \\\n--dns=host \\\n- --volume dns,kind=host,source=/run/systemd/resolve/resolv.conf,readOnly=true \\\n- --mount volume=dns,target=/etc/resolv.conf \\\n--volume var-lib-kubelet,kind=host,source=/var/lib/kubelet,readOnly=true \\\n--mount volume=var-lib-kubelet,target=/var/lib/kubelet \\\n--volume etc-kubernetes-certs,kind=host,source=/etc/kubernetes/certs,readOnly=true \\\n@@ -214,17 +213,16 @@ systemd:\ncontents: |\n[Unit]\nDescription=Kube-Proxy\n- Requires=network-online.target\n- After=network-online.target\n+ After=network-online.target nss-lookup.target\n+ Wants=network-online.target nss-lookup.target\n[Service]\nSlice=machine.slice\n+ ExecStartPre=/usr/bin/host identity-3.{{ .OpenstackRegion }}.cloud.sap\nExecStart=/opt/bin/rkt run \\\n--trust-keys-from-https \\\n--inherit-env \\\n--net=host \\\n--dns=host \\\n- --volume dns,kind=host,source=/run/systemd/resolve/resolv.conf,readOnly=true \\\n- --mount volume=dns,target=/etc/resolv.conf \\\n--volume etc-kubernetes,kind=host,source=/etc/kubernetes,readOnly=true \\\n--mount volume=etc-kubernetes,target=/etc/kubernetes \\\n--volume lib-modules,kind=host,source=/lib/modules,readOnly=true \\\n"
},
{
"change_type": "MODIFY",
"old_path": "pkg/templates/node_1.14.go",
"new_path": "pkg/templates/node_1.14.go",
"diff": "@@ -49,8 +49,9 @@ systemd:\n[Unit]\nDescription=flannel - Network fabric for containers (System Application Container)\nDocumentation=https://github.com/coreos/flannel\n- After=etcd.service etcd2.service etcd-member.service\n+ After=etcd.service etcd2.service etcd-member.service network-online.target nss-lookup.target\nRequires=flannel-docker-opts.service\n+ Wants=network-online.target nss-lookup.target\n[Service]\nType=notify\n@@ -65,6 +66,7 @@ systemd:\nEnvironment=\"RKT_RUN_ARGS=--uuid-file-save=/var/lib/flatcar/flannel-wrapper.uuid\"\nEnvironmentFile=-/run/flannel/options.env\n+ ExecStartPre=/usr/bin/host identity-3.{{ .OpenstackRegion }}.cloud.sap\nExecStartPre=/sbin/modprobe ip_tables\nExecStartPre=/usr/bin/mkdir --parents /var/lib/flatcar /run/flannel\nExecStartPre=-/opt/bin/rkt rm --uuid-file=/var/lib/flatcar/flannel-wrapper.uuid\n@@ -90,10 +92,7 @@ systemd:\n--volume etc-kubernetes-certs,kind=host,source=/etc/kubernetes/certs,readOnly=true \\\n--mount volume=etc-kubernetes-certs,target=/etc/kubernetes/certs \\\n--volume etc-kube-flannel,kind=host,source=/etc/kube-flannel,readOnly=true \\\n- --mount volume=etc-kube-flannel,target=/etc/kube-flannel \\\n- --dns=host \\\n- --volume dns,kind=host,source=/run/systemd/resolve/resolv.conf,readOnly=true \\\n- --mount volume=dns,target=/etc/resolv.conf\"\n+ --mount volume=etc-kube-flannel,target=/etc/kube-flannel\"\n- name: flannel-docker-opts.service\nenable: true\ncontents: |\n@@ -109,14 +108,14 @@ systemd:\ncontents: |\n[Unit]\nDescription=Kubelet via Hyperkube ACI\n+ After=network-online.target nss-lookup.target\n+ Wants=network-online.target nss-lookup.target\n[Service]\nEnvironment=\"RKT_RUN_ARGS=--uuid-file-save=/var/run/kubelet-pod.uuid \\\n--inherit-env \\\n- --net=host \\\n--dns=host \\\n- --volume dns,kind=host,source=/run/systemd/resolve/resolv.conf,readOnly=true \\\n- --mount volume=dns,target=/etc/resolv.conf \\\n+ --net=host \\\n--volume var-lib-cni,kind=host,source=/var/lib/cni \\\n--volume var-log,kind=host,source=/var/log \\\n--volume etc-machine-id,kind=host,source=/etc/machine-id,readOnly=true \\\n@@ -137,6 +136,7 @@ systemd:\nEnvironment=\"KUBELET_IMAGE_TAG={{ .HyperkubeImageTag }}\"\nEnvironment=\"KUBELET_IMAGE_URL=docker://{{ .HyperkubeImage }}\"\nEnvironment=\"KUBELET_IMAGE_ARGS=--name=kubelet --exec=/kubelet\"\n+ ExecStartPre=/usr/bin/host identity-3.{{ .OpenstackRegion }}.cloud.sap\n{{- if .CalicoNetworking }}\nExecStartPre=/bin/mkdir -p /etc/cni /opt/cni /var/lib/calico\n{{- end }}\n@@ -179,17 +179,16 @@ systemd:\ncontents: |\n[Unit]\nDescription=Kubernikus Wormhole\n- Requires=network-online.target\n- After=network-online.target\n+ After=network-online.target nss-lookup.target\n+ Wants=network-online.target nss-lookup.target\n[Service]\nSlice=machine.slice\n+ ExecStartPre=/usr/bin/host identity-3.{{ .OpenstackRegion }}.cloud.sap\nExecStartPre=/opt/bin/rkt fetch --insecure-options=image --pull-policy=new docker://{{ .KubernikusImage }}:{{ .KubernikusImageTag }}\nExecStart=/opt/bin/rkt run \\\n--inherit-env \\\n--net=host \\\n--dns=host \\\n- --volume dns,kind=host,source=/run/systemd/resolve/resolv.conf,readOnly=true \\\n- --mount volume=dns,target=/etc/resolv.conf \\\n--volume var-lib-kubelet,kind=host,source=/var/lib/kubelet,readOnly=true \\\n--mount volume=var-lib-kubelet,target=/var/lib/kubelet \\\n--volume etc-kubernetes-certs,kind=host,source=/etc/kubernetes/certs,readOnly=true \\\n@@ -214,17 +213,16 @@ systemd:\ncontents: |\n[Unit]\nDescription=Kube-Proxy\n- Requires=network-online.target\n- After=network-online.target\n+ After=network-online.target nss-lookup.target\n+ Wants=network-online.target nss-lookup.target\n[Service]\nSlice=machine.slice\n+ ExecStartPre=/usr/bin/host identity-3.{{ .OpenstackRegion }}.cloud.sap\nExecStart=/opt/bin/rkt run \\\n--trust-keys-from-https \\\n--inherit-env \\\n--net=host \\\n--dns=host \\\n- --volume dns,kind=host,source=/run/systemd/resolve/resolv.conf,readOnly=true \\\n- --mount volume=dns,target=/etc/resolv.conf \\\n--volume etc-kubernetes,kind=host,source=/etc/kubernetes,readOnly=true \\\n--mount volume=etc-kubernetes,target=/etc/kubernetes \\\n--volume lib-modules,kind=host,source=/lib/modules,readOnly=true \\\n"
},
{
"change_type": "MODIFY",
"old_path": "pkg/templates/node_1.17.go",
"new_path": "pkg/templates/node_1.17.go",
"diff": "@@ -49,8 +49,9 @@ systemd:\n[Unit]\nDescription=flannel - Network fabric for containers (System Application Container)\nDocumentation=https://github.com/coreos/flannel\n- After=etcd.service etcd2.service etcd-member.service\n+ After=etcd.service etcd2.service etcd-member.service network-online.target nss-lookup.target\nRequires=flannel-docker-opts.service\n+ Wants=network-online.target nss-lookup.target\n[Service]\nType=notify\n@@ -65,6 +66,7 @@ systemd:\nEnvironment=\"RKT_RUN_ARGS=--uuid-file-save=/var/lib/flatcar/flannel-wrapper.uuid\"\nEnvironmentFile=-/run/flannel/options.env\n+ ExecStartPre=/usr/bin/host identity-3.{{ .OpenstackRegion }}.cloud.sap\nExecStartPre=/sbin/modprobe ip_tables\nExecStartPre=/usr/bin/mkdir --parents /var/lib/flatcar /run/flannel\nExecStartPre=-/opt/bin/rkt rm --uuid-file=/var/lib/flatcar/flannel-wrapper.uuid\n@@ -90,10 +92,7 @@ systemd:\n--volume etc-kubernetes-certs,kind=host,source=/etc/kubernetes/certs,readOnly=true \\\n--mount volume=etc-kubernetes-certs,target=/etc/kubernetes/certs \\\n--volume etc-kube-flannel,kind=host,source=/etc/kube-flannel,readOnly=true \\\n- --mount volume=etc-kube-flannel,target=/etc/kube-flannel \\\n- --dns=host \\\n- --volume dns,kind=host,source=/run/systemd/resolve/resolv.conf,readOnly=true \\\n- --mount volume=dns,target=/etc/resolv.conf\"\n+ --mount volume=etc-kube-flannel,target=/etc/kube-flannel\"\n- name: flannel-docker-opts.service\nenable: true\ncontents: |\n@@ -109,14 +108,14 @@ systemd:\ncontents: |\n[Unit]\nDescription=Kubelet via Hyperkube ACI\n+ After=network-online.target nss-lookup.target\n+ Wants=network-online.target nss-lookup.target\n[Service]\nEnvironment=\"RKT_RUN_ARGS=--uuid-file-save=/var/run/kubelet-pod.uuid \\\n--inherit-env \\\n- --net=host \\\n--dns=host \\\n- --volume dns,kind=host,source=/run/systemd/resolve/resolv.conf,readOnly=true \\\n- --mount volume=dns,target=/etc/resolv.conf \\\n+ --net=host \\\n--volume var-lib-cni,kind=host,source=/var/lib/cni \\\n--volume var-log,kind=host,source=/var/log \\\n--volume etc-machine-id,kind=host,source=/etc/machine-id,readOnly=true \\\n@@ -137,6 +136,7 @@ systemd:\nEnvironment=\"KUBELET_IMAGE_TAG={{ .HyperkubeImageTag }}\"\nEnvironment=\"KUBELET_IMAGE_URL=docker://{{ .HyperkubeImage }}\"\nEnvironment=\"KUBELET_IMAGE_ARGS=--name=kubelet --exec=/usr/local/bin/kubelet\"\n+ ExecStartPre=/usr/bin/host identity-3.{{ .OpenstackRegion }}.cloud.sap\n{{- if .CalicoNetworking }}\nExecStartPre=/bin/mkdir -p /etc/cni /opt/cni /var/lib/calico\n{{- end }}\n@@ -179,17 +179,16 @@ systemd:\ncontents: |\n[Unit]\nDescription=Kubernikus Wormhole\n- Requires=network-online.target\n- After=network-online.target\n+ After=network-online.target nss-lookup.target\n+ Wants=network-online.target nss-lookup.target\n[Service]\nSlice=machine.slice\n+ ExecStartPre=/usr/bin/host identity-3.{{ .OpenstackRegion }}.cloud.sap\nExecStartPre=/opt/bin/rkt fetch --insecure-options=image --pull-policy=new docker://{{ .KubernikusImage }}:{{ .KubernikusImageTag }}\nExecStart=/opt/bin/rkt run \\\n--inherit-env \\\n--net=host \\\n--dns=host \\\n- --volume dns,kind=host,source=/run/systemd/resolve/resolv.conf,readOnly=true \\\n- --mount volume=dns,target=/etc/resolv.conf \\\n--volume var-lib-kubelet,kind=host,source=/var/lib/kubelet,readOnly=true \\\n--mount volume=var-lib-kubelet,target=/var/lib/kubelet \\\n--volume etc-kubernetes-certs,kind=host,source=/etc/kubernetes/certs,readOnly=true \\\n@@ -214,17 +213,16 @@ systemd:\ncontents: |\n[Unit]\nDescription=Kube-Proxy\n- Requires=network-online.target\n- After=network-online.target\n+ After=network-online.target nss-lookup.target\n+ Wants=network-online.target nss-lookup.target\n[Service]\nSlice=machine.slice\n+ ExecStartPre=/usr/bin/host identity-3.{{ .OpenstackRegion }}.cloud.sap\nExecStart=/opt/bin/rkt run \\\n--trust-keys-from-https \\\n--inherit-env \\\n--net=host \\\n--dns=host \\\n- --volume dns,kind=host,source=/run/systemd/resolve/resolv.conf,readOnly=true \\\n- --mount volume=dns,target=/etc/resolv.conf \\\n--volume etc-kubernetes,kind=host,source=/etc/kubernetes,readOnly=true \\\n--mount volume=etc-kubernetes,target=/etc/kubernetes \\\n--volume lib-modules,kind=host,source=/lib/modules,readOnly=true \\\n"
},
{
"change_type": "MODIFY",
"old_path": "pkg/templates/node_1.19.go",
"new_path": "pkg/templates/node_1.19.go",
"diff": "@@ -49,8 +49,9 @@ systemd:\n[Unit]\nDescription=flannel - Network fabric for containers (System Application Container)\nDocumentation=https://github.com/coreos/flannel\n- After=etcd.service etcd2.service etcd-member.service\n+ After=etcd.service etcd2.service etcd-member.service network-online.target nss-lookup.target\nRequires=flannel-docker-opts.service\n+ Wants=network-online.target nss-lookup.target\n[Service]\nType=notify\n@@ -65,6 +66,7 @@ systemd:\nEnvironment=\"RKT_RUN_ARGS=--uuid-file-save=/var/lib/flatcar/flannel-wrapper.uuid\"\nEnvironmentFile=-/run/flannel/options.env\n+ ExecStartPre=/usr/bin/host identity-3.{{ .OpenstackRegion }}.cloud.sap\nExecStartPre=/sbin/modprobe ip_tables\nExecStartPre=/usr/bin/mkdir --parents /var/lib/flatcar /run/flannel\nExecStartPre=-/opt/bin/rkt rm --uuid-file=/var/lib/flatcar/flannel-wrapper.uuid\n@@ -90,10 +92,7 @@ systemd:\n--volume etc-kubernetes-certs,kind=host,source=/etc/kubernetes/certs,readOnly=true \\\n--mount volume=etc-kubernetes-certs,target=/etc/kubernetes/certs \\\n--volume etc-kube-flannel,kind=host,source=/etc/kube-flannel,readOnly=true \\\n- --mount volume=etc-kube-flannel,target=/etc/kube-flannel \\\n- --dns=host \\\n- --volume dns,kind=host,source=/run/systemd/resolve/resolv.conf,readOnly=true \\\n- --mount volume=dns,target=/etc/resolv.conf\"\n+ --mount volume=etc-kube-flannel,target=/etc/kube-flannel\"\n- name: flannel-docker-opts.service\nenable: true\ncontents: |\n@@ -109,14 +108,14 @@ systemd:\ncontents: |\n[Unit]\nDescription=Kubelet\n+ After=network-online.target nss-lookup.target\n+ Wants=network-online.target nss-lookup.target\n[Service]\nEnvironment=\"RKT_RUN_ARGS=--uuid-file-save=/var/run/kubelet-pod.uuid \\\n--inherit-env \\\n- --net=host \\\n--dns=host \\\n- --volume dns,kind=host,source=/run/systemd/resolve/resolv.conf,readOnly=true \\\n- --mount volume=dns,target=/etc/resolv.conf \\\n+ --net=host \\\n--volume var-lib-cni,kind=host,source=/var/lib/cni \\\n--volume var-log,kind=host,source=/var/log \\\n--volume etc-machine-id,kind=host,source=/etc/machine-id,readOnly=true \\\n@@ -137,6 +136,7 @@ systemd:\nEnvironment=\"KUBELET_IMAGE_TAG={{ .KubeletImageTag }}\"\nEnvironment=\"KUBELET_IMAGE_URL=docker://{{ .KubeletImage }}\"\nEnvironment=\"KUBELET_IMAGE_ARGS=--name=kubelet --exec=/usr/local/bin/kubelet\"\n+ ExecStartPre=/usr/bin/host identity-3.{{ .OpenstackRegion }}.cloud.sap\n{{- if .CalicoNetworking }}\nExecStartPre=/bin/mkdir -p /etc/cni /opt/cni /var/lib/calico\n{{- end }}\n@@ -179,17 +179,16 @@ systemd:\ncontents: |\n[Unit]\nDescription=Kubernikus Wormhole\n- Requires=network-online.target\n- After=network-online.target\n+ After=network-online.target nss-lookup.target\n+ Wants=network-online.target nss-lookup.target\n[Service]\nSlice=machine.slice\n+ ExecStartPre=/usr/bin/host identity-3.{{ .OpenstackRegion }}.cloud.sap\nExecStartPre=/opt/bin/rkt fetch --insecure-options=image --pull-policy=new docker://{{ .KubernikusImage }}:{{ .KubernikusImageTag }}\nExecStart=/opt/bin/rkt run \\\n--inherit-env \\\n--net=host \\\n--dns=host \\\n- --volume dns,kind=host,source=/run/systemd/resolve/resolv.conf,readOnly=true \\\n- --mount volume=dns,target=/etc/resolv.conf \\\n--volume var-lib-kubelet,kind=host,source=/var/lib/kubelet,readOnly=true \\\n--mount volume=var-lib-kubelet,target=/var/lib/kubelet \\\n--volume etc-kubernetes-certs,kind=host,source=/etc/kubernetes/certs,readOnly=true \\\n@@ -214,17 +213,16 @@ systemd:\ncontents: |\n[Unit]\nDescription=Kube-Proxy\n- Requires=network-online.target\n- After=network-online.target\n+ After=network-online.target nss-lookup.target\n+ Wants=network-online.target nss-lookup.target\n[Service]\nSlice=machine.slice\n+ ExecStartPre=/usr/bin/host identity-3.{{ .OpenstackRegion }}.cloud.sap\nExecStart=/opt/bin/rkt run \\\n--trust-keys-from-https \\\n--inherit-env \\\n--net=host \\\n--dns=host \\\n- --volume dns,kind=host,source=/run/systemd/resolve/resolv.conf,readOnly=true \\\n- --mount volume=dns,target=/etc/resolv.conf \\\n--volume etc-kubernetes,kind=host,source=/etc/kubernetes,readOnly=true \\\n--mount volume=etc-kubernetes,target=/etc/kubernetes \\\n--volume lib-modules,kind=host,source=/lib/modules,readOnly=true \\\n"
},
{
"change_type": "MODIFY",
"old_path": "pkg/templates/node_1.20.go",
"new_path": "pkg/templates/node_1.20.go",
"diff": "@@ -49,8 +49,9 @@ systemd:\n[Unit]\nDescription=flannel - Network fabric for containers (System Application Container)\nDocumentation=https://github.com/coreos/flannel\n- After=etcd.service etcd2.service etcd-member.service\n+ After=etcd.service etcd2.service etcd-member.service network-online.target nss-lookup.target\nRequires=flannel-docker-opts.service\n+ Wants=network-online.target nss-lookup.target\n[Service]\nType=notify\n@@ -65,6 +66,7 @@ systemd:\nEnvironment=\"RKT_RUN_ARGS=--uuid-file-save=/var/lib/flatcar/flannel-wrapper.uuid\"\nEnvironmentFile=-/run/flannel/options.env\n+ ExecStartPre=/usr/bin/host identity-3.{{ .OpenstackRegion }}.cloud.sap\nExecStartPre=/sbin/modprobe ip_tables\nExecStartPre=/usr/bin/mkdir --parents /var/lib/flatcar /run/flannel\nExecStartPre=-/opt/bin/rkt rm --uuid-file=/var/lib/flatcar/flannel-wrapper.uuid\n@@ -90,10 +92,7 @@ systemd:\n--volume etc-kubernetes-certs,kind=host,source=/etc/kubernetes/certs,readOnly=true \\\n--mount volume=etc-kubernetes-certs,target=/etc/kubernetes/certs \\\n--volume etc-kube-flannel,kind=host,source=/etc/kube-flannel,readOnly=true \\\n- --mount volume=etc-kube-flannel,target=/etc/kube-flannel \\\n- --dns=host \\\n- --volume dns,kind=host,source=/run/systemd/resolve/resolv.conf,readOnly=true \\\n- --mount volume=dns,target=/etc/resolv.conf\"\n+ --mount volume=etc-kube-flannel,target=/etc/kube-flannel\"\n- name: flannel-docker-opts.service\nenable: true\ncontents: |\n@@ -109,16 +108,14 @@ systemd:\ncontents: |\n[Unit]\nDescription=Kubelet\n- After=network-online.target\n- Wants=network-online.target\n+ After=network-online.target nss-lookup.target\n+ Wants=network-online.target nss-lookup.target\n[Service]\nEnvironment=\"RKT_RUN_ARGS=--uuid-file-save=/var/run/kubelet-pod.uuid \\\n--inherit-env \\\n- --net=host \\\n--dns=host \\\n- --volume dns,kind=host,source=/run/systemd/resolve/resolv.conf,readOnly=true \\\n- --mount volume=dns,target=/etc/resolv.conf \\\n+ --net=host \\\n--volume var-lib-cni,kind=host,source=/var/lib/cni \\\n--volume var-log,kind=host,source=/var/log \\\n--volume etc-machine-id,kind=host,source=/etc/machine-id,readOnly=true \\\n@@ -139,6 +136,7 @@ systemd:\nEnvironment=\"KUBELET_IMAGE_TAG={{ .KubeletImageTag }}\"\nEnvironment=\"KUBELET_IMAGE_URL=docker://{{ .KubeletImage }}\"\nEnvironment=\"KUBELET_IMAGE_ARGS=--name=kubelet --exec=/usr/local/bin/kubelet\"\n+ ExecStartPre=/usr/bin/host identity-3.{{ .OpenstackRegion }}.cloud.sap\n{{- if .CalicoNetworking }}\nExecStartPre=/bin/mkdir -p /etc/cni /opt/cni /var/lib/calico\n{{- end }}\n@@ -180,17 +178,16 @@ systemd:\ncontents: |\n[Unit]\nDescription=Kubernikus Wormhole\n- Requires=network-online.target\n- After=network-online.target\n+ After=network-online.target nss-lookup.target\n+ Wants=network-online.target nss-lookup.target\n[Service]\nSlice=machine.slice\n+ ExecStartPre=/usr/bin/host identity-3.{{ .OpenstackRegion }}.cloud.sap\nExecStartPre=/opt/bin/rkt fetch --insecure-options=image --pull-policy=new docker://{{ .KubernikusImage }}:{{ .KubernikusImageTag }}\nExecStart=/opt/bin/rkt run \\\n--inherit-env \\\n--net=host \\\n--dns=host \\\n- --volume dns,kind=host,source=/run/systemd/resolve/resolv.conf,readOnly=true \\\n- --mount volume=dns,target=/etc/resolv.conf \\\n--volume var-lib-kubelet,kind=host,source=/var/lib/kubelet,readOnly=true \\\n--mount volume=var-lib-kubelet,target=/var/lib/kubelet \\\n--volume etc-kubernetes-certs,kind=host,source=/etc/kubernetes/certs,readOnly=true \\\n@@ -215,17 +212,16 @@ systemd:\ncontents: |\n[Unit]\nDescription=Kube-Proxy\n- Requires=network-online.target\n- After=network-online.target\n+ After=network-online.target nss-lookup.target\n+ Wants=network-online.target nss-lookup.target\n[Service]\nSlice=machine.slice\n+ ExecStartPre=/usr/bin/host identity-3.{{ .OpenstackRegion }}.cloud.sap\nExecStart=/opt/bin/rkt run \\\n--trust-keys-from-https \\\n--inherit-env \\\n--net=host \\\n--dns=host \\\n- --volume dns,kind=host,source=/run/systemd/resolve/resolv.conf,readOnly=true \\\n- --mount volume=dns,target=/etc/resolv.conf \\\n--volume etc-kubernetes,kind=host,source=/etc/kubernetes,readOnly=true \\\n--mount volume=etc-kubernetes,target=/etc/kubernetes \\\n--volume lib-modules,kind=host,source=/lib/modules,readOnly=true \\\n"
},
{
"change_type": "MODIFY",
"old_path": "pkg/templates/node_1.21.go",
"new_path": "pkg/templates/node_1.21.go",
"diff": "@@ -15,7 +15,6 @@ passwd:\ngroups:\n- name: rkt\nsystem: true\n-\nsystemd:\nunits:\n- name: iptables-restore.service\n@@ -25,7 +24,6 @@ systemd:\ncontents: |\n[Unit]\nDescription=Workaround for coreos-metadata hostname bug\n-\n[Service]\nExecStartPre=/usr/bin/curl -s http://169.254.169.254/latest/meta-data/hostname\nExecStartPre=/usr/bin/bash -c \"/usr/bin/systemctl set-environment COREOS_OPENSTACK_HOSTNAME=$(curl -s http://169.254.169.254/latest/meta-data/hostname)\"\n@@ -33,7 +31,6 @@ systemd:\nRestart=on-failure\nRestartSec=5\nRemainAfterExit=yes\n-\n[Install]\nWantedBy=multi-user.target\n- name: docker.service\n@@ -49,9 +46,9 @@ systemd:\n[Unit]\nDescription=flannel - Network fabric for containers (System Application Container)\nDocumentation=https://github.com/coreos/flannel\n- After=etcd.service etcd2.service etcd-member.service\n+ After=etcd.service etcd2.service etcd-member.service network-online.target nss-lookup.target\nRequires=flannel-docker-opts.service\n-\n+ Wants=network-online.target nss-lookup.target\n[Service]\nType=notify\nRestart=always\n@@ -59,18 +56,16 @@ systemd:\nTimeoutStartSec=300\nLimitNOFILE=40000\nLimitNPROC=1048576\n-\nEnvironment=\"FLANNEL_IMAGE_TAG=v0.12.0\"\nEnvironment=\"FLANNEL_OPTS=--ip-masq=true\"\nEnvironment=\"RKT_RUN_ARGS=--uuid-file-save=/var/lib/flatcar/flannel-wrapper.uuid\"\nEnvironmentFile=-/run/flannel/options.env\n-\n+ ExecStartPre=/usr/bin/host identity-3.{{ .OpenstackRegion }}.cloud.sap\nExecStartPre=/sbin/modprobe ip_tables\nExecStartPre=/usr/bin/mkdir --parents /var/lib/flatcar /run/flannel\nExecStartPre=-/opt/bin/rkt rm --uuid-file=/var/lib/flatcar/flannel-wrapper.uuid\nExecStart=/opt/bin/flannel-wrapper $FLANNEL_OPTS\nExecStop=-/opt/bin/rkt stop --uuid-file=/var/lib/flatcar/flannel-wrapper.uuid\n-\n[Install]\nWantedBy=multi-user.target\n- name: flanneld.service\n@@ -90,10 +85,7 @@ systemd:\n--volume etc-kubernetes-certs,kind=host,source=/etc/kubernetes/certs,readOnly=true \\\n--mount volume=etc-kubernetes-certs,target=/etc/kubernetes/certs \\\n--volume etc-kube-flannel,kind=host,source=/etc/kube-flannel,readOnly=true \\\n- --mount volume=etc-kube-flannel,target=/etc/kube-flannel \\\n- --dns=host \\\n- --volume dns,kind=host,source=/run/systemd/resolve/resolv.conf,readOnly=true \\\n- --mount volume=dns,target=/etc/resolv.conf\"\n+ --mount volume=etc-kube-flannel,target=/etc/kube-flannel\"\n- name: flannel-docker-opts.service\nenable: true\ncontents: |\n@@ -109,16 +101,13 @@ systemd:\ncontents: |\n[Unit]\nDescription=Kubelet\n- After=network-online.target\n- Wants=network-online.target\n-\n+ After=network-online.target nss-lookup.target\n+ Wants=network-online.target nss-lookup.target\n[Service]\nEnvironment=\"RKT_RUN_ARGS=--uuid-file-save=/var/run/kubelet-pod.uuid \\\n--inherit-env \\\n--net=host \\\n--dns=host \\\n- --volume dns,kind=host,source=/run/systemd/resolve/resolv.conf,readOnly=true \\\n- --mount volume=dns,target=/etc/resolv.conf \\\n--volume var-lib-cni,kind=host,source=/var/lib/cni \\\n--volume var-log,kind=host,source=/var/log \\\n--volume etc-machine-id,kind=host,source=/etc/machine-id,readOnly=true \\\n@@ -139,6 +128,7 @@ systemd:\nEnvironment=\"KUBELET_IMAGE_TAG={{ .KubeletImageTag }}\"\nEnvironment=\"KUBELET_IMAGE_URL=docker://{{ .KubeletImage }}\"\nEnvironment=\"KUBELET_IMAGE_ARGS=--name=kubelet --exec=/usr/local/bin/kubelet\"\n+ ExecStartPre=/usr/bin/host identity-3.{{ .OpenstackRegion }}.cloud.sap\n{{- if .CalicoNetworking }}\nExecStartPre=/bin/mkdir -p /etc/cni /opt/cni /var/lib/calico\n{{- end }}\n@@ -173,24 +163,22 @@ systemd:\nExecStop=-/opt/bin/rkt stop --uuid-file=/var/run/kubelet-pod.uuid\nRestart=always\nRestartSec=10\n-\n[Install]\nWantedBy=multi-user.target\n- name: wormhole.service\ncontents: |\n[Unit]\nDescription=Kubernikus Wormhole\n- Requires=network-online.target\n- After=network-online.target\n+ After=network-online.target nss-lookup.target\n+ Wants=network-online.target nss-lookup.target\n[Service]\nSlice=machine.slice\n+ ExecStartPre=/usr/bin/host identity-3.{{ .OpenstackRegion }}.cloud.sap\nExecStartPre=/opt/bin/rkt fetch --insecure-options=image --pull-policy=new docker://{{ .KubernikusImage }}:{{ .KubernikusImageTag }}\nExecStart=/opt/bin/rkt run \\\n--inherit-env \\\n--net=host \\\n--dns=host \\\n- --volume dns,kind=host,source=/run/systemd/resolve/resolv.conf,readOnly=true \\\n- --mount volume=dns,target=/etc/resolv.conf \\\n--volume var-lib-kubelet,kind=host,source=/var/lib/kubelet,readOnly=true \\\n--mount volume=var-lib-kubelet,target=/var/lib/kubelet \\\n--volume etc-kubernetes-certs,kind=host,source=/etc/kubernetes/certs,readOnly=true \\\n@@ -215,17 +203,16 @@ systemd:\ncontents: |\n[Unit]\nDescription=Kube-Proxy\n- Requires=network-online.target\n- After=network-online.target\n+ After=network-online.target nss-lookup.target\n+ Wants=network-online.target nss-lookup.target\n[Service]\nSlice=machine.slice\n+ ExecStartPre=/usr/bin/host identity-3.{{ .OpenstackRegion }}.cloud.sap\nExecStart=/opt/bin/rkt run \\\n--trust-keys-from-https \\\n--inherit-env \\\n--net=host \\\n--dns=host \\\n- --volume dns,kind=host,source=/run/systemd/resolve/resolv.conf,readOnly=true \\\n- --mount volume=dns,target=/etc/resolv.conf \\\n--volume etc-kubernetes,kind=host,source=/etc/kubernetes,readOnly=true \\\n--mount volume=etc-kubernetes,target=/etc/kubernetes \\\n--volume lib-modules,kind=host,source=/lib/modules,readOnly=true \\\n@@ -259,7 +246,6 @@ systemd:\ncontents: |\n[Unit]\nDescription=Garbage Collection for rkt\n-\n[Service]\nEnvironment=GRACE_PERIOD=24h\nType=oneshot\n@@ -270,14 +256,11 @@ systemd:\ncontents: |\n[Unit]\nDescription=Periodic Garbage Collection for rkt\n-\n[Timer]\nOnActiveSec=0s\nOnUnitActiveSec=12h\n-\n[Install]\nWantedBy=multi-user.target\n-\nnetworkd:\nunits:\n- name: 50-kubernikus.netdev\n@@ -293,7 +276,6 @@ networkd:\n[Network]\nDHCP=no\nAddress={{ .ApiserverIP }}/32\n-\nstorage:\nfiles:\n- path: /etc/udev/rules.d/99-vmware-scsi-udev.rules\n@@ -307,7 +289,6 @@ storage:\n# Modify the timeout value for VMware SCSI devices so that\n# in the event of a failover, we don't time out.\n# See Bug 271286 for more information.\n-\nACTION==\"add\", SUBSYSTEMS==\"scsi\", ATTRS{vendor}==\"VMware \", ATTRS{model}==\"Virtual disk\", RUN+=\"/bin/sh -c 'echo 180 >/sys$DEVPATH/timeout'\"\n- path: /etc/ssl/certs/SAPGlobalRootCA.pem\nfilesystem: root\n@@ -723,7 +704,6 @@ storage:\n${KUBELET_IMAGE} \\\n${KUBELET_IMAGE_ARGS} \\\n-- \"$@\"\n-\n- path: /opt/bin/flannel-wrapper\nfilesystem: root\nmode: 0755\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
Fix resolv.conf race condition (#622)
* Revert "Mount resolv.conf in rkt container (#608)"
This reverts commit e643bf149d2447b37b2bb0ff9cdfd532156ed6a6.
* Test for dns resolution before starting rkt containers
* Add nss-lookup.target to templates
* Compact node template, add unit test
|
596,226 |
08.10.2021 10:55:39
| -7,200 |
7e6d00cef59489bafb422e50ecc3d8785fdcf154
|
Compact ignition templates
|
[
{
"change_type": "MODIFY",
"old_path": "pkg/templates/node_1.10.go",
"new_path": "pkg/templates/node_1.10.go",
"diff": "@@ -684,11 +684,6 @@ storage:\ncontents:\ninline: |-\n#!/bin/bash\n- # Wrapper for launching kubelet via rkt-fly.\n- #\n- # Make sure to set KUBELET_IMAGE_TAG to an image tag published here:\n- # https://quay.io/repository/coreos/hyperkube?tab=tags Alternatively,\n- # override KUBELET_IMAGE to a custom image.\nset -e\nfunction require_ev_all() {\nfor rev in $@ ; do\n@@ -768,12 +763,6 @@ storage:\ncontents:\ninline: |-\n#!/bin/bash -e\n- # Wrapper for launching flannel via rkt.\n- #\n- # Make sure to set FLANNEL_IMAGE_TAG to an image tag published here:\n- # https://quay.io/repository/coreos/flannel?tab=tags Alternatively,\n- # override FLANNEL_IMAGE to a custom image.\n-\nfunction require_ev_all() {\nfor rev in $@ ; do\nif [[ -z \"${!rev}\" ]]; then\n@@ -782,7 +771,6 @@ storage:\nfi\ndone\n}\n-\nfunction require_ev_one() {\nfor rev in $@ ; do\nif [[ ! -z \"${!rev}\" ]]; then\n@@ -792,28 +780,21 @@ storage:\necho One of $@ must be set\nexit 1\n}\n-\nif [[ -n \"${FLANNEL_VER}\" ]]; then\necho FLANNEL_VER environment variable is deprecated, please use FLANNEL_IMAGE_TAG instead\nfi\n-\nif [[ -n \"${FLANNEL_IMG}\" ]]; then\necho FLANNEL_IMG environment variable is deprecated, please use FLANNEL_IMAGE_URL instead\nfi\n-\nFLANNEL_IMAGE_TAG=\"${FLANNEL_IMAGE_TAG:-${FLANNEL_VER}}\"\n-\nrequire_ev_one FLANNEL_IMAGE FLANNEL_IMAGE_TAG\n-\nFLANNEL_IMAGE_URL=\"${FLANNEL_IMAGE_URL:-${FLANNEL_IMG:-docker://quay.io/coreos/flannel}}\"\nFLANNEL_IMAGE=\"${FLANNEL_IMAGE:-${FLANNEL_IMAGE_URL}:${FLANNEL_IMAGE_TAG}}\"\n-\nif [[ \"${FLANNEL_IMAGE%%/*}\" == \"quay.io\" ]] && ! (echo \"${RKT_RUN_ARGS}\" | grep -q trust-keys-from-https); then\nRKT_RUN_ARGS=\"${RKT_RUN_ARGS} --trust-keys-from-https\"\nelif [[ \"${FLANNEL_IMAGE%%/*}\" == \"docker:\" ]] && ! (echo \"${RKT_RUN_ARGS}\" | grep -q insecure-options); then\nRKT_RUN_ARGS=\"${RKT_RUN_ARGS} --insecure-options=image\"\nfi\n-\nETCD_SSL_DIR=\"${ETCD_SSL_DIR:-/etc/ssl/etcd}\"\nif [[ -d \"${ETCD_SSL_DIR}\" ]]; then\nRKT_RUN_ARGS=\"${RKT_RUN_ARGS} \\\n@@ -821,7 +802,6 @@ storage:\n--mount volume=coreos-ssl,target=${ETCD_SSL_DIR} \\\n\"\nfi\n-\nif [[ -S \"${NOTIFY_SOCKET}\" ]]; then\nRKT_RUN_ARGS=\"${RKT_RUN_ARGS} \\\n--mount volume=coreos-notify,target=/run/systemd/notify \\\n@@ -829,9 +809,7 @@ storage:\n--set-env=NOTIFY_SOCKET=/run/systemd/notify \\\n\"\nfi\n-\nmkdir --parents /run/flannel\n-\nRKT=\"${RKT:-/opt/bin/rkt}\"\nRKT_STAGE1_ARG=\"${RKT_STAGE1_ARG:---stage1-from-dir=stage1-fly.aci}\"\nset -x\n"
},
{
"change_type": "MODIFY",
"old_path": "pkg/templates/node_1.11.go",
"new_path": "pkg/templates/node_1.11.go",
"diff": "@@ -636,11 +636,6 @@ storage:\ncontents:\ninline: |-\n#!/bin/bash\n- # Wrapper for launching kubelet via rkt-fly.\n- #\n- # Make sure to set KUBELET_IMAGE_TAG to an image tag published here:\n- # https://quay.io/repository/coreos/hyperkube?tab=tags Alternatively,\n- # override KUBELET_IMAGE to a custom image.\nset -e\nfunction require_ev_all() {\nfor rev in $@ ; do\n@@ -720,12 +715,6 @@ storage:\ncontents:\ninline: |-\n#!/bin/bash -e\n- # Wrapper for launching flannel via rkt.\n- #\n- # Make sure to set FLANNEL_IMAGE_TAG to an image tag published here:\n- # https://quay.io/repository/coreos/flannel?tab=tags Alternatively,\n- # override FLANNEL_IMAGE to a custom image.\n-\nfunction require_ev_all() {\nfor rev in $@ ; do\nif [[ -z \"${!rev}\" ]]; then\n@@ -734,7 +723,6 @@ storage:\nfi\ndone\n}\n-\nfunction require_ev_one() {\nfor rev in $@ ; do\nif [[ ! -z \"${!rev}\" ]]; then\n@@ -744,28 +732,21 @@ storage:\necho One of $@ must be set\nexit 1\n}\n-\nif [[ -n \"${FLANNEL_VER}\" ]]; then\necho FLANNEL_VER environment variable is deprecated, please use FLANNEL_IMAGE_TAG instead\nfi\n-\nif [[ -n \"${FLANNEL_IMG}\" ]]; then\necho FLANNEL_IMG environment variable is deprecated, please use FLANNEL_IMAGE_URL instead\nfi\n-\nFLANNEL_IMAGE_TAG=\"${FLANNEL_IMAGE_TAG:-${FLANNEL_VER}}\"\n-\nrequire_ev_one FLANNEL_IMAGE FLANNEL_IMAGE_TAG\n-\nFLANNEL_IMAGE_URL=\"${FLANNEL_IMAGE_URL:-${FLANNEL_IMG:-docker://quay.io/coreos/flannel}}\"\nFLANNEL_IMAGE=\"${FLANNEL_IMAGE:-${FLANNEL_IMAGE_URL}:${FLANNEL_IMAGE_TAG}}\"\n-\nif [[ \"${FLANNEL_IMAGE%%/*}\" == \"quay.io\" ]] && ! (echo \"${RKT_RUN_ARGS}\" | grep -q trust-keys-from-https); then\nRKT_RUN_ARGS=\"${RKT_RUN_ARGS} --trust-keys-from-https\"\nelif [[ \"${FLANNEL_IMAGE%%/*}\" == \"docker:\" ]] && ! (echo \"${RKT_RUN_ARGS}\" | grep -q insecure-options); then\nRKT_RUN_ARGS=\"${RKT_RUN_ARGS} --insecure-options=image\"\nfi\n-\nETCD_SSL_DIR=\"${ETCD_SSL_DIR:-/etc/ssl/etcd}\"\nif [[ -d \"${ETCD_SSL_DIR}\" ]]; then\nRKT_RUN_ARGS=\"${RKT_RUN_ARGS} \\\n@@ -773,7 +754,6 @@ storage:\n--mount volume=coreos-ssl,target=${ETCD_SSL_DIR} \\\n\"\nfi\n-\nif [[ -S \"${NOTIFY_SOCKET}\" ]]; then\nRKT_RUN_ARGS=\"${RKT_RUN_ARGS} \\\n--mount volume=coreos-notify,target=/run/systemd/notify \\\n@@ -781,9 +761,7 @@ storage:\n--set-env=NOTIFY_SOCKET=/run/systemd/notify \\\n\"\nfi\n-\nmkdir --parents /run/flannel\n-\nRKT=\"${RKT:-/opt/bin/rkt}\"\nRKT_STAGE1_ARG=\"${RKT_STAGE1_ARG:---stage1-from-dir=stage1-fly.aci}\"\nset -x\n"
},
{
"change_type": "MODIFY",
"old_path": "pkg/templates/node_1.12.go",
"new_path": "pkg/templates/node_1.12.go",
"diff": "@@ -638,11 +638,6 @@ storage:\ncontents:\ninline: |-\n#!/bin/bash\n- # Wrapper for launching kubelet via rkt-fly.\n- #\n- # Make sure to set KUBELET_IMAGE_TAG to an image tag published here:\n- # https://quay.io/repository/coreos/hyperkube?tab=tags Alternatively,\n- # override KUBELET_IMAGE to a custom image.\nset -e\nfunction require_ev_all() {\nfor rev in $@ ; do\n@@ -722,12 +717,6 @@ storage:\ncontents:\ninline: |-\n#!/bin/bash -e\n- # Wrapper for launching flannel via rkt.\n- #\n- # Make sure to set FLANNEL_IMAGE_TAG to an image tag published here:\n- # https://quay.io/repository/coreos/flannel?tab=tags Alternatively,\n- # override FLANNEL_IMAGE to a custom image.\n-\nfunction require_ev_all() {\nfor rev in $@ ; do\nif [[ -z \"${!rev}\" ]]; then\n@@ -736,7 +725,6 @@ storage:\nfi\ndone\n}\n-\nfunction require_ev_one() {\nfor rev in $@ ; do\nif [[ ! -z \"${!rev}\" ]]; then\n@@ -746,28 +734,21 @@ storage:\necho One of $@ must be set\nexit 1\n}\n-\nif [[ -n \"${FLANNEL_VER}\" ]]; then\necho FLANNEL_VER environment variable is deprecated, please use FLANNEL_IMAGE_TAG instead\nfi\n-\nif [[ -n \"${FLANNEL_IMG}\" ]]; then\necho FLANNEL_IMG environment variable is deprecated, please use FLANNEL_IMAGE_URL instead\nfi\n-\nFLANNEL_IMAGE_TAG=\"${FLANNEL_IMAGE_TAG:-${FLANNEL_VER}}\"\n-\nrequire_ev_one FLANNEL_IMAGE FLANNEL_IMAGE_TAG\n-\nFLANNEL_IMAGE_URL=\"${FLANNEL_IMAGE_URL:-${FLANNEL_IMG:-docker://quay.io/coreos/flannel}}\"\nFLANNEL_IMAGE=\"${FLANNEL_IMAGE:-${FLANNEL_IMAGE_URL}:${FLANNEL_IMAGE_TAG}}\"\n-\nif [[ \"${FLANNEL_IMAGE%%/*}\" == \"quay.io\" ]] && ! (echo \"${RKT_RUN_ARGS}\" | grep -q trust-keys-from-https); then\nRKT_RUN_ARGS=\"${RKT_RUN_ARGS} --trust-keys-from-https\"\nelif [[ \"${FLANNEL_IMAGE%%/*}\" == \"docker:\" ]] && ! (echo \"${RKT_RUN_ARGS}\" | grep -q insecure-options); then\nRKT_RUN_ARGS=\"${RKT_RUN_ARGS} --insecure-options=image\"\nfi\n-\nETCD_SSL_DIR=\"${ETCD_SSL_DIR:-/etc/ssl/etcd}\"\nif [[ -d \"${ETCD_SSL_DIR}\" ]]; then\nRKT_RUN_ARGS=\"${RKT_RUN_ARGS} \\\n@@ -775,7 +756,6 @@ storage:\n--mount volume=coreos-ssl,target=${ETCD_SSL_DIR} \\\n\"\nfi\n-\nif [[ -S \"${NOTIFY_SOCKET}\" ]]; then\nRKT_RUN_ARGS=\"${RKT_RUN_ARGS} \\\n--mount volume=coreos-notify,target=/run/systemd/notify \\\n@@ -783,9 +763,7 @@ storage:\n--set-env=NOTIFY_SOCKET=/run/systemd/notify \\\n\"\nfi\n-\nmkdir --parents /run/flannel\n-\nRKT=\"${RKT:-/opt/bin/rkt}\"\nRKT_STAGE1_ARG=\"${RKT_STAGE1_ARG:---stage1-from-dir=stage1-fly.aci}\"\nset -x\n"
},
{
"change_type": "MODIFY",
"old_path": "pkg/templates/node_1.14.go",
"new_path": "pkg/templates/node_1.14.go",
"diff": "@@ -640,11 +640,6 @@ storage:\ncontents:\ninline: |-\n#!/bin/bash\n- # Wrapper for launching kubelet via rkt-fly.\n- #\n- # Make sure to set KUBELET_IMAGE_TAG to an image tag published here:\n- # https://quay.io/repository/coreos/hyperkube?tab=tags Alternatively,\n- # override KUBELET_IMAGE to a custom image.\nset -e\nfunction require_ev_all() {\nfor rev in $@ ; do\n@@ -724,12 +719,6 @@ storage:\ncontents:\ninline: |-\n#!/bin/bash -e\n- # Wrapper for launching flannel via rkt.\n- #\n- # Make sure to set FLANNEL_IMAGE_TAG to an image tag published here:\n- # https://quay.io/repository/coreos/flannel?tab=tags Alternatively,\n- # override FLANNEL_IMAGE to a custom image.\n-\nfunction require_ev_all() {\nfor rev in $@ ; do\nif [[ -z \"${!rev}\" ]]; then\n@@ -738,7 +727,6 @@ storage:\nfi\ndone\n}\n-\nfunction require_ev_one() {\nfor rev in $@ ; do\nif [[ ! -z \"${!rev}\" ]]; then\n@@ -748,28 +736,21 @@ storage:\necho One of $@ must be set\nexit 1\n}\n-\nif [[ -n \"${FLANNEL_VER}\" ]]; then\necho FLANNEL_VER environment variable is deprecated, please use FLANNEL_IMAGE_TAG instead\nfi\n-\nif [[ -n \"${FLANNEL_IMG}\" ]]; then\necho FLANNEL_IMG environment variable is deprecated, please use FLANNEL_IMAGE_URL instead\nfi\n-\nFLANNEL_IMAGE_TAG=\"${FLANNEL_IMAGE_TAG:-${FLANNEL_VER}}\"\n-\nrequire_ev_one FLANNEL_IMAGE FLANNEL_IMAGE_TAG\n-\nFLANNEL_IMAGE_URL=\"${FLANNEL_IMAGE_URL:-${FLANNEL_IMG:-docker://quay.io/coreos/flannel}}\"\nFLANNEL_IMAGE=\"${FLANNEL_IMAGE:-${FLANNEL_IMAGE_URL}:${FLANNEL_IMAGE_TAG}}\"\n-\nif [[ \"${FLANNEL_IMAGE%%/*}\" == \"quay.io\" ]] && ! (echo \"${RKT_RUN_ARGS}\" | grep -q trust-keys-from-https); then\nRKT_RUN_ARGS=\"${RKT_RUN_ARGS} --trust-keys-from-https\"\nelif [[ \"${FLANNEL_IMAGE%%/*}\" == \"docker:\" ]] && ! (echo \"${RKT_RUN_ARGS}\" | grep -q insecure-options); then\nRKT_RUN_ARGS=\"${RKT_RUN_ARGS} --insecure-options=image\"\nfi\n-\nETCD_SSL_DIR=\"${ETCD_SSL_DIR:-/etc/ssl/etcd}\"\nif [[ -d \"${ETCD_SSL_DIR}\" ]]; then\nRKT_RUN_ARGS=\"${RKT_RUN_ARGS} \\\n@@ -777,7 +758,6 @@ storage:\n--mount volume=coreos-ssl,target=${ETCD_SSL_DIR} \\\n\"\nfi\n-\nif [[ -S \"${NOTIFY_SOCKET}\" ]]; then\nRKT_RUN_ARGS=\"${RKT_RUN_ARGS} \\\n--mount volume=coreos-notify,target=/run/systemd/notify \\\n@@ -785,9 +765,7 @@ storage:\n--set-env=NOTIFY_SOCKET=/run/systemd/notify \\\n\"\nfi\n-\nmkdir --parents /run/flannel\n-\nRKT=\"${RKT:-/opt/bin/rkt}\"\nRKT_STAGE1_ARG=\"${RKT_STAGE1_ARG:---stage1-from-dir=stage1-fly.aci}\"\nset -x\n"
},
{
"change_type": "MODIFY",
"old_path": "pkg/templates/node_1.17.go",
"new_path": "pkg/templates/node_1.17.go",
"diff": "@@ -640,11 +640,6 @@ storage:\ncontents:\ninline: |-\n#!/bin/bash\n- # Wrapper for launching kubelet via rkt-fly.\n- #\n- # Make sure to set KUBELET_IMAGE_TAG to an image tag published here:\n- # https://quay.io/repository/coreos/hyperkube?tab=tags Alternatively,\n- # override KUBELET_IMAGE to a custom image.\nset -e\nfunction require_ev_all() {\nfor rev in $@ ; do\n@@ -724,12 +719,6 @@ storage:\ncontents:\ninline: |-\n#!/bin/bash -e\n- # Wrapper for launching flannel via rkt.\n- #\n- # Make sure to set FLANNEL_IMAGE_TAG to an image tag published here:\n- # https://quay.io/repository/coreos/flannel?tab=tags Alternatively,\n- # override FLANNEL_IMAGE to a custom image.\n-\nfunction require_ev_all() {\nfor rev in $@ ; do\nif [[ -z \"${!rev}\" ]]; then\n@@ -738,7 +727,6 @@ storage:\nfi\ndone\n}\n-\nfunction require_ev_one() {\nfor rev in $@ ; do\nif [[ ! -z \"${!rev}\" ]]; then\n@@ -748,28 +736,21 @@ storage:\necho One of $@ must be set\nexit 1\n}\n-\nif [[ -n \"${FLANNEL_VER}\" ]]; then\necho FLANNEL_VER environment variable is deprecated, please use FLANNEL_IMAGE_TAG instead\nfi\n-\nif [[ -n \"${FLANNEL_IMG}\" ]]; then\necho FLANNEL_IMG environment variable is deprecated, please use FLANNEL_IMAGE_URL instead\nfi\n-\nFLANNEL_IMAGE_TAG=\"${FLANNEL_IMAGE_TAG:-${FLANNEL_VER}}\"\n-\nrequire_ev_one FLANNEL_IMAGE FLANNEL_IMAGE_TAG\n-\nFLANNEL_IMAGE_URL=\"${FLANNEL_IMAGE_URL:-${FLANNEL_IMG:-docker://quay.io/coreos/flannel}}\"\nFLANNEL_IMAGE=\"${FLANNEL_IMAGE:-${FLANNEL_IMAGE_URL}:${FLANNEL_IMAGE_TAG}}\"\n-\nif [[ \"${FLANNEL_IMAGE%%/*}\" == \"quay.io\" ]] && ! (echo \"${RKT_RUN_ARGS}\" | grep -q trust-keys-from-https); then\nRKT_RUN_ARGS=\"${RKT_RUN_ARGS} --trust-keys-from-https\"\nelif [[ \"${FLANNEL_IMAGE%%/*}\" == \"docker:\" ]] && ! (echo \"${RKT_RUN_ARGS}\" | grep -q insecure-options); then\nRKT_RUN_ARGS=\"${RKT_RUN_ARGS} --insecure-options=image\"\nfi\n-\nETCD_SSL_DIR=\"${ETCD_SSL_DIR:-/etc/ssl/etcd}\"\nif [[ -d \"${ETCD_SSL_DIR}\" ]]; then\nRKT_RUN_ARGS=\"${RKT_RUN_ARGS} \\\n@@ -777,7 +758,6 @@ storage:\n--mount volume=coreos-ssl,target=${ETCD_SSL_DIR} \\\n\"\nfi\n-\nif [[ -S \"${NOTIFY_SOCKET}\" ]]; then\nRKT_RUN_ARGS=\"${RKT_RUN_ARGS} \\\n--mount volume=coreos-notify,target=/run/systemd/notify \\\n@@ -785,9 +765,7 @@ storage:\n--set-env=NOTIFY_SOCKET=/run/systemd/notify \\\n\"\nfi\n-\nmkdir --parents /run/flannel\n-\nRKT=\"${RKT:-/opt/bin/rkt}\"\nRKT_STAGE1_ARG=\"${RKT_STAGE1_ARG:---stage1-from-dir=stage1-fly.aci}\"\nset -x\n"
},
{
"change_type": "MODIFY",
"old_path": "pkg/templates/node_1.19.go",
"new_path": "pkg/templates/node_1.19.go",
"diff": "@@ -637,11 +637,6 @@ storage:\ncontents:\ninline: |-\n#!/bin/bash\n- # Wrapper for launching kubelet via rkt-fly.\n- #\n- # Make sure to set KUBELET_IMAGE_TAG to an image tag published here:\n- # https://quay.io/repository/coreos/hyperkube?tab=tags Alternatively,\n- # override KUBELET_IMAGE to a custom image.\nset -e\nfunction require_ev_all() {\nfor rev in $@ ; do\n@@ -721,12 +716,6 @@ storage:\ncontents:\ninline: |-\n#!/bin/bash -e\n- # Wrapper for launching flannel via rkt.\n- #\n- # Make sure to set FLANNEL_IMAGE_TAG to an image tag published here:\n- # https://quay.io/repository/coreos/flannel?tab=tags Alternatively,\n- # override FLANNEL_IMAGE to a custom image.\n-\nfunction require_ev_all() {\nfor rev in $@ ; do\nif [[ -z \"${!rev}\" ]]; then\n@@ -735,7 +724,6 @@ storage:\nfi\ndone\n}\n-\nfunction require_ev_one() {\nfor rev in $@ ; do\nif [[ ! -z \"${!rev}\" ]]; then\n@@ -745,28 +733,21 @@ storage:\necho One of $@ must be set\nexit 1\n}\n-\nif [[ -n \"${FLANNEL_VER}\" ]]; then\necho FLANNEL_VER environment variable is deprecated, please use FLANNEL_IMAGE_TAG instead\nfi\n-\nif [[ -n \"${FLANNEL_IMG}\" ]]; then\necho FLANNEL_IMG environment variable is deprecated, please use FLANNEL_IMAGE_URL instead\nfi\n-\nFLANNEL_IMAGE_TAG=\"${FLANNEL_IMAGE_TAG:-${FLANNEL_VER}}\"\n-\nrequire_ev_one FLANNEL_IMAGE FLANNEL_IMAGE_TAG\n-\nFLANNEL_IMAGE_URL=\"${FLANNEL_IMAGE_URL:-${FLANNEL_IMG:-docker://quay.io/coreos/flannel}}\"\nFLANNEL_IMAGE=\"${FLANNEL_IMAGE:-${FLANNEL_IMAGE_URL}:${FLANNEL_IMAGE_TAG}}\"\n-\nif [[ \"${FLANNEL_IMAGE%%/*}\" == \"quay.io\" ]] && ! (echo \"${RKT_RUN_ARGS}\" | grep -q trust-keys-from-https); then\nRKT_RUN_ARGS=\"${RKT_RUN_ARGS} --trust-keys-from-https\"\nelif [[ \"${FLANNEL_IMAGE%%/*}\" == \"docker:\" ]] && ! (echo \"${RKT_RUN_ARGS}\" | grep -q insecure-options); then\nRKT_RUN_ARGS=\"${RKT_RUN_ARGS} --insecure-options=image\"\nfi\n-\nETCD_SSL_DIR=\"${ETCD_SSL_DIR:-/etc/ssl/etcd}\"\nif [[ -d \"${ETCD_SSL_DIR}\" ]]; then\nRKT_RUN_ARGS=\"${RKT_RUN_ARGS} \\\n@@ -774,7 +755,6 @@ storage:\n--mount volume=coreos-ssl,target=${ETCD_SSL_DIR} \\\n\"\nfi\n-\nif [[ -S \"${NOTIFY_SOCKET}\" ]]; then\nRKT_RUN_ARGS=\"${RKT_RUN_ARGS} \\\n--mount volume=coreos-notify,target=/run/systemd/notify \\\n@@ -782,9 +762,7 @@ storage:\n--set-env=NOTIFY_SOCKET=/run/systemd/notify \\\n\"\nfi\n-\nmkdir --parents /run/flannel\n-\nRKT=\"${RKT:-/opt/bin/rkt}\"\nRKT_STAGE1_ARG=\"${RKT_STAGE1_ARG:---stage1-from-dir=stage1-fly.aci}\"\nset -x\n"
},
{
"change_type": "MODIFY",
"old_path": "pkg/templates/node_1.20.go",
"new_path": "pkg/templates/node_1.20.go",
"diff": "@@ -641,11 +641,6 @@ storage:\ncontents:\ninline: |-\n#!/bin/bash\n- # Wrapper for launching kubelet via rkt-fly.\n- #\n- # Make sure to set KUBELET_IMAGE_TAG to an image tag published here:\n- # https://quay.io/repository/coreos/hyperkube?tab=tags Alternatively,\n- # override KUBELET_IMAGE to a custom image.\nset -e\nfunction require_ev_all() {\nfor rev in $@ ; do\n@@ -725,12 +720,6 @@ storage:\ncontents:\ninline: |-\n#!/bin/bash -e\n- # Wrapper for launching flannel via rkt.\n- #\n- # Make sure to set FLANNEL_IMAGE_TAG to an image tag published here:\n- # https://quay.io/repository/coreos/flannel?tab=tags Alternatively,\n- # override FLANNEL_IMAGE to a custom image.\n-\nfunction require_ev_all() {\nfor rev in $@ ; do\nif [[ -z \"${!rev}\" ]]; then\n@@ -739,7 +728,6 @@ storage:\nfi\ndone\n}\n-\nfunction require_ev_one() {\nfor rev in $@ ; do\nif [[ ! -z \"${!rev}\" ]]; then\n@@ -749,28 +737,21 @@ storage:\necho One of $@ must be set\nexit 1\n}\n-\nif [[ -n \"${FLANNEL_VER}\" ]]; then\necho FLANNEL_VER environment variable is deprecated, please use FLANNEL_IMAGE_TAG instead\nfi\n-\nif [[ -n \"${FLANNEL_IMG}\" ]]; then\necho FLANNEL_IMG environment variable is deprecated, please use FLANNEL_IMAGE_URL instead\nfi\n-\nFLANNEL_IMAGE_TAG=\"${FLANNEL_IMAGE_TAG:-${FLANNEL_VER}}\"\n-\nrequire_ev_one FLANNEL_IMAGE FLANNEL_IMAGE_TAG\n-\nFLANNEL_IMAGE_URL=\"${FLANNEL_IMAGE_URL:-${FLANNEL_IMG:-docker://quay.io/coreos/flannel}}\"\nFLANNEL_IMAGE=\"${FLANNEL_IMAGE:-${FLANNEL_IMAGE_URL}:${FLANNEL_IMAGE_TAG}}\"\n-\nif [[ \"${FLANNEL_IMAGE%%/*}\" == \"quay.io\" ]] && ! (echo \"${RKT_RUN_ARGS}\" | grep -q trust-keys-from-https); then\nRKT_RUN_ARGS=\"${RKT_RUN_ARGS} --trust-keys-from-https\"\nelif [[ \"${FLANNEL_IMAGE%%/*}\" == \"docker:\" ]] && ! (echo \"${RKT_RUN_ARGS}\" | grep -q insecure-options); then\nRKT_RUN_ARGS=\"${RKT_RUN_ARGS} --insecure-options=image\"\nfi\n-\nETCD_SSL_DIR=\"${ETCD_SSL_DIR:-/etc/ssl/etcd}\"\nif [[ -d \"${ETCD_SSL_DIR}\" ]]; then\nRKT_RUN_ARGS=\"${RKT_RUN_ARGS} \\\n@@ -778,7 +759,6 @@ storage:\n--mount volume=coreos-ssl,target=${ETCD_SSL_DIR} \\\n\"\nfi\n-\nif [[ -S \"${NOTIFY_SOCKET}\" ]]; then\nRKT_RUN_ARGS=\"${RKT_RUN_ARGS} \\\n--mount volume=coreos-notify,target=/run/systemd/notify \\\n@@ -786,9 +766,7 @@ storage:\n--set-env=NOTIFY_SOCKET=/run/systemd/notify \\\n\"\nfi\n-\nmkdir --parents /run/flannel\n-\nRKT=\"${RKT:-/opt/bin/rkt}\"\nRKT_STAGE1_ARG=\"${RKT_STAGE1_ARG:---stage1-from-dir=stage1-fly.aci}\"\nset -x\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
Compact ignition templates
|
596,226 |
08.10.2021 11:13:54
| -7,200 |
7229d0a6bf8eb481cf4b8e10204b32ab93706a52
|
Set default version to 1.20.8
|
[
{
"change_type": "MODIFY",
"old_path": "charts/images.yaml",
"new_path": "charts/images.yaml",
"diff": "@@ -71,6 +71,7 @@ imagesForVersion:\nrepository: 'keppel.$REGION.cloud.sap/ccloud-quay-mirror/coreos/flannel'\ntag: 'v0.12.0'\n'1.20.8':\n+ default: true\nsupported: true\napiserver:\nrepository: 'keppel.$REGION.cloud.sap/ccloud/kube-apiserver'\n@@ -446,7 +447,6 @@ imagesForVersion:\nrepository: 'keppel.$REGION.cloud.sap/ccloud-quay-mirror/coreos/flannel'\ntag: 'v0.12.0'\n'1.15.9':\n- default: true\nsupported: true\nhyperkube:\nrepository: 'keppel.$REGION.cloud.sap/ccloud-dockerhub-mirror/sapcc/hyperkube'\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
Set default version to 1.20.8
|
596,226 |
12.10.2021 15:06:51
| -7,200 |
31e85c6e264d4a9bfc093e36255bdb4b9f72739a
|
Make cidr and image configurable in e2e test
|
[
{
"change_type": "MODIFY",
"old_path": "test/e2e/main_test.go",
"new_path": "test/e2e/main_test.go",
"diff": "@@ -107,6 +107,15 @@ func TestRunner(t *testing.T) {\nif os.Getenv(\"CP_KLUSTER\") != \"\" {\nfmt.Printf(\"CP Kluster Name: %v\\n\", os.Getenv(\"CP_KLUSTER\"))\n}\n+ if os.Getenv(\"KLUSTER_VERSION\") != \"\" {\n+ fmt.Printf(\"Kubernetes Version: %v\\n\", os.Getenv(\"KLUSTER_VERSION\"))\n+ }\n+ if os.Getenv(\"KLUSTER_CIDR\") != \"\" {\n+ fmt.Printf(\"Cluster CIDR: %v\\n\", os.Getenv(\"KLUSTER_CIDR\"))\n+ }\n+ if os.Getenv(\"KLUSTER_OS_IMAGE\") != \"\" {\n+ fmt.Printf(\"OS Image: %v\\n\", os.Getenv(\"KLUSTER_OS_IMAGE\"))\n+ }\nfmt.Printf(\"\\n\\n\")\nauthOptions := &tokens.AuthOptions{\n"
},
{
"change_type": "MODIFY",
"old_path": "test/e2e/setup_test.go",
"new_path": "test/e2e/setup_test.go",
"diff": "@@ -42,17 +42,30 @@ func (s *SetupTests) CreateCluster(t *testing.T) {\nif v := os.Getenv(\"KLUSTER_VERSION\"); v != \"\" {\nversion = v\n}\n+\n+ clusterCidr := \"100.100.0.0/16\"\n+ if cidr := os.Getenv(\"KLUSTER_CIDR\"); cidr != \"\" {\n+ clusterCidr = cidr\n+ }\n+\n+ osImage := \"flatcar-stable-amd64\"\n+ if image := os.Getenv(\"KLUSTER_OS_IMAGE\"); image != \"\" {\n+ osImage = image\n+ }\n+\nkluster := &models.Kluster{\nName: s.KlusterName,\nSpec: models.KlusterSpec{\nVersion: version,\nSSHPublicKey: \"ssh-rsa AAAAB3NzaC1yc2EAAAADAQABAAACAQCXIxVEUgtUVkvk2VM1hmIb8MxvxsmvYoiq9OBy3J8akTGNybqKsA2uhcwxSJX5Cn3si8kfMfka9EWiJT+e1ybvtsGILO5XRZPxyhYzexwb3TcALwc3LuzpF3Z/Dg2jYTRELTGhYmyca3mxzTlCjNXvYayLNedjJ8fIBzoCuSXNqDRToHru7h0Glz+wtuE74mNkOiXSvhtuJtJs7VCNVjobFQNfC1aeDsri2bPRHJJZJ0QF4LLYSayMEz3lVwIDyAviQR2Aa97WfuXiofiAemfGqiH47Kq6b8X7j3bOYGBvJKMUV7XeWhGsskAmTsvvnFxkc5PAD3Ct+liULjiQWlzDrmpTE8aMqLK4l0YQw7/8iRVz6gli42iEc2ZG56ob1ErpTLAKFWyCNOebZuGoygdEQaGTIIunAncXg5Rz07TdPl0Tf5ZZLpiAgR5ck0H1SETnjDTZ/S83CiVZWJgmCpu8YOKWyYRD4orWwdnA77L4+ixeojLIhEoNL8KlBgsP9Twx+fFMWLfxMmiuX+yksM6Hu+Lsm+Ao7Q284VPp36EB1rxP1JM7HCiEOEm50Jb6hNKjgN4aoLhG5yg+GnDhwCZqUwcRJo1bWtm3QvRA+rzrGZkId4EY3cyOK5QnYV5+24x93Ex0UspHMn7HGsHUESsVeV0fLqlfXyd2RbHTmDMP6w== Kubernikus Master Key\",\n+ ClusterCIDR: &clusterCidr,\nNodePools: []models.NodePool{\n{\nName: \"small\",\nFlavor: \"m1.small\",\nSize: SmokeTestNodeCount,\nAvailabilityZone: os.Getenv(\"NODEPOOL_AVZ\"),\n+ Image: osImage,\n},\n},\n},\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
Make cidr and image configurable in e2e test
|
596,226 |
12.10.2021 17:14:01
| -7,200 |
58d21884717e0201a6bdc96c626efaf607ff7c5c
|
Isolate e2e version tests
|
[
{
"change_type": "MODIFY",
"old_path": "test/e2e/cleanup_test.go",
"new_path": "test/e2e/cleanup_test.go",
"diff": "@@ -32,6 +32,7 @@ type CleanupTests struct {\nOpenStack *framework.OpenStack\nKlusterName string\nReuse bool\n+ Isolate bool\n}\nfunc (s *CleanupTests) Run(t *testing.T) {\n@@ -40,7 +41,9 @@ func (s *CleanupTests) Run(t *testing.T) {\nt.Run(\"Cluster/IsDeleted\", s.WaitForKlusterToBeDeleted)\nif s.Reuse == false {\n+ if s.Isolate == false {\nt.Run(\"QuotaPostFlightCheck\", s.QuotaPostFlightCheck)\n+ }\nt.Run(\"ServerGroupsGotDeleted\", s.ServerGroupsGotDeleted)\nt.Run(\"LoadbalancerGotDeleted\", s.LoadbalancerGotDeleted)\n}\n"
},
{
"change_type": "MODIFY",
"old_path": "test/e2e/main_test.go",
"new_path": "test/e2e/main_test.go",
"diff": "@@ -22,6 +22,7 @@ var (\nkluster = flag.String(\"kluster\", \"\", \"Use existing Kluster\")\nreuse = flag.Bool(\"reuse\", false, \"Reuse exisiting Kluster\")\ncleanup = flag.Bool(\"cleanup\", true, \"Cleanup after tests have been run\")\n+ isolate = flag.Bool(\"isolate\", false, \"Do not destroy or depend on resources of other tests running in the same project\")\n)\nconst (\n@@ -74,6 +75,10 @@ func TestRunner(t *testing.T) {\nrequire.NoError(t, err, \"Must be able to parse Kubernikus URL\")\nrequire.NotEmpty(t, kurl.Host, \"There must be a host in the Kubernikus URL\")\n+ if os.Getenv(\"ISOLATE_TEST\") == \"true\" {\n+ *isolate = true\n+ }\n+\nfmt.Printf(\"========================================================================\\n\")\nfmt.Printf(\"Authentication\\n\")\nfmt.Printf(\"========================================================================\\n\")\n@@ -102,6 +107,7 @@ func TestRunner(t *testing.T) {\nfmt.Printf(\"Kluster Name: %v\\n\", klusterName)\nfmt.Printf(\"Reuse: %v\\n\", *reuse)\nfmt.Printf(\"Cleanup: %v\\n\", *cleanup)\n+ fmt.Printf(\"Isolate: %v\\n\", *isolate)\nfmt.Println(\"\")\nfmt.Printf(\"Dashboard: https://dashboard.%s.cloud.sap/%s/%s/kubernetes\\n\", os.Getenv(\"OS_REGION_NAME\"), os.Getenv(\"OS_PROJECT_DOMAIN_NAME\"), os.Getenv(\"OS_PROJECT_NAME\"))\nif os.Getenv(\"CP_KLUSTER\") != \"\" {\n@@ -171,7 +177,7 @@ func TestRunner(t *testing.T) {\nfullKlusterName := fmt.Sprintf(\"%s-%s\", klusterName, project.ID)\n// Pyrolize garbage left from previous e2e runs\n- pyrolisisTests := &PyrolisisTests{kubernikus, openstack, *reuse}\n+ pyrolisisTests := &PyrolisisTests{kubernikus, openstack, *reuse, *isolate}\nif !t.Run(\"Pyrolisis\", pyrolisisTests.Run) {\nreturn\n}\n@@ -182,7 +188,7 @@ func TestRunner(t *testing.T) {\n}\nif cleanup != nil && *cleanup == true {\n- cleanupTests := &CleanupTests{kubernikus, openstack, klusterName, *reuse}\n+ cleanupTests := &CleanupTests{kubernikus, openstack, klusterName, *reuse, *isolate}\ndefer t.Run(\"Cleanup\", cleanupTests.Run)\n}\n"
},
{
"change_type": "MODIFY",
"old_path": "test/e2e/pyrolisis_test.go",
"new_path": "test/e2e/pyrolisis_test.go",
"diff": "@@ -32,10 +32,11 @@ type PyrolisisTests struct {\nKubernikus *framework.Kubernikus\nOpenStack *framework.OpenStack\nReuse bool\n+ Isolate bool\n}\nfunc (p *PyrolisisTests) Run(t *testing.T) {\n- if p.Reuse == false {\n+ if p.Reuse == false && p.Isolate == false {\nquota := t.Run(\"SettingKlustersOnFire\", p.SettingKlustersOnFire)\nrequire.True(t, quota, \"Klusters must burn\")\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
Isolate e2e version tests
|
596,226 |
12.10.2021 17:41:31
| -7,200 |
85be290900a918eed3582db416b6ef69a6ff885d
|
Skip cleanup tests
|
[
{
"change_type": "MODIFY",
"old_path": "test/e2e/cleanup_test.go",
"new_path": "test/e2e/cleanup_test.go",
"diff": "@@ -40,10 +40,8 @@ func (s *CleanupTests) Run(t *testing.T) {\nt.Run(\"Cluster/BecomesTerminating\", s.KlusterPhaseBecomesTerminating)\nt.Run(\"Cluster/IsDeleted\", s.WaitForKlusterToBeDeleted)\n- if s.Reuse == false {\n- if s.Isolate == false {\n+ if s.Reuse == false && s.Isolate == false {\nt.Run(\"QuotaPostFlightCheck\", s.QuotaPostFlightCheck)\n- }\nt.Run(\"ServerGroupsGotDeleted\", s.ServerGroupsGotDeleted)\nt.Run(\"LoadbalancerGotDeleted\", s.LoadbalancerGotDeleted)\n}\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
Skip cleanup tests
|
596,226 |
13.10.2021 10:42:49
| -7,200 |
51f1bcbd231c61c05c1e1327fa86c1857d535330
|
Test all versions in k-master
|
[
{
"change_type": "MODIFY",
"old_path": "ci/pipeline.yaml",
"new_path": "ci/pipeline.yaml",
"diff": "@@ -743,6 +743,10 @@ task_e2e_tests: &task_e2e_tests\nCP_KUBERNIKUS_URL:\nCP_KLUSTER:\nNODEPOOL_AVZ:\n+ KLUSTER_VERSION:\n+ KLUSTER_CIDR:\n+ KLUSTER_OS_IMAGE:\n+ ISOLATE_TEST:\ntask_github_compare_url: &task_github_compare_url\n@@ -1100,7 +1104,6 @@ slack: &slack\nresources:\n-\n- name: docs.git\ntype: git\nsource:\n@@ -1221,12 +1224,19 @@ resources:\ntype: time\nsource: {interval: 1h}\n+ - name: all-versions.trigger\n+ type: time-version-resource\n+ check_every: 525600h\n+ source:\n+ pipeline: kubernikus\n+\nresource_types:\n- name: slack-alert\ntype: docker-image\ncheck_every: 24h\nsource:\nrepository: keppel.eu-de-1.cloud.sap/ccloud-dockerhub-mirror/arbourd/concourse-slack-alert-resource\n+\n- name: slack-notification\ntype: docker-image\ncheck_every: 24h\n@@ -1359,7 +1369,151 @@ jobs:\nPRODUCT_TOKEN: ((whitesource-product-token))\n<<: *slack\n- - name: e2e\n+ - name: all-versions\n+ serial: true\n+ plan:\n+ - put: all-versions.trigger\n+\n+\n+ - name: e2e-1.21.5\n+ serial: true\n+ plan:\n+ - in_parallel:\n+ - get: secrets.git\n+ - get: kubernikus.builds\n+ resource: kubernikus.git\n+ passed: [master]\n+ trigger: true\n+ - get: all-versions.trigger\n+ trigger: true\n+ passed: [all-versions]\n+ - task: e2e_tests\n+ config:\n+ <<: *task_e2e_tests\n+ timeout: 2h\n+ params:\n+ KLUSTER_VERSION: \"1.21.5\"\n+ KLUSTER_CIDR: \"100.100.0.0/16\"\n+ ISOLATE_TEST: \"true\"\n+ <<: *auth_e2e_master\n+ <<: *slack\n+\n+ - name: e2e-1.20.8\n+ serial: true\n+ plan:\n+ - in_parallel:\n+ - get: secrets.git\n+ - get: kubernikus.builds\n+ resource: kubernikus.git\n+ passed: [master]\n+ trigger: true\n+ - get: all-versions.trigger\n+ trigger: true\n+ passed: [all-versions]\n+ - task: e2e_tests\n+ config:\n+ <<: *task_e2e_tests\n+ timeout: 2h\n+ params:\n+ KLUSTER_VERSION: \"1.20.8\"\n+ KLUSTER_CIDR: \"100.101.0.0/16\"\n+ ISOLATE_TEST: \"true\"\n+ <<: *auth_e2e_master\n+ <<: *slack\n+\n+ - name: e2e-1.19.11\n+ serial: true\n+ plan:\n+ - in_parallel:\n+ - get: secrets.git\n+ - get: kubernikus.builds\n+ resource: kubernikus.git\n+ passed: [master]\n+ trigger: true\n+ - get: all-versions.trigger\n+ trigger: true\n+ passed: [all-versions]\n+ - task: e2e_tests\n+ config:\n+ <<: *task_e2e_tests\n+ timeout: 2h\n+ params:\n+ KLUSTER_VERSION: \"1.19.11\"\n+ KLUSTER_CIDR: \"100.102.0.0/16\"\n+ ISOLATE_TEST: \"true\"\n+ <<: *auth_e2e_master\n+ <<: *slack\n+\n+ - name: e2e-1.18.10\n+ serial: true\n+ plan:\n+ - in_parallel:\n+ - get: secrets.git\n+ - get: kubernikus.builds\n+ resource: kubernikus.git\n+ passed: [master]\n+ trigger: true\n+ - get: all-versions.trigger\n+ trigger: true\n+ passed: [all-versions]\n+ - task: e2e_tests\n+ config:\n+ <<: *task_e2e_tests\n+ timeout: 2h\n+ params:\n+ KLUSTER_VERSION: \"1.18.10\"\n+ KLUSTER_CIDR: \"100.103.0.0/16\"\n+ ISOLATE_TEST: \"true\"\n+ <<: *auth_e2e_master\n+ <<: *slack\n+\n+ - name: e2e-1.17.13\n+ serial: true\n+ plan:\n+ - in_parallel:\n+ - get: secrets.git\n+ - get: kubernikus.builds\n+ resource: kubernikus.git\n+ passed: [master]\n+ trigger: true\n+ - get: all-versions.trigger\n+ trigger: true\n+ passed: [all-versions]\n+ - task: e2e_tests\n+ config:\n+ <<: *task_e2e_tests\n+ timeout: 2h\n+ params:\n+ KLUSTER_VERSION: \"1.17.13\"\n+ KLUSTER_CIDR: \"100.104.0.0/16\"\n+ ISOLATE_TEST: \"true\"\n+ <<: *auth_e2e_master\n+ <<: *slack\n+\n+ - name: e2e-1.16.14\n+ serial: true\n+ plan:\n+ - in_parallel:\n+ - get: secrets.git\n+ - get: kubernikus.builds\n+ resource: kubernikus.git\n+ passed: [master]\n+ trigger: true\n+ - get: all-versions.trigger\n+ trigger: true\n+ passed: [all-versions]\n+ - task: e2e_tests\n+ config:\n+ <<: *task_e2e_tests\n+ timeout: 2h\n+ params:\n+ KLUSTER_VERSION: \"1.16.14\"\n+ KLUSTER_CIDR: \"100.105.0.0/16\"\n+ ISOLATE_TEST: \"true\"\n+ <<: *auth_e2e_master\n+ <<: *slack\n+\n+ - name: e2e-1.15.9\nserial: true\nplan:\n- in_parallel:\n@@ -1368,20 +1522,96 @@ jobs:\nresource: kubernikus.git\npassed: [master]\ntrigger: true\n+ - get: all-versions.trigger\n+ trigger: true\n+ passed: [all-versions]\n- task: e2e_tests\nconfig:\n<<: *task_e2e_tests\ntimeout: 2h\nparams:\n+ KLUSTER_VERSION: \"1.15.9\"\n+ KLUSTER_CIDR: \"100.106.0.0/16\"\n+ ISOLATE_TEST: \"true\"\n<<: *auth_e2e_master\n<<: *slack\n+ - name: e2e-1.14.5\n+ serial: true\n+ plan:\n+ - in_parallel:\n+ - get: secrets.git\n+ - get: kubernikus.builds\n+ resource: kubernikus.git\n+ passed: [master]\n+ trigger: true\n+ - get: all-versions.trigger\n+ trigger: true\n+ passed: [all-versions]\n+ - task: e2e_tests\n+ config:\n+ <<: *task_e2e_tests\n+ timeout: 2h\n+ params:\n+ KLUSTER_VERSION: \"1.14.5\"\n+ KLUSTER_CIDR: \"100.107.0.0/16\"\n+ ISOLATE_TEST: \"true\"\n+ <<: *auth_e2e_master\n+ <<: *slack\n+\n+ - name: e2e-1.13.9\n+ serial: true\n+ plan:\n+ - in_parallel:\n+ - get: secrets.git\n+ - get: kubernikus.builds\n+ resource: kubernikus.git\n+ passed: [master]\n+ trigger: true\n+ - get: all-versions.trigger\n+ trigger: true\n+ passed: [all-versions]\n+ - task: e2e_tests\n+ config:\n+ <<: *task_e2e_tests\n+ timeout: 2h\n+ params:\n+ KLUSTER_VERSION: \"1.13.9\"\n+ KLUSTER_CIDR: \"100.108.0.0/16\"\n+ ISOLATE_TEST: \"true\"\n+ <<: *auth_e2e_master\n+ <<: *slack\n+\n+ - name: e2e-1.12.10\n+ serial: true\n+ plan:\n+ - in_parallel:\n+ - get: secrets.git\n+ - get: kubernikus.builds\n+ resource: kubernikus.git\n+ passed: [master]\n+ trigger: true\n+ - get: all-versions.trigger\n+ trigger: true\n+ passed: [all-versions]\n+ - task: e2e_tests\n+ config:\n+ <<: *task_e2e_tests\n+ timeout: 2h\n+ params:\n+ KLUSTER_VERSION: \"1.12.10\"\n+ KLUSTER_CIDR: \"100.109.0.0/16\"\n+ ISOLATE_TEST: \"true\"\n+ <<: *auth_e2e_master\n+ <<: *slack\n+\n+\n- name: prod\nserial: true\nplan:\n- get: kubernikus.builds\nresource: kubernikus.git\n- passed: [e2e]\n+ passed: [e2e-1.21.5]\n- task: changes\nconfig:\n<<: *task_github_compare_url\n@@ -3313,7 +3543,28 @@ groups:\n- cli\n- whitesource\n- tiller_master\n- - e2e\n+ - all-versions\n+\n+ - e2e-1.21.5\n+\n+ - e2e-1.20.8\n+\n+ - e2e-1.19.11\n+\n+ - e2e-1.18.10\n+\n+ - e2e-1.17.13\n+\n+ - e2e-1.16.14\n+\n+ - e2e-1.15.9\n+\n+ - e2e-1.14.5\n+\n+ - e2e-1.13.9\n+\n+ - e2e-1.12.10\n+\n- prod\n- ASIA\n@@ -3469,7 +3720,28 @@ groups:\n- cli\n- whitesource\n- tiller_master\n- - e2e\n+ - all-versions\n+\n+ - e2e-1.21.5\n+\n+ - e2e-1.20.8\n+\n+ - e2e-1.19.11\n+\n+ - e2e-1.18.10\n+\n+ - e2e-1.17.13\n+\n+ - e2e-1.16.14\n+\n+ - e2e-1.15.9\n+\n+ - e2e-1.14.5\n+\n+ - e2e-1.13.9\n+\n+ - e2e-1.12.10\n+\n- prod\n- ASIA\n"
},
{
"change_type": "MODIFY",
"old_path": "ci/pipeline.yaml.erb",
"new_path": "ci/pipeline.yaml.erb",
"diff": "<%\n+VERSIONS = [\n+ '1.21.5',\n+ '1.20.8',\n+ '1.19.11',\n+ '1.18.10',\n+ '1.17.13',\n+ '1.16.14',\n+ '1.15.9',\n+ '1.14.5',\n+ '1.13.9',\n+ '1.12.10'\n+]\n+\nREGIONS = {\n\"ap-ae-1\": { continent: 'ASIA', install: 'terraform', e2e: 'terraform', oidc: '9a2d0e1181194fb1bca52e7951666bc3'},\n\"ap-au-1\": { continent: 'ASIA', install: 'manual' , e2e: 'manual', oidc: 'f34304c03d2147cc964470c31d00ce97' },\n@@ -91,7 +104,6 @@ slack: &slack\nresources:\n-\n- name: docs.git\ntype: git\nsource:\n@@ -191,12 +203,19 @@ resources:\ntype: time\nsource: {interval: 1h}\n+ - name: all-versions.trigger\n+ type: time-version-resource\n+ check_every: 525600h\n+ source:\n+ pipeline: kubernikus\n+\nresource_types:\n- name: slack-alert\ntype: docker-image\ncheck_every: 24h\nsource:\nrepository: keppel.eu-de-1.cloud.sap/ccloud-dockerhub-mirror/arbourd/concourse-slack-alert-resource\n+\n- name: slack-notification\ntype: docker-image\ncheck_every: 24h\n@@ -329,7 +348,13 @@ jobs:\nPRODUCT_TOKEN: ((whitesource-product-token))\n<<: *slack\n- - name: e2e\n+ - name: all-versions\n+ serial: true\n+ plan:\n+ - put: all-versions.trigger\n+\n+<% VERSIONS.each_with_index do |version, index| %>\n+ - name: e2e-<%= version %>\nserial: true\nplan:\n- in_parallel:\n@@ -338,20 +363,27 @@ jobs:\nresource: kubernikus.git\npassed: [master]\ntrigger: true\n+ - get: all-versions.trigger\n+ trigger: true\n+ passed: [all-versions]\n- task: e2e_tests\nconfig:\n<<: *task_e2e_tests\ntimeout: 2h\nparams:\n+ KLUSTER_VERSION: \"<%= version %>\"\n+ KLUSTER_CIDR: \"100.<%= 100+index %>.0.0/16\"\n+ ISOLATE_TEST: \"true\"\n<<: *auth_e2e_master\n<<: *slack\n+<% end %>\n- name: prod\nserial: true\nplan:\n- get: kubernikus.builds\nresource: kubernikus.git\n- passed: [e2e]\n+ passed: [e2e-<%= VERSIONS[0] %>]\n- task: changes\nconfig:\n<<: *task_github_compare_url\n@@ -492,7 +524,10 @@ groups:\n- cli\n- whitesource\n- tiller_master\n- - e2e\n+ - all-versions\n+<% VERSIONS.each do |version| %>\n+ - e2e-<%= version %>\n+<% end %>\n- prod\n<% GROUPS.each do |group| %>\n- <%= group %>\n@@ -514,7 +549,10 @@ groups:\n- cli\n- whitesource\n- tiller_master\n- - e2e\n+ - all-versions\n+<% VERSIONS.each do |version| %>\n+ - e2e-<%= version %>\n+<% end %>\n- prod\n<% GROUPS.each do |group| %>\n- <%= group %>\n"
},
{
"change_type": "MODIFY",
"old_path": "ci/task_e2e_tests.yaml",
"new_path": "ci/task_e2e_tests.yaml",
"diff": "@@ -54,3 +54,7 @@ params:\nCP_KUBERNIKUS_URL:\nCP_KLUSTER:\nNODEPOOL_AVZ:\n+ KLUSTER_VERSION:\n+ KLUSTER_CIDR:\n+ KLUSTER_OS_IMAGE:\n+ ISOLATE_TEST:\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
Test all versions in k-master
|
596,226 |
14.10.2021 11:54:04
| -7,200 |
007fa674f4585b2a5ae829c84231c225f9fbbbaf
|
Only test flatcar stable version
|
[
{
"change_type": "MODIFY",
"old_path": "test/e2e/node_test.go",
"new_path": "test/e2e/node_test.go",
"diff": "@@ -6,6 +6,7 @@ import (\n\"fmt\"\n\"io/ioutil\"\n\"net/http\"\n+ \"os\"\n\"strings\"\n\"testing\"\n\"time\"\n@@ -133,6 +134,10 @@ func (k *NodeTests) Registered(t *testing.T) {\n}\nfunc (k NodeTests) LatestStableContainerLinux(t *testing.T) {\n+ if os.Getenv(\"KLUSTER_OS_IMAGE\") != \"\" && os.Getenv(\"KLUSTER_OS_IMAGE\") != \"flatcar-stable-amd64\" {\n+ return\n+ }\n+\nnodes, err := k.Kubernetes.ClientSet.CoreV1().Nodes().List(meta_v1.ListOptions{})\nif !assert.NoError(t, err) {\nreturn\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
Only test flatcar stable version
|
596,226 |
26.10.2021 17:37:20
| -7,200 |
25376dd113f489205fd81a0b93cdd1db8b2af7ac
|
[ci skip] Add coredns scripts
|
[
{
"change_type": "ADD",
"old_path": null,
"new_path": "scripts/coredns/clusterrole.yaml",
"diff": "+apiVersion: rbac.authorization.k8s.io/v1\n+kind: ClusterRole\n+metadata:\n+ labels:\n+ addonmanager.kubernetes.io/mode: Reconcile\n+ kubernetes.io/bootstrapping: rbac-defaults\n+ name: system:coredns\n+rules:\n+- apiGroups:\n+ - \"\"\n+ resources:\n+ - endpoints\n+ - services\n+ - pods\n+ - namespaces\n+ verbs:\n+ - list\n+ - watch\n+- apiGroups:\n+ - \"\"\n+ resources:\n+ - nodes\n+ verbs:\n+ - get\n"
},
{
"change_type": "ADD",
"old_path": null,
"new_path": "scripts/coredns/clusterrolebinding.yaml",
"diff": "+apiVersion: rbac.authorization.k8s.io/v1\n+kind: ClusterRoleBinding\n+metadata:\n+ labels:\n+ addonmanager.kubernetes.io/mode: EnsureExists\n+ kubernetes.io/bootstrapping: rbac-defaults\n+ name: system:coredns\n+roleRef:\n+ apiGroup: rbac.authorization.k8s.io\n+ kind: ClusterRole\n+ name: system:coredns\n+subjects:\n+- kind: ServiceAccount\n+ name: coredns\n+ namespace: kube-system\n"
},
{
"change_type": "ADD",
"old_path": null,
"new_path": "scripts/coredns/configmap.yaml",
"diff": "+apiVersion: v1\n+data:\n+ Corefile: |2-\n+\n+ .:53 {\n+ errors\n+ health\n+ kubernetes cluster.local in-addr.arpa ip6.arpa {\n+ pods insecure\n+ fallthrough in-addr.arpa ip6.arpa\n+ ttl 30\n+ }\n+ prometheus :9153\n+ forward . /etc/resolv.conf\n+ cache 30\n+ loop\n+ reload\n+ loadbalance\n+ }\n+kind: ConfigMap\n+metadata:\n+ labels:\n+ addonmanager.kubernetes.io/mode: EnsureExists\n+ name: coredns\n+ namespace: kube-system\n"
},
{
"change_type": "ADD",
"old_path": null,
"new_path": "scripts/coredns/delete-kube-dns.sh",
"diff": "+kubectl -n kube-system delete svc kube-dns\n+kubectl -n kube-system delete sa kube-dns\n+kubectl -n kube-system delete cm kube-dns\n+kubectl -n kube-system delete deployment kube-dns\n"
},
{
"change_type": "ADD",
"old_path": null,
"new_path": "scripts/coredns/deployment.yaml",
"diff": "+apiVersion: apps/v1\n+kind: Deployment\n+metadata:\n+ labels:\n+ addonmanager.kubernetes.io/mode: Reconcile\n+ k8s-app: kube-dns\n+ kubernetes.io/cluster-service: \"true\"\n+ kubernetes.io/name: CoreDNS\n+ name: coredns\n+ namespace: kube-system\n+spec:\n+ progressDeadlineSeconds: 600\n+ replicas: 3\n+ revisionHistoryLimit: 10\n+ selector:\n+ matchLabels:\n+ k8s-app: kube-dns\n+ strategy:\n+ rollingUpdate:\n+ maxSurge: 25%\n+ maxUnavailable: 1\n+ type: RollingUpdate\n+ template:\n+ metadata:\n+ annotations:\n+ seccomp.security.alpha.kubernetes.io/pod: docker/default\n+ labels:\n+ k8s-app: kube-dns\n+ spec:\n+ affinity:\n+ podAntiAffinity:\n+ requiredDuringSchedulingIgnoredDuringExecution:\n+ - labelSelector:\n+ matchExpressions:\n+ - key: k8s-app\n+ operator: In\n+ values:\n+ - kube-dns\n+ topologyKey: kubernetes.io/hostname\n+ containers:\n+ - args:\n+ - -conf\n+ - /etc/coredns/Corefile\n+ image: keppel.eu-de-1.cloud.sap/ccloud-dockerhub-mirror/sapcc/coredns:1.6.2\n+ imagePullPolicy: IfNotPresent\n+ livenessProbe:\n+ failureThreshold: 5\n+ httpGet:\n+ path: /health\n+ port: 8080\n+ scheme: HTTP\n+ initialDelaySeconds: 60\n+ periodSeconds: 10\n+ successThreshold: 1\n+ timeoutSeconds: 5\n+ name: coredns\n+ ports:\n+ - containerPort: 53\n+ name: dns\n+ protocol: UDP\n+ - containerPort: 53\n+ name: dns-tcp\n+ protocol: TCP\n+ - containerPort: 9153\n+ name: metrics\n+ protocol: TCP\n+ readinessProbe:\n+ failureThreshold: 3\n+ httpGet:\n+ path: /health\n+ port: 8080\n+ scheme: HTTP\n+ periodSeconds: 10\n+ successThreshold: 1\n+ timeoutSeconds: 1\n+ resources:\n+ limits:\n+ memory: 170Mi\n+ requests:\n+ cpu: 100m\n+ memory: 70Mi\n+ securityContext:\n+ allowPrivilegeEscalation: false\n+ capabilities:\n+ add:\n+ - NET_BIND_SERVICE\n+ drop:\n+ - all\n+ readOnlyRootFilesystem: true\n+ terminationMessagePath: /dev/termination-log\n+ terminationMessagePolicy: File\n+ volumeMounts:\n+ - mountPath: /etc/coredns\n+ name: config-volume\n+ readOnly: true\n+ dnsPolicy: Default\n+ nodeSelector:\n+ beta.kubernetes.io/os: linux\n+ priorityClassName: system-cluster-critical\n+ restartPolicy: Always\n+ schedulerName: default-scheduler\n+ securityContext: {}\n+ serviceAccount: coredns\n+ serviceAccountName: coredns\n+ terminationGracePeriodSeconds: 30\n+ tolerations:\n+ - key: CriticalAddonsOnly\n+ operator: Exists\n+ volumes:\n+ - configMap:\n+ defaultMode: 420\n+ items:\n+ - key: Corefile\n+ path: Corefile\n+ name: coredns\n+ name: config-volume\n"
},
{
"change_type": "ADD",
"old_path": null,
"new_path": "scripts/coredns/dnsutils.yaml",
"diff": "+apiVersion: v1\n+kind: Pod\n+metadata:\n+ name: dnsutils\n+spec:\n+ containers:\n+ - name: dnsutils\n+ image: gcr.io/kubernetes-e2e-test-images/dnsutils:1.3\n+ command:\n+ - sleep\n+ - \"3600\"\n+ imagePullPolicy: IfNotPresent\n+ restartPolicy: Always\n"
},
{
"change_type": "ADD",
"old_path": null,
"new_path": "scripts/coredns/install-coredns.sh",
"diff": "+kubectl -n kube-system apply -f clusterrolebinding.yaml -f clusterrole.yaml -f configmap.yaml -f deployment.yaml -f serviceaccount.yaml -f service.yaml\n+\n"
},
{
"change_type": "ADD",
"old_path": null,
"new_path": "scripts/coredns/service.yaml",
"diff": "+apiVersion: v1\n+kind: Service\n+metadata:\n+ annotations:\n+ prometheus.io/port: \"9153\"\n+ prometheus.io/scrape: \"true\"\n+ labels:\n+ addonmanager.kubernetes.io/mode: Reconcile\n+ k8s-app: kube-dns\n+ kubernetes.io/cluster-service: \"true\"\n+ kubernetes.io/name: CoreDNS\n+ name: kube-dns\n+ namespace: kube-system\n+spec:\n+ clusterIP: 198.18.128.2\n+ ports:\n+ - name: dns\n+ port: 53\n+ protocol: UDP\n+ targetPort: 53\n+ - name: dns-tcp\n+ port: 53\n+ protocol: TCP\n+ targetPort: 53\n+ - name: metrics\n+ port: 9153\n+ protocol: TCP\n+ targetPort: 9153\n+ selector:\n+ k8s-app: kube-dns\n+ sessionAffinity: None\n+ type: ClusterIP\n"
},
{
"change_type": "ADD",
"old_path": null,
"new_path": "scripts/coredns/serviceaccount.yaml",
"diff": "+apiVersion: v1\n+kind: ServiceAccount\n+metadata:\n+ labels:\n+ addonmanager.kubernetes.io/mode: Reconcile\n+ kubernetes.io/cluster-service: \"true\"\n+ name: coredns\n+ namespace: kube-system\n"
},
{
"change_type": "ADD",
"old_path": null,
"new_path": "scripts/coredns/test-dns.sh",
"diff": "+kubectl -n default apply -f dnsutils.yaml\n+sleep 30\n+kubectl -n default exec -i -t dnsutils -- nslookup kubernetes.default\n+\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
[ci skip] Add coredns scripts
|
596,227 |
28.10.2021 16:43:21
| -7,200 |
2c733d7d31a91b6351240b8f461e4bf3abdd7096
|
Skip servicing if maintenance-controller is assumed
The servicing controller now skips clusters, which have at least one node
with the "cloud.sap/maintenance-profile" label set, which in turn indicates
the presence of sapcc/maintenance-controller.
|
[
{
"change_type": "MODIFY",
"old_path": "pkg/controller/servicing/controller_test.go",
"new_path": "pkg/controller/servicing/controller_test.go",
"diff": "@@ -281,6 +281,26 @@ func TestServicingControllerReconcile(t *testing.T) {\nexpectedReboot: false,\nexpectedReplace: false,\n},\n+ {\n+ message: \"Klusters assumed to be maintained by the maintenance-controller should neither be rebooted nor replaced\",\n+ options: &FakeKlusterOptions{\n+ Phase: models.KlusterPhaseRunning,\n+ NodePools: []FakeNodePoolOptions{\n+ {\n+ AllowReboot: true,\n+ AllowReplace: true,\n+ NodeHealthy: true,\n+ NodeOSOutdated: true,\n+ NodeKubeletOutdated: true,\n+ Size: 2,\n+ Labels: []string{\"cloud.sap/maintenance-profile=flatcar\"},\n+ },\n+ },\n+ },\n+ expectedDrain: false,\n+ expectedReboot: false,\n+ expectedReplace: false,\n+ },\n} {\nt.Run(string(subject.message), func(t *testing.T) {\nkluster, nodes := NewFakeKluster(subject.options, true)\n"
},
{
"change_type": "MODIFY",
"old_path": "pkg/controller/servicing/lister.go",
"new_path": "pkg/controller/servicing/lister.go",
"diff": "@@ -24,6 +24,7 @@ import (\nconst (\nAnnotationNodeForceReplace = \"kubernikus.cloud.sap/forceReplace\"\nAnnotationNodeSkipReplace = \"kubernikus.cloud.sap/skipReplace\"\n+ LabelMaintenanceController = \"cloud.sap/maintenance-profile\"\n)\ntype (\n@@ -36,6 +37,7 @@ type (\nSuccessful() []*core_v1.Node\nFailed() []*core_v1.Node\nNotReady() []*core_v1.Node\n+ Maintained() []*core_v1.Node\n}\n// ListerFactory produces a Lister\n@@ -416,6 +418,20 @@ func (d *NodeLister) Failed() []*core_v1.Node {\nreturn found\n}\n+// Returns all nodes that are assumed to be maintained by the maintenance-controller.\n+func (d *NodeLister) Maintained() []*core_v1.Node {\n+ var found []*core_v1.Node\n+\n+ for _, node := range d.All() {\n+ _, ok := node.Labels[LabelMaintenanceController]\n+ if ok {\n+ found = append(found, node)\n+ }\n+ }\n+\n+ return found\n+}\n+\nfunc (d *NodeLister) updateTimeout() []*core_v1.Node {\nvar found []*core_v1.Node\n@@ -572,6 +588,18 @@ func (l *LoggingLister) Failed() (nodes []*core_v1.Node) {\nreturn l.Lister.Failed()\n}\n+func (l *LoggingLister) Maintained() (nodes []*core_v1.Node) {\n+ defer func(begin time.Time) {\n+ l.Logger.Log(\n+ \"msg\", \"listing nodes assumed to be maintained by the maintenance-controller\",\n+ \"took\", time.Since(begin),\n+ \"count\", len(nodes),\n+ \"v\", 3,\n+ )\n+ }(time.Now())\n+ return l.Lister.Maintained()\n+}\n+\nfunc getKubeletVersion(node *core_v1.Node) (*version.Version, error) {\nreturn version.ParseSemantic(node.Status.NodeInfo.KubeletVersion)\n}\n"
},
{
"change_type": "MODIFY",
"old_path": "pkg/controller/servicing/reconciler.go",
"new_path": "pkg/controller/servicing/reconciler.go",
"diff": "@@ -118,7 +118,12 @@ func (r *KlusterReconciler) Do() error {\n}\nif util.DisabledValue(r.Kluster.ObjectMeta.Annotations[AnnotationServicingSafeguard]) {\n- r.Logger.Log(\"msg\", \"Skippig upgrades. Manually disabled with safeguard annotation.\")\n+ r.Logger.Log(\"msg\", \"Skippig upgrades. Manually disabled with safeguard annotation.\", \"v\", 2)\n+ return nil\n+ }\n+\n+ if len(r.Lister.Maintained()) > 0 {\n+ r.Logger.Log(\"msg\", \"Skipping upgrades. At least one node seems to be maintained by the maintenance-controller.\", \"v\", 2)\nreturn nil\n}\n"
},
{
"change_type": "MODIFY",
"old_path": "pkg/controller/servicing/testing.go",
"new_path": "pkg/controller/servicing/testing.go",
"diff": "@@ -3,6 +3,7 @@ package servicing\nimport (\n\"fmt\"\n\"os\"\n+ \"strings\"\n\"time\"\n\"github.com/go-kit/kit/log\"\n@@ -47,6 +48,7 @@ type FakeNodePoolOptions struct {\nNodeKubeletOutdated bool\nNodeUpdating *time.Time\nSize int\n+ Labels []string\n}\n// NewFakeKluster creates a Kluster Object for tests\n@@ -81,15 +83,23 @@ func NewFakeKluster(opts *FakeKlusterOptions, afterFlatCarRktRemoval bool) (*v1.\nAllowReplace: &allowReboot,\nAllowReboot: &allowReplace,\n},\n+ Labels: p.Labels,\n}\nkluster.Spec.NodePools = append(kluster.Spec.NodePools, pool)\nfor j := 0; j < p.Size; j++ {\n+ labels := make(map[string]string)\n+ for _, label := range p.Labels {\n+ splitted := strings.Split(label, \"=\")\n+ labels[splitted[0]] = splitted[1]\n+ }\n+\nnodeName := fmt.Sprintf(\"test-%s-0000%d\", poolName, j)\nnode := &core_v1.Node{\nObjectMeta: meta_v1.ObjectMeta{\nName: nodeName,\nAnnotations: map[string]string{},\n+ Labels: labels,\n},\nStatus: core_v1.NodeStatus{\nPhase: core_v1.NodeRunning,\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
Skip servicing if maintenance-controller is assumed (#623)
The servicing controller now skips clusters, which have at least one node
with the "cloud.sap/maintenance-profile" label set, which in turn indicates
the presence of sapcc/maintenance-controller.
|
596,226 |
29.10.2021 11:16:06
| -7,200 |
7a156ae3c36f23027ec2bfeceff9c3de41b27af3
|
Only cleanup containers for not running klusters in e2e test
|
[
{
"change_type": "MODIFY",
"old_path": "test/e2e/pyrolisis_test.go",
"new_path": "test/e2e/pyrolisis_test.go",
"diff": "@@ -6,6 +6,7 @@ import (\n\"testing\"\n\"time\"\n+ \"github.com/go-openapi/runtime\"\n\"github.com/gophercloud/gophercloud\"\n\"github.com/gophercloud/gophercloud/openstack\"\n\"github.com/gophercloud/gophercloud/openstack/blockstorage/v3/volumes\"\n@@ -19,6 +20,7 @@ import (\n\"k8s.io/apimachinery/pkg/util/wait\"\n\"github.com/sapcc/kubernikus/pkg/api/client/operations\"\n+ \"github.com/sapcc/kubernikus/pkg/api/models\"\netcd_util \"github.com/sapcc/kubernikus/pkg/util/etcd\"\n\"github.com/sapcc/kubernikus/test/e2e/framework\"\n)\n@@ -46,22 +48,18 @@ func (p *PyrolisisTests) Run(t *testing.T) {\nt.Run(\"CleanupVolumes\", p.CleanupVolumes)\nt.Run(\"CleanupInstances\", p.CleanupInstances)\n+ t.Run(\"CleanupLoadbalancers\", p.CleanupLoadbalancers)\n+ }\ncleanupStorageContainer := t.Run(\"CleanupBackupStorageContainers\", p.CleanupBackupStorageContainers)\nrequire.True(t, cleanupStorageContainer, \"Etcd backup storage container cleanup failed\")\n-\n- t.Run(\"CleanupLoadbalancers\", p.CleanupLoadbalancers)\n- }\n}\nfunc (p *PyrolisisTests) SettingKlustersOnFire(t *testing.T) {\n- res, err := p.Kubernikus.Client.Operations.ListClusters(\n- operations.NewListClustersParams(),\n- p.Kubernikus.AuthInfo,\n- )\n+ klusters, err := listKlusters(p.Kubernikus.Client.Operations, p.Kubernikus.AuthInfo)\nrequire.NoError(t, err, \"There should be no error while listing klusters\")\n- for _, kluster := range res.Payload {\n+ for _, kluster := range klusters {\nif strings.HasPrefix(kluster.Name, \"e2e-\") {\nt.Run(fmt.Sprintf(\"TerminatingKluster-%v\", kluster.Name), func(t *testing.T) {\n_, err := p.Kubernikus.Client.Operations.TerminateCluster(\n@@ -74,6 +72,19 @@ func (p *PyrolisisTests) SettingKlustersOnFire(t *testing.T) {\n}\n}\n+func listKlusters(client *operations.Client, authinfo runtime.ClientAuthInfoWriter) ([]*models.Kluster, error) {\n+ res, err := client.ListClusters(\n+ operations.NewListClustersParams(),\n+ authinfo,\n+ )\n+\n+ if err != nil {\n+ return nil, err\n+ }\n+\n+ return res.Payload, nil\n+}\n+\nfunc (p *PyrolisisTests) WaitForE2EKlustersTerminated(t *testing.T) {\nerr := p.Kubernikus.WaitForKlusters(\"e2e-\", 0, WaitForKlusterToBeDeletedTimeout)\nassert.NoError(t, err, \"E2E Klusters didn't burn down in time\")\n@@ -92,10 +103,30 @@ func (p *PyrolisisTests) CleanupBackupStorageContainers(t *testing.T) {\nallContainers, err := containers.ExtractNames(allPages)\nrequire.NoError(t, err, \"There should be no error while extracting storage containers\")\n+ klusters, err := listKlusters(p.Kubernikus.Client.Operations, p.Kubernikus.AuthInfo)\n+ require.NoError(t, err, \"There should be no error while listing klusters\")\n+\n+ // do not delete containers where there is still a kluster running\n+ var containersToDelete []string\n+ for _, container := range allContainers {\n+ found := false\n+\n+ for _, kluster := range klusters {\n+ if strings.HasPrefix(container, fmt.Sprintf(\"%s-%s\", etcd_util.BackupStorageContainerBase, kluster.Name)) {\n+ found = true\n+ }\n+ }\n+\n+ if !found {\n+ containersToDelete = append(containersToDelete, container)\n+ }\n+ }\n+\nobjectsListOpts := objects.ListOpts{\nFull: false,\n}\n- for _, container := range allContainers {\n+\n+ for _, container := range containersToDelete {\nif strings.HasPrefix(container, etcd_util.BackupStorageContainerBase) {\nallPages, err := objects.List(storageClient, container, objectsListOpts).AllPages()\nrequire.NoError(t, err, \"There should be no error while lising objetcs in container %s\", container)\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
Only cleanup containers for not running klusters in e2e test
|
596,226 |
29.10.2021 12:13:20
| -7,200 |
f6731106daa0e9ea1b95316f06d02ed1ee4ce08a
|
Only cleanup loadbalancers for not running klusters in e2e test
|
[
{
"change_type": "MODIFY",
"old_path": "test/e2e/pyrolisis_test.go",
"new_path": "test/e2e/pyrolisis_test.go",
"diff": "@@ -48,11 +48,12 @@ func (p *PyrolisisTests) Run(t *testing.T) {\nt.Run(\"CleanupVolumes\", p.CleanupVolumes)\nt.Run(\"CleanupInstances\", p.CleanupInstances)\n- t.Run(\"CleanupLoadbalancers\", p.CleanupLoadbalancers)\n}\ncleanupStorageContainer := t.Run(\"CleanupBackupStorageContainers\", p.CleanupBackupStorageContainers)\nrequire.True(t, cleanupStorageContainer, \"Etcd backup storage container cleanup failed\")\n+\n+ t.Run(\"CleanupLoadbalancers\", p.CleanupLoadbalancers)\n}\nfunc (p *PyrolisisTests) SettingKlustersOnFire(t *testing.T) {\n@@ -227,7 +228,26 @@ func (p *PyrolisisTests) CleanupLoadbalancers(t *testing.T) {\nallLoadbalancers, err := loadbalancers.ExtractLoadBalancers(allPages)\nrequire.NoError(t, err, \"There should be no error while extracting loadbalancers\")\n+ klusters, err := listKlusters(p.Kubernikus.Client.Operations, p.Kubernikus.AuthInfo)\n+ require.NoError(t, err, \"There should be no error while listing klusters\")\n+\n+ // do not delete loadbalancers where there is still a kluster running\n+ var lbsToDelete []loadbalancers.LoadBalancer\nfor _, lb := range allLoadbalancers {\n+ found := false\n+\n+ for _, kluster := range klusters {\n+ if strings.HasPrefix(lb.Name, fmt.Sprintf(\"kube_service_%s\", kluster.Name)) {\n+ found = true\n+ }\n+ }\n+\n+ if !found {\n+ lbsToDelete = append(lbsToDelete, lb)\n+ }\n+ }\n+\n+ for _, lb := range lbsToDelete {\nif strings.HasSuffix(lb.Name, \"_e2e-lb\") {\nerr := loadbalancers.Delete(lbClient, lb.ID, loadbalancers.DeleteOpts{Cascade: true}).ExtractErr()\nrequire.NoError(t, err, \"There should be no error while deleting loadbalancer %s\", lb.Name)\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
Only cleanup loadbalancers for not running klusters in e2e test
|
596,226 |
29.10.2021 13:26:35
| -7,200 |
eef32578eba6ae9cc66e9148451530472fdb760d
|
Ignore PENDING_DELETE on lb deletion
|
[
{
"change_type": "MODIFY",
"old_path": "test/e2e/pyrolisis_test.go",
"new_path": "test/e2e/pyrolisis_test.go",
"diff": "@@ -250,6 +250,12 @@ func (p *PyrolisisTests) CleanupLoadbalancers(t *testing.T) {\nfor _, lb := range lbsToDelete {\nif strings.HasSuffix(lb.Name, \"_e2e-lb\") {\nerr := loadbalancers.Delete(lbClient, lb.ID, loadbalancers.DeleteOpts{Cascade: true}).ExtractErr()\n+\n+ // Ignore PENDING_DELETE error\n+ if _, ok := err.(gophercloud.Err409er); ok {\n+ continue\n+ }\n+\nrequire.NoError(t, err, \"There should be no error while deleting loadbalancer %s\", lb.Name)\n}\n}\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
Ignore PENDING_DELETE on lb deletion
|
596,226 |
29.10.2021 14:22:29
| -7,200 |
f907a71ee96b36736a017d18dbd8913d8338dc6a
|
Container might already be deleted
|
[
{
"change_type": "MODIFY",
"old_path": "test/e2e/pyrolisis_test.go",
"new_path": "test/e2e/pyrolisis_test.go",
"diff": "@@ -130,7 +130,10 @@ func (p *PyrolisisTests) CleanupBackupStorageContainers(t *testing.T) {\nfor _, container := range containersToDelete {\nif strings.HasPrefix(container, etcd_util.BackupStorageContainerBase) {\nallPages, err := objects.List(storageClient, container, objectsListOpts).AllPages()\n- require.NoError(t, err, \"There should be no error while lising objetcs in container %s\", container)\n+ if _, ok := err.(gophercloud.ErrDefault404); ok {\n+ continue\n+ }\n+ require.NoError(t, err, \"There should be no error while lising objetcs in container %s:\", container)\nallObjects, err := objects.ExtractNames(allPages)\nrequire.NoError(t, err, \"There should be no error while extracting objetcs names for container %s\", container)\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
Container might already be deleted
|
596,226 |
11.11.2021 10:52:49
| -3,600 |
b91770d68fce11721ae07daaf4404a431d894947
|
Fix e2e test lb deletion
|
[
{
"change_type": "MODIFY",
"old_path": "test/e2e/pyrolisis_test.go",
"new_path": "test/e2e/pyrolisis_test.go",
"diff": "@@ -255,7 +255,7 @@ func (p *PyrolisisTests) CleanupLoadbalancers(t *testing.T) {\nerr := loadbalancers.Delete(lbClient, lb.ID, loadbalancers.DeleteOpts{Cascade: true}).ExtractErr()\n// Ignore PENDING_DELETE error\n- if _, ok := err.(gophercloud.Err409er); ok {\n+ if _, ok := err.(gophercloud.ErrDefault409); ok {\ncontinue\n}\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
Fix e2e test lb deletion
|
596,226 |
15.11.2021 13:56:18
| -3,600 |
42051de35ae31406a34e7149940c52920f55757e
|
Raise csi plugin limits
|
[
{
"change_type": "MODIFY",
"old_path": "charts/kube-master/values.yaml",
"new_path": "charts/kube-master/values.yaml",
"diff": "@@ -128,8 +128,8 @@ csi:\ncpu: 20m\nmemory: 50Mi\nlimits:\n- cpu: 50m\n- memory: 80Mi\n+ cpu: 100m\n+ memory: 100Mi\nprovisioner:\nrequests:\ncpu: 10m\n@@ -142,8 +142,8 @@ csi:\ncpu: 10m\nmemory: 32Mi\nlimits:\n- cpu: 30m\n- memory: 50Mi\n+ cpu: 100m\n+ memory: 100Mi\nsnapshotter:\nrequests:\ncpu: 10m\n@@ -163,5 +163,5 @@ csi:\ncpu: 10m\nmemory: 32Mi\nlimits:\n- cpu: 30m\n- memory: 50Mi\n+ cpu: 100m\n+ memory: 100Mi\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
Raise csi plugin limits
|
596,226 |
30.11.2021 11:36:05
| -3,600 |
df436eb8b4529be87c04a8812059be6e78ba1faf
|
Update oci-build-task image
|
[
{
"change_type": "MODIFY",
"old_path": "ci/pipeline.yaml",
"new_path": "ci/pipeline.yaml",
"diff": "@@ -919,8 +919,8 @@ task_oci_build: &task_oci_build\nimage_resource:\ntype: registry-image\nsource:\n- repository: keppel.eu-de-2.cloud.sap/ci/oci-build-task\n- #repository: vito/oci-build-task\n+ repository: keppel.eu-de-2.cloud.sap/ccloud-dockerhub-mirror/concourse/oci-build-task\n+ tag: 0.8.1\ninputs:\n- name: context\n"
},
{
"change_type": "MODIFY",
"old_path": "ci/task_oci_build.yaml",
"new_path": "ci/task_oci_build.yaml",
"diff": "@@ -2,8 +2,8 @@ platform: linux\nimage_resource:\ntype: registry-image\nsource:\n- repository: keppel.eu-de-2.cloud.sap/ci/oci-build-task\n- #repository: vito/oci-build-task\n+ repository: keppel.eu-de-2.cloud.sap/ccloud-dockerhub-mirror/concourse/oci-build-task\n+ tag: 0.8.1\ninputs:\n- name: context\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
Update oci-build-task image
|
596,226 |
02.12.2021 16:43:07
| -3,600 |
3d768210318ba91d10b3f1746f29520c39f62492
|
Use git-resource-proxy in ci
|
[
{
"change_type": "MODIFY",
"old_path": "ci/pipeline.yaml",
"new_path": "ci/pipeline.yaml",
"diff": "@@ -1126,7 +1126,7 @@ resources:\nsource:\nusername: ((keppel.username))\npassword: ((keppel.password))\n- repository: keppel.eu-de-1.cloud.sap/ccloud/kubernikus-docs-builder\n+ repository: keppel.eu-de-2.cloud.sap/ccloud/kubernikus-docs-builder\ntag: latest\n- name: kubernikus.image\n@@ -1136,7 +1136,7 @@ resources:\nsource:\nusername: ((keppel.username))\npassword: ((keppel.password))\n- repository: keppel.eu-de-1.cloud.sap/ccloud/kubernikus\n+ repository: keppel.eu-de-2.cloud.sap/ccloud/kubernikus\ntag: latest\n- name: kubernikusctl.release\n@@ -1148,7 +1148,7 @@ resources:\n- name: secrets.git\nicon: key-variant\n- type: git\n+ type: git-proxy\nwebhook_token: aldhjalkdhahdjkahdjkhjadhjadhkjadlkjhAAdd\ncheck_every: 1h\nsource:\n@@ -1159,7 +1159,7 @@ resources:\n- name: tiller-version.git\nicon: key-variant\n- type: git\n+ type: git-proxy\nwebhook_token: aldhjalkdhahdjkahdjkhjadhjadhkjadlkjhAAdd\ncheck_every: 1h\nsource:\n@@ -1235,22 +1235,29 @@ resource_types:\ntype: docker-image\ncheck_every: 24h\nsource:\n- repository: keppel.eu-de-1.cloud.sap/ccloud-dockerhub-mirror/arbourd/concourse-slack-alert-resource\n+ repository: keppel.eu-de-2.cloud.sap/ccloud-dockerhub-mirror/arbourd/concourse-slack-alert-resource\n- name: slack-notification\ntype: docker-image\ncheck_every: 24h\nsource:\n- repository: keppel.eu-de-1.cloud.sap/ccloud-dockerhub-mirror/cfcommunity/slack-notification-resource\n+ repository: keppel.eu-de-2.cloud.sap/ccloud-dockerhub-mirror/cfcommunity/slack-notification-resource\ntag: latest\n- name: time-version-resource\ntype: docker-image\ncheck_every: 24h\nsource:\n- repository: keppel.eu-de-1.cloud.sap/ccloud/concourse-time-version-resource\n+ repository: keppel.eu-de-2.cloud.sap/ccloud/concourse-time-version-resource\ntag: v2\n+ - name: git-proxy\n+ type: registry-image\n+ check_every: 24h\n+ source:\n+ repository: keppel.eu-de-2.cloud.sap/ccloud-dockerhub-mirror/sapcc/concourse-git-resource-proxy\n+ tag: 0.5.0\n+\njobs:\n- name: docs\nserial: true\n"
},
{
"change_type": "MODIFY",
"old_path": "ci/pipeline.yaml.erb",
"new_path": "ci/pipeline.yaml.erb",
"diff": "@@ -126,7 +126,7 @@ resources:\nsource:\nusername: ((keppel.username))\npassword: ((keppel.password))\n- repository: keppel.eu-de-1.cloud.sap/ccloud/kubernikus-docs-builder\n+ repository: keppel.eu-de-2.cloud.sap/ccloud/kubernikus-docs-builder\ntag: latest\n- name: kubernikus.image\n@@ -136,7 +136,7 @@ resources:\nsource:\nusername: ((keppel.username))\npassword: ((keppel.password))\n- repository: keppel.eu-de-1.cloud.sap/ccloud/kubernikus\n+ repository: keppel.eu-de-2.cloud.sap/ccloud/kubernikus\ntag: latest\n- name: kubernikusctl.release\n@@ -148,7 +148,7 @@ resources:\n- name: secrets.git\nicon: key-variant\n- type: git\n+ type: git-proxy\nwebhook_token: aldhjalkdhahdjkahdjkhjadhjadhkjadlkjhAAdd\ncheck_every: 1h\nsource:\n@@ -159,7 +159,7 @@ resources:\n- name: tiller-version.git\nicon: key-variant\n- type: git\n+ type: git-proxy\nwebhook_token: aldhjalkdhahdjkahdjkhjadhjadhkjadlkjhAAdd\ncheck_every: 1h\nsource:\n@@ -214,22 +214,29 @@ resource_types:\ntype: docker-image\ncheck_every: 24h\nsource:\n- repository: keppel.eu-de-1.cloud.sap/ccloud-dockerhub-mirror/arbourd/concourse-slack-alert-resource\n+ repository: keppel.eu-de-2.cloud.sap/ccloud-dockerhub-mirror/arbourd/concourse-slack-alert-resource\n- name: slack-notification\ntype: docker-image\ncheck_every: 24h\nsource:\n- repository: keppel.eu-de-1.cloud.sap/ccloud-dockerhub-mirror/cfcommunity/slack-notification-resource\n+ repository: keppel.eu-de-2.cloud.sap/ccloud-dockerhub-mirror/cfcommunity/slack-notification-resource\ntag: latest\n- name: time-version-resource\ntype: docker-image\ncheck_every: 24h\nsource:\n- repository: keppel.eu-de-1.cloud.sap/ccloud/concourse-time-version-resource\n+ repository: keppel.eu-de-2.cloud.sap/ccloud/concourse-time-version-resource\ntag: v2\n+ - name: git-proxy\n+ type: registry-image\n+ check_every: 24h\n+ source:\n+ repository: keppel.eu-de-2.cloud.sap/ccloud-dockerhub-mirror/sapcc/concourse-git-resource-proxy\n+ tag: 0.5.0\n+\njobs:\n- name: docs\nserial: true\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
Use git-resource-proxy in ci
|
596,227 |
04.01.2022 14:30:42
| -3,600 |
7635353de2fe9c1e5ed9273a33800c079ded193a
|
Add HTTP ingestion for audit logs
Can be used to ship logs to logstash.
|
[
{
"change_type": "MODIFY",
"old_path": "charts/kube-master/templates/api.yaml",
"new_path": "charts/kube-master/templates/api.yaml",
"diff": "@@ -402,6 +402,23 @@ spec:\nname: {{ include \"master.fullname\" . }}-secret\nkey: openstack-region\n{{- end}}\n+ {{- if eq .Values.audit \"http\" }}\n+ - name: HTTP_USERNAME\n+ valueFrom:\n+ secretKeyRef:\n+ name: {{ include \"master.fullname\" . }}-http\n+ key: http-username\n+ - name: HTTP_PASSWORD\n+ valueFrom:\n+ secretKeyRef:\n+ name: {{ include \"master.fullname\" . }}-http\n+ key: http-password\n+ - name: HTTP_ENDPOINT\n+ valueFrom:\n+ secretKeyRef:\n+ name: {{ include \"master.fullname\" . }}-http\n+ key: http-endpoint\n+ {{- end }}\nvolumeMounts:\n- mountPath: /fluentd/etc\nname: config\n"
},
{
"change_type": "MODIFY",
"old_path": "charts/kube-master/templates/configmap.yaml",
"new_path": "charts/kube-master/templates/configmap.yaml",
"diff": "@@ -171,4 +171,25 @@ data:\n</buffer>\n</match>\n{{- end }}\n+\n+ {{- if eq .Values.audit \"http\" }}\n+ <match apiserver.audit>\n+ @type http\n+ endpoint \"#{ENV['HTTP_ENDPOINT']}\"\n+ json_array true\n+ <auth>\n+ method basic\n+ username \"#{ENV['HTTP_USERNAME']}\"\n+ password \"#{ENV['HTTP_PASSWORD']}\"\n+ </auth>\n+ <format>\n+ @type json\n+ </format>\n+ <buffer>\n+ @type file\n+ path /var/log/audit.buf\n+ chunk_limit_size 8m\n+ </buffer>\n+ </match>\n+ {{- end}}\n{{- end }}\n"
},
{
"change_type": "MODIFY",
"old_path": "pkg/api/models/kluster_spec.go",
"new_path": "pkg/api/models/kluster_spec.go",
"diff": "@@ -27,7 +27,7 @@ type KlusterSpec struct {\nAdvertisePort int64 `json:\"advertisePort\"`\n// audit\n- // Enum: [elasticsearch swift]\n+ // Enum: [elasticsearch swift http]\nAudit *string `json:\"audit,omitempty\"`\n// backup\n@@ -127,7 +127,7 @@ var klusterSpecTypeAuditPropEnum []interface{}\nfunc init() {\nvar res []string\n- if err := json.Unmarshal([]byte(`[\"elasticsearch\",\"swift\"]`), &res); err != nil {\n+ if err := json.Unmarshal([]byte(`[\"elasticsearch\",\"swift\",\"http\"]`), &res); err != nil {\npanic(err)\n}\nfor _, v := range res {\n@@ -142,6 +142,9 @@ const (\n// KlusterSpecAuditSwift captures enum value \"swift\"\nKlusterSpecAuditSwift string = \"swift\"\n+\n+ // KlusterSpecAuditHTTP captures enum value \"http\"\n+ KlusterSpecAuditHTTP string = \"http\"\n)\n// prop value enum\n"
},
{
"change_type": "MODIFY",
"old_path": "pkg/api/spec/embedded_spec.go",
"new_path": "pkg/api/spec/embedded_spec.go",
"diff": "@@ -598,7 +598,8 @@ func init() {\n\"type\": \"string\",\n\"enum\": [\n\"elasticsearch\",\n- \"swift\"\n+ \"swift\",\n+ \"http\"\n],\n\"x-nullable\": true\n},\n@@ -1643,7 +1644,8 @@ func init() {\n\"type\": \"string\",\n\"enum\": [\n\"elasticsearch\",\n- \"swift\"\n+ \"swift\",\n+ \"http\"\n],\n\"x-nullable\": true\n},\n"
},
{
"change_type": "MODIFY",
"old_path": "swagger.yml",
"new_path": "swagger.yml",
"diff": "@@ -470,7 +470,7 @@ definitions:\naudit:\ntype: string\nx-nullable: true\n- enum: [\"elasticsearch\", \"swift\"]\n+ enum: [\"elasticsearch\", \"swift\", \"http\"]\ndefault: null\nnoCloud:\ntype: boolean\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
Add HTTP ingestion for audit logs (#632)
Can be used to ship logs to logstash.
|
596,226 |
05.01.2022 15:33:51
| -3,600 |
c41e18141cedf3fd80f3c5f3cd9cd28c0378b211
|
Add AVZ to e2e tests
|
[
{
"change_type": "MODIFY",
"old_path": "ci/pipeline.yaml",
"new_path": "ci/pipeline.yaml",
"diff": "@@ -20,6 +20,8 @@ auth_e2e_ap-ae-1: &auth_e2e_ap-ae-1\nKUBERNIKUS_NAME: e2e\nOS_PROJECT_DOMAIN_NAME: ccadmin\n+ NODEPOOL_AVZ: ap-ae-1a\n+\nOS_AUTH_URL: https://identity-3.ap-ae-1.cloud.sap/v3\nOS_USERNAME: kubernikus-pipeline\n@@ -52,6 +54,8 @@ auth_e2e_ap-au-1: &auth_e2e_ap-au-1\nKUBERNIKUS_NAME: e2e\nOS_PROJECT_DOMAIN_NAME: ccadmin\n+ NODEPOOL_AVZ: ap-au-1a\n+\nOS_AUTH_URL: https://identity-3.ap-au-1.cloud.sap/v3\n@@ -84,6 +88,8 @@ auth_e2e_ap-cn-1: &auth_e2e_ap-cn-1\nKUBERNIKUS_NAME: e2e\nOS_PROJECT_DOMAIN_NAME: ccadmin\n+ NODEPOOL_AVZ: ap-cn-1a\n+\nOS_AUTH_URL: https://identity-3.ap-cn-1.cloud.sap/v3\nOS_USERNAME: kubernikus-pipeline\n@@ -116,6 +122,8 @@ auth_e2e_ap-jp-1: &auth_e2e_ap-jp-1\nKUBERNIKUS_NAME: e2e\nOS_PROJECT_DOMAIN_NAME: ccadmin\n+ NODEPOOL_AVZ: ap-jp-1a\n+\nOS_AUTH_URL: https://identity-3.ap-jp-1.cloud.sap/v3\nOS_USERNAME: kubernikus-pipeline\n@@ -148,6 +156,8 @@ auth_e2e_ap-jp-2: &auth_e2e_ap-jp-2\nKUBERNIKUS_NAME: e2e\nOS_PROJECT_DOMAIN_NAME: ccadmin\n+ NODEPOOL_AVZ: ap-jp-2a\n+\nOS_AUTH_URL: https://identity-3.ap-jp-2.cloud.sap/v3\nOS_USERNAME: kubernikus-pipeline\n@@ -180,6 +190,8 @@ auth_e2e_ap-sa-1: &auth_e2e_ap-sa-1\nKUBERNIKUS_NAME: e2e\nOS_PROJECT_DOMAIN_NAME: ccadmin\n+ NODEPOOL_AVZ: ap-sa-1a\n+\nOS_AUTH_URL: https://identity-3.ap-sa-1.cloud.sap/v3\nOS_USERNAME: kubernikus-pipeline\n@@ -212,6 +224,8 @@ auth_e2e_ap-sa-2: &auth_e2e_ap-sa-2\nKUBERNIKUS_NAME: e2e\nOS_PROJECT_DOMAIN_NAME: ccadmin\n+ NODEPOOL_AVZ: ap-sa-2a\n+\nOS_AUTH_URL: https://identity-3.ap-sa-2.cloud.sap/v3\nOS_USERNAME: kubernikus-pipeline\n@@ -244,6 +258,8 @@ auth_e2e_eu-de-1: &auth_e2e_eu-de-1\nKUBERNIKUS_NAME: e2e\nOS_PROJECT_DOMAIN_NAME: ccadmin\n+ NODEPOOL_AVZ: eu-de-1d\n+\nOS_AUTH_URL: https://identity-3.eu-de-1.cloud.sap/v3\n@@ -276,6 +292,8 @@ auth_e2e_eu-de-2: &auth_e2e_eu-de-2\nKUBERNIKUS_NAME: e2e\nOS_PROJECT_DOMAIN_NAME: ccadmin\n+ NODEPOOL_AVZ: eu-de-2b\n+\nOS_AUTH_URL: https://identity-3.eu-de-2.cloud.sap/v3\n@@ -308,6 +326,8 @@ auth_e2e_eu-nl-1: &auth_e2e_eu-nl-1\nKUBERNIKUS_NAME: e2e\nOS_PROJECT_DOMAIN_NAME: ccadmin\n+ NODEPOOL_AVZ: eu-nl-1a\n+\nOS_AUTH_URL: https://identity-3.eu-nl-1.cloud.sap/v3\nOS_USERNAME: kubernikus-pipeline\n@@ -340,6 +360,8 @@ auth_e2e_eu-ru-1: &auth_e2e_eu-ru-1\nKUBERNIKUS_NAME: e2e\nOS_PROJECT_DOMAIN_NAME: ccadmin\n+ NODEPOOL_AVZ: eu-ru-1a\n+\nOS_AUTH_URL: https://identity-3.eu-ru-1.cloud.sap/v3\nOS_USERNAME: kubernikus-pipeline\n@@ -372,6 +394,8 @@ auth_e2e_la-br-1: &auth_e2e_la-br-1\nKUBERNIKUS_NAME: e2e\nOS_PROJECT_DOMAIN_NAME: ccadmin\n+ NODEPOOL_AVZ: la-br-1b\n+\nOS_AUTH_URL: https://identity-3.la-br-1.cloud.sap/v3\nOS_USERNAME: kubernikus-pipeline\n@@ -404,6 +428,8 @@ auth_e2e_na-ca-1: &auth_e2e_na-ca-1\nKUBERNIKUS_NAME: e2e\nOS_PROJECT_DOMAIN_NAME: ccadmin\n+ NODEPOOL_AVZ: na-ca-1b\n+\nOS_AUTH_URL: https://identity-3.na-ca-1.cloud.sap/v3\nOS_USERNAME: kubernikus-pipeline\n@@ -470,6 +496,8 @@ auth_e2e_na-us-2: &auth_e2e_na-us-2\nKUBERNIKUS_NAME: e2e\nOS_PROJECT_DOMAIN_NAME: ccadmin\n+ NODEPOOL_AVZ: na-us-2b\n+\nOS_AUTH_URL: https://identity-3.na-us-2.cloud.sap/v3\nOS_USERNAME: kubernikus-pipeline\n@@ -502,6 +530,8 @@ auth_e2e_na-us-3: &auth_e2e_na-us-3\nKUBERNIKUS_NAME: e2e\nOS_PROJECT_DOMAIN_NAME: ccadmin\n+ NODEPOOL_AVZ: na-us-3a\n+\nOS_AUTH_URL: https://identity-3.na-us-3.cloud.sap/v3\nOS_USERNAME: kubernikus-pipeline\n@@ -568,6 +598,8 @@ auth_e2e_master: &auth_e2e_master\nKUBERNIKUS_NAME: e2e\nOS_PROJECT_DOMAIN_NAME: ccadmin\n+ NODEPOOL_AVZ: eu-nl-1a\n+\n@@ -1126,7 +1158,7 @@ resources:\nsource:\nusername: ((keppel.username))\npassword: ((keppel.password))\n- repository: keppel.eu-de-2.cloud.sap/ccloud/kubernikus-docs-builder\n+ repository: keppel.eu-de-1.cloud.sap/ccloud/kubernikus-docs-builder\ntag: latest\n- name: kubernikus.image\n@@ -1136,7 +1168,7 @@ resources:\nsource:\nusername: ((keppel.username))\npassword: ((keppel.password))\n- repository: keppel.eu-de-2.cloud.sap/ccloud/kubernikus\n+ repository: keppel.eu-de-1.cloud.sap/ccloud/kubernikus\ntag: latest\n- name: kubernikusctl.release\n@@ -1382,6 +1414,29 @@ jobs:\n- put: all-versions.trigger\n+ - name: e2e-1.22.4\n+ serial: true\n+ plan:\n+ - in_parallel:\n+ - get: secrets.git\n+ - get: kubernikus.builds\n+ resource: kubernikus.git\n+ passed: [master]\n+ trigger: true\n+ - get: all-versions.trigger\n+ trigger: true\n+ passed: [all-versions]\n+ - task: e2e_tests\n+ config:\n+ <<: *task_e2e_tests\n+ timeout: 2h\n+ params:\n+ KLUSTER_VERSION: \"1.22.4\"\n+ KLUSTER_CIDR: \"100.100.0.0/16\"\n+ ISOLATE_TEST: \"true\"\n+ <<: *auth_e2e_master\n+ <<: *slack\n+\n- name: e2e-1.21.5\nserial: true\nplan:\n@@ -1400,7 +1455,7 @@ jobs:\ntimeout: 2h\nparams:\nKLUSTER_VERSION: \"1.21.5\"\n- KLUSTER_CIDR: \"100.100.0.0/16\"\n+ KLUSTER_CIDR: \"100.101.0.0/16\"\nISOLATE_TEST: \"true\"\n<<: *auth_e2e_master\n<<: *slack\n@@ -1423,7 +1478,7 @@ jobs:\ntimeout: 2h\nparams:\nKLUSTER_VERSION: \"1.20.8\"\n- KLUSTER_CIDR: \"100.101.0.0/16\"\n+ KLUSTER_CIDR: \"100.102.0.0/16\"\nISOLATE_TEST: \"true\"\n<<: *auth_e2e_master\n<<: *slack\n@@ -1446,7 +1501,7 @@ jobs:\ntimeout: 2h\nparams:\nKLUSTER_VERSION: \"1.19.11\"\n- KLUSTER_CIDR: \"100.102.0.0/16\"\n+ KLUSTER_CIDR: \"100.103.0.0/16\"\nISOLATE_TEST: \"true\"\n<<: *auth_e2e_master\n<<: *slack\n@@ -1469,7 +1524,7 @@ jobs:\ntimeout: 2h\nparams:\nKLUSTER_VERSION: \"1.18.10\"\n- KLUSTER_CIDR: \"100.103.0.0/16\"\n+ KLUSTER_CIDR: \"100.104.0.0/16\"\nISOLATE_TEST: \"true\"\n<<: *auth_e2e_master\n<<: *slack\n@@ -1492,7 +1547,7 @@ jobs:\ntimeout: 2h\nparams:\nKLUSTER_VERSION: \"1.17.13\"\n- KLUSTER_CIDR: \"100.104.0.0/16\"\n+ KLUSTER_CIDR: \"100.105.0.0/16\"\nISOLATE_TEST: \"true\"\n<<: *auth_e2e_master\n<<: *slack\n@@ -1515,7 +1570,7 @@ jobs:\ntimeout: 2h\nparams:\nKLUSTER_VERSION: \"1.16.14\"\n- KLUSTER_CIDR: \"100.105.0.0/16\"\n+ KLUSTER_CIDR: \"100.106.0.0/16\"\nISOLATE_TEST: \"true\"\n<<: *auth_e2e_master\n<<: *slack\n@@ -1538,7 +1593,7 @@ jobs:\ntimeout: 2h\nparams:\nKLUSTER_VERSION: \"1.15.9\"\n- KLUSTER_CIDR: \"100.106.0.0/16\"\n+ KLUSTER_CIDR: \"100.107.0.0/16\"\nISOLATE_TEST: \"true\"\n<<: *auth_e2e_master\n<<: *slack\n@@ -1561,7 +1616,7 @@ jobs:\ntimeout: 2h\nparams:\nKLUSTER_VERSION: \"1.14.5\"\n- KLUSTER_CIDR: \"100.107.0.0/16\"\n+ KLUSTER_CIDR: \"100.108.0.0/16\"\nISOLATE_TEST: \"true\"\n<<: *auth_e2e_master\n<<: *slack\n@@ -1584,7 +1639,7 @@ jobs:\ntimeout: 2h\nparams:\nKLUSTER_VERSION: \"1.13.9\"\n- KLUSTER_CIDR: \"100.108.0.0/16\"\n+ KLUSTER_CIDR: \"100.109.0.0/16\"\nISOLATE_TEST: \"true\"\n<<: *auth_e2e_master\n<<: *slack\n@@ -1607,7 +1662,7 @@ jobs:\ntimeout: 2h\nparams:\nKLUSTER_VERSION: \"1.12.10\"\n- KLUSTER_CIDR: \"100.109.0.0/16\"\n+ KLUSTER_CIDR: \"100.110.0.0/16\"\nISOLATE_TEST: \"true\"\n<<: *auth_e2e_master\n<<: *slack\n@@ -1618,7 +1673,7 @@ jobs:\nplan:\n- get: kubernikus.builds\nresource: kubernikus.git\n- passed: [e2e-1.21.5]\n+ passed: [e2e-1.22.4]\n- task: changes\nconfig:\n<<: *task_github_compare_url\n@@ -3552,6 +3607,8 @@ groups:\n- tiller_master\n- all-versions\n+ - e2e-1.22.4\n+\n- e2e-1.21.5\n- e2e-1.20.8\n@@ -3729,6 +3786,8 @@ groups:\n- tiller_master\n- all-versions\n+ - e2e-1.22.4\n+\n- e2e-1.21.5\n- e2e-1.20.8\n"
},
{
"change_type": "MODIFY",
"old_path": "ci/pipeline.yaml.erb",
"new_path": "ci/pipeline.yaml.erb",
"diff": "@@ -14,24 +14,24 @@ VERSIONS = [\n]\nREGIONS = {\n- \"ap-ae-1\": { continent: 'ASIA', install: 'terraform', e2e: 'terraform', oidc: '9a2d0e1181194fb1bca52e7951666bc3'},\n- \"ap-au-1\": { continent: 'ASIA', install: 'manual' , e2e: 'manual', oidc: 'f34304c03d2147cc964470c31d00ce97' },\n- \"ap-cn-1\": { continent: 'ASIA', install: 'terraform', e2e: 'terraform', oidc: 'e83ddace725b4c5da605e3f66d266f2c' },\n- \"ap-jp-1\": { continent: 'ASIA', install: 'terraform', e2e: 'terraform', oidc: 'b505497ca03e4942a2d0aa50dfbdc7a0' },\n- \"ap-jp-2\": { continent: 'ASIA', install: 'terraform', e2e: 'terraform', oidc: 'f34b8aa879334a7fb20e1b96cb1c576e' },\n- \"ap-sa-1\": { continent: 'ASIA', install: 'terraform', e2e: 'terraform', oidc: '436297dc67b9428f97da11227dc070ec' },\n- \"ap-sa-2\": { continent: 'ASIA', install: 'terraform', e2e: 'terraform', oidc: '37c3822fe4194ab586d9625d81b9d532' },\n- \"eu-de-1\": { continent: 'EMEA', install: 'manual' , e2e: 'manual', oidc: 'd7df5ce6c37643e49b3c93528c29818b' },\n- \"eu-de-2\": { continent: 'EMEA', install: 'manual' , e2e: 'manual', oidc: '57da32d8d50d49a9a9fbbe6b890e9e13' },\n- \"eu-nl-1\": { continent: 'EMEA', install: 'manual' , e2e: 'terraform', oidc: 'd9b056a6762649f684a8a02b30f38324' },\n- \"eu-ru-1\": { continent: 'EMEA', install: 'terraform', e2e: 'terraform', oidc: 'eb31733cfb9d4ab283bb19e00863b780' },\n- \"la-br-1\": { continent: 'AMERICAS', install: 'terraform', e2e: 'terraform', oidc: 'b33ac00cb0004d50bb34eb4e4cbbeb54' },\n- \"na-ca-1\": { continent: 'AMERICAS', install: 'terraform', e2e: 'terraform', oidc: 'b7269ea1dde54a6383453ab69f25cce4' },\n+ \"ap-ae-1\": { continent: 'ASIA', install: 'terraform', e2e: 'terraform', nodepool_avz: 'ap-ae-1a', oidc: '9a2d0e1181194fb1bca52e7951666bc3'},\n+ \"ap-au-1\": { continent: 'ASIA', install: 'manual' , e2e: 'manual', nodepool_avz: 'ap-au-1a', oidc: 'f34304c03d2147cc964470c31d00ce97' },\n+ \"ap-cn-1\": { continent: 'ASIA', install: 'terraform', e2e: 'terraform', nodepool_avz: 'ap-cn-1a', oidc: 'e83ddace725b4c5da605e3f66d266f2c' },\n+ \"ap-jp-1\": { continent: 'ASIA', install: 'terraform', e2e: 'terraform', nodepool_avz: 'ap-jp-1a', oidc: 'b505497ca03e4942a2d0aa50dfbdc7a0' },\n+ \"ap-jp-2\": { continent: 'ASIA', install: 'terraform', e2e: 'terraform', nodepool_avz: 'ap-jp-2a', oidc: 'f34b8aa879334a7fb20e1b96cb1c576e' },\n+ \"ap-sa-1\": { continent: 'ASIA', install: 'terraform', e2e: 'terraform', nodepool_avz: 'ap-sa-1a', oidc: '436297dc67b9428f97da11227dc070ec' },\n+ \"ap-sa-2\": { continent: 'ASIA', install: 'terraform', e2e: 'terraform', nodepool_avz: 'ap-sa-2a', oidc: '37c3822fe4194ab586d9625d81b9d532' },\n+ \"eu-de-1\": { continent: 'EMEA', install: 'manual' , e2e: 'manual', nodepool_avz: 'eu-de-1d', oidc: 'd7df5ce6c37643e49b3c93528c29818b' },\n+ \"eu-de-2\": { continent: 'EMEA', install: 'manual' , e2e: 'manual', nodepool_avz: 'eu-de-2b', oidc: '57da32d8d50d49a9a9fbbe6b890e9e13' },\n+ \"eu-nl-1\": { continent: 'EMEA', install: 'manual' , e2e: 'terraform', nodepool_avz: 'eu-nl-1a', oidc: 'd9b056a6762649f684a8a02b30f38324' },\n+ \"eu-ru-1\": { continent: 'EMEA', install: 'terraform', e2e: 'terraform', nodepool_avz: 'eu-ru-1a', oidc: 'eb31733cfb9d4ab283bb19e00863b780' },\n+ \"la-br-1\": { continent: 'AMERICAS', install: 'terraform', e2e: 'terraform', nodepool_avz: 'la-br-1b', oidc: 'b33ac00cb0004d50bb34eb4e4cbbeb54' },\n+ \"na-ca-1\": { continent: 'AMERICAS', install: 'terraform', e2e: 'terraform', nodepool_avz: 'na-ca-1b', oidc: 'b7269ea1dde54a6383453ab69f25cce4' },\n\"na-us-1\": { continent: 'AMERICAS', install: 'manual' , e2e: 'manual', nodepool_avz: 'na-us-1b', oidc: '040ed45cb5d8403790cf341a26c50601' },\n- \"na-us-2\": { continent: 'AMERICAS', install: 'terraform', e2e: 'terraform', oidc: 'fc0f2b8450e1444197609800c8b5c4cb' },\n- \"na-us-3\": { continent: 'AMERICAS', install: 'terraform', e2e: 'terraform', oidc: 'aa579369ca944a41bccb76f49f20924f' },\n+ \"na-us-2\": { continent: 'AMERICAS', install: 'terraform', e2e: 'terraform', nodepool_avz: 'na-us-2b', oidc: 'fc0f2b8450e1444197609800c8b5c4cb' },\n+ \"na-us-3\": { continent: 'AMERICAS', install: 'terraform', e2e: 'terraform', nodepool_avz: 'na-us-3a', oidc: 'aa579369ca944a41bccb76f49f20924f' },\n\"qa-de-1\": { continent: 'INFRA', install: 'terraform', e2e: 'terraform', nodepool_avz: 'qa-de-1a', oidc: 'ce5462b173184391a9b47b486fabe580'},\n- \"master\": { continent: \"INFRA\", install: 'virtual' , e2e: 'virtual', oidc: 'a8b6853024494e1eb56136025b04866b' }\n+ \"master\": { continent: \"INFRA\", install: 'virtual' , e2e: 'virtual', nodepool_avz: 'eu-nl-1a', oidc: 'a8b6853024494e1eb56136025b04866b' }\n}\nGROUPS = REGIONS.values.map{ |v| v[:continent]}.uniq\n@@ -127,7 +127,7 @@ resources:\nsource:\nusername: ((keppel.username))\npassword: ((keppel.password))\n- repository: keppel.eu-de-2.cloud.sap/ccloud/kubernikus-docs-builder\n+ repository: keppel.eu-de-1.cloud.sap/ccloud/kubernikus-docs-builder\ntag: latest\n- name: kubernikus.image\n@@ -137,7 +137,7 @@ resources:\nsource:\nusername: ((keppel.username))\npassword: ((keppel.password))\n- repository: keppel.eu-de-2.cloud.sap/ccloud/kubernikus\n+ repository: keppel.eu-de-1.cloud.sap/ccloud/kubernikus\ntag: latest\n- name: kubernikusctl.release\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
Add AVZ to e2e tests
|
596,226 |
05.01.2022 16:40:29
| -3,600 |
99ca14e8e939815e79367606aac12858d32175d7
|
Use ap-cn-1b in e2e test
|
[
{
"change_type": "MODIFY",
"old_path": "ci/pipeline.yaml",
"new_path": "ci/pipeline.yaml",
"diff": "@@ -88,7 +88,7 @@ auth_e2e_ap-cn-1: &auth_e2e_ap-cn-1\nKUBERNIKUS_NAME: e2e\nOS_PROJECT_DOMAIN_NAME: ccadmin\n- NODEPOOL_AVZ: ap-cn-1a\n+ NODEPOOL_AVZ: ap-cn-1b\nOS_AUTH_URL: https://identity-3.ap-cn-1.cloud.sap/v3\n"
},
{
"change_type": "MODIFY",
"old_path": "ci/pipeline.yaml.erb",
"new_path": "ci/pipeline.yaml.erb",
"diff": "@@ -16,7 +16,7 @@ VERSIONS = [\nREGIONS = {\n\"ap-ae-1\": { continent: 'ASIA', install: 'terraform', e2e: 'terraform', nodepool_avz: 'ap-ae-1a', oidc: '9a2d0e1181194fb1bca52e7951666bc3'},\n\"ap-au-1\": { continent: 'ASIA', install: 'manual' , e2e: 'manual', nodepool_avz: 'ap-au-1a', oidc: 'f34304c03d2147cc964470c31d00ce97' },\n- \"ap-cn-1\": { continent: 'ASIA', install: 'terraform', e2e: 'terraform', nodepool_avz: 'ap-cn-1a', oidc: 'e83ddace725b4c5da605e3f66d266f2c' },\n+ \"ap-cn-1\": { continent: 'ASIA', install: 'terraform', e2e: 'terraform', nodepool_avz: 'ap-cn-1b', oidc: 'e83ddace725b4c5da605e3f66d266f2c' },\n\"ap-jp-1\": { continent: 'ASIA', install: 'terraform', e2e: 'terraform', nodepool_avz: 'ap-jp-1a', oidc: 'b505497ca03e4942a2d0aa50dfbdc7a0' },\n\"ap-jp-2\": { continent: 'ASIA', install: 'terraform', e2e: 'terraform', nodepool_avz: 'ap-jp-2a', oidc: 'f34b8aa879334a7fb20e1b96cb1c576e' },\n\"ap-sa-1\": { continent: 'ASIA', install: 'terraform', e2e: 'terraform', nodepool_avz: 'ap-sa-1a', oidc: '436297dc67b9428f97da11227dc070ec' },\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
Use ap-cn-1b in e2e test
|
596,226 |
10.01.2022 11:35:04
| -3,600 |
9bb576b2e9aa1425e67520e3bf1361c05d00d969
|
Wait for fs resize in cgroupsv1 reboot fix
|
[
{
"change_type": "MODIFY",
"old_path": "pkg/templates/node_1.10.go",
"new_path": "pkg/templates/node_1.10.go",
"diff": "@@ -23,6 +23,7 @@ systemd:\n[Unit]\nDescription=Reboot if legacy cgroups are not enabled yet\nFailureAction=reboot\n+ After=extend-filesystems.service\n[Service]\nType=simple\nExecStart=/usr/bin/grep 'systemd.unified_cgroup_hierarchy=0' /proc/cmdline\n"
},
{
"change_type": "MODIFY",
"old_path": "pkg/templates/node_1.11.go",
"new_path": "pkg/templates/node_1.11.go",
"diff": "@@ -15,7 +15,6 @@ passwd:\ngroups:\n- name: rkt\nsystem: true\n-\nsystemd:\nunits:\n- name: legacy-cgroup-reboot.service\n@@ -24,6 +23,7 @@ systemd:\n[Unit]\nDescription=Reboot if legacy cgroups are not enabled yet\nFailureAction=reboot\n+ After=extend-filesystems.service\n[Service]\nType=simple\nExecStart=/usr/bin/grep 'systemd.unified_cgroup_hierarchy=0' /proc/cmdline\n"
},
{
"change_type": "MODIFY",
"old_path": "pkg/templates/node_1.12.go",
"new_path": "pkg/templates/node_1.12.go",
"diff": "@@ -23,6 +23,7 @@ systemd:\n[Unit]\nDescription=Reboot if legacy cgroups are not enabled yet\nFailureAction=reboot\n+ After=extend-filesystems.service\n[Service]\nType=simple\nExecStart=/usr/bin/grep 'systemd.unified_cgroup_hierarchy=0' /proc/cmdline\n"
},
{
"change_type": "MODIFY",
"old_path": "pkg/templates/node_1.14.go",
"new_path": "pkg/templates/node_1.14.go",
"diff": "@@ -23,6 +23,7 @@ systemd:\n[Unit]\nDescription=Reboot if legacy cgroups are not enabled yet\nFailureAction=reboot\n+ After=extend-filesystems.service\n[Service]\nType=simple\nExecStart=/usr/bin/grep 'systemd.unified_cgroup_hierarchy=0' /proc/cmdline\n"
},
{
"change_type": "MODIFY",
"old_path": "pkg/templates/node_1.17.go",
"new_path": "pkg/templates/node_1.17.go",
"diff": "@@ -23,6 +23,7 @@ systemd:\n[Unit]\nDescription=Reboot if legacy cgroups are not enabled yet\nFailureAction=reboot\n+ After=extend-filesystems.service\n[Service]\nType=simple\nExecStart=/usr/bin/grep 'systemd.unified_cgroup_hierarchy=0' /proc/cmdline\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
Wait for fs resize in cgroupsv1 reboot fix
|
596,242 |
20.01.2022 07:43:04
| -3,600 |
447fdfdf2647a4c29df07ff6f4f6d6c3526c0f53
|
Add new lines to kluster prints
`kubernikusctl get cluster` results in only one line, because new lines are not added
|
[
{
"change_type": "MODIFY",
"old_path": "pkg/api/models/kluster_print.go",
"new_path": "pkg/api/models/kluster_print.go",
"diff": "@@ -50,11 +50,11 @@ func (k *Kluster) printTable(options printers.PrintOptions) {\nif options.WithHeaders {\nfmt.Print(\"NAME\")\nfmt.Print(\"\\t\")\n- fmt.Print(\"STATUS\")\n+ fmt.Println(\"STATUS\")\n}\nfmt.Print(k.Name)\nfmt.Print(\"\\t\")\n- fmt.Print(k.Status.Phase)\n+ fmt.Println(k.Status.Phase)\n}\nfunc (p NodePool) GetFormats() map[printers.PrintFormat]struct{} {\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
Add new lines to kluster prints (#634)
`kubernikusctl get cluster` results in only one line, because new lines are not added
|
596,227 |
20.01.2022 13:41:37
| -3,600 |
37a843122535386b93d33c67461b12b13413fa01
|
Add stdout output for audit logs
|
[
{
"change_type": "MODIFY",
"old_path": "charts/kube-master/templates/configmap.yaml",
"new_path": "charts/kube-master/templates/configmap.yaml",
"diff": "@@ -215,4 +215,19 @@ data:\n</buffer>\n</match>\n{{- end}}\n+\n+ {{- if eq .Values.audit \"stdout\" }}\n+ <match apiserver.audit>\n+ @type stdout\n+ <buffer>\n+ @type memory\n+ flush_at_shutdown true\n+ flush_mode immediate\n+ chunk_limit_size 8m\n+ </buffer>\n+ <format>\n+ @type json\n+ </format>\n+ </match>\n+ {{- end}}\n{{- end }}\n"
},
{
"change_type": "MODIFY",
"old_path": "pkg/api/models/kluster_spec.go",
"new_path": "pkg/api/models/kluster_spec.go",
"diff": "@@ -27,7 +27,7 @@ type KlusterSpec struct {\nAdvertisePort int64 `json:\"advertisePort\"`\n// audit\n- // Enum: [elasticsearch swift http]\n+ // Enum: [elasticsearch swift http stdout]\nAudit *string `json:\"audit,omitempty\"`\n// backup\n@@ -127,7 +127,7 @@ var klusterSpecTypeAuditPropEnum []interface{}\nfunc init() {\nvar res []string\n- if err := json.Unmarshal([]byte(`[\"elasticsearch\",\"swift\",\"http\"]`), &res); err != nil {\n+ if err := json.Unmarshal([]byte(`[\"elasticsearch\",\"swift\",\"http\",\"stdout\"]`), &res); err != nil {\npanic(err)\n}\nfor _, v := range res {\n@@ -145,6 +145,9 @@ const (\n// KlusterSpecAuditHTTP captures enum value \"http\"\nKlusterSpecAuditHTTP string = \"http\"\n+\n+ // KlusterSpecAuditStdout captures enum value \"stdout\"\n+ KlusterSpecAuditStdout string = \"stdout\"\n)\n// prop value enum\n"
},
{
"change_type": "MODIFY",
"old_path": "pkg/api/spec/embedded_spec.go",
"new_path": "pkg/api/spec/embedded_spec.go",
"diff": "@@ -599,7 +599,8 @@ func init() {\n\"enum\": [\n\"elasticsearch\",\n\"swift\",\n- \"http\"\n+ \"http\",\n+ \"stdout\"\n],\n\"x-nullable\": true\n},\n@@ -1646,7 +1647,8 @@ func init() {\n\"enum\": [\n\"elasticsearch\",\n\"swift\",\n- \"http\"\n+ \"http\",\n+ \"stdout\"\n],\n\"x-nullable\": true\n},\n"
},
{
"change_type": "MODIFY",
"old_path": "swagger.yml",
"new_path": "swagger.yml",
"diff": "@@ -470,7 +470,7 @@ definitions:\naudit:\ntype: string\nx-nullable: true\n- enum: [\"elasticsearch\", \"swift\", \"http\"]\n+ enum: [\"elasticsearch\", \"swift\", \"http\", \"stdout\"]\ndefault: null\nnoCloud:\ntype: boolean\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
Add stdout output for audit logs (#635)
Co-authored-by: Fabian Ruff <fabian.ruff@sap.com>
|
596,226 |
25.01.2022 13:51:43
| -3,600 |
1f30bf02daa7cbee0b2b24012d578992ec2ed898
|
Bump csi liveness probe
|
[
{
"change_type": "MODIFY",
"old_path": "charts/images.yaml",
"new_path": "charts/images.yaml",
"diff": "@@ -209,7 +209,7 @@ imagesForVersion:\ntag: 'v0.5.1'\ncsiLivenessProbe:\nrepository: 'keppel.$REGION.cloud.sap/ccloud/csi-livenessprobe'\n- tag: 'v2.1.0'\n+ tag: 'v2.4.0'\ncsiNodeDriver:\nrepository: 'keppel.$REGION.cloud.sap/ccloud/csi-node-driver-registrar'\ntag: 'v1.3.0'\n@@ -283,7 +283,7 @@ imagesForVersion:\ntag: 'v0.5.1'\ncsiLivenessProbe:\nrepository: 'keppel.$REGION.cloud.sap/ccloud/csi-livenessprobe'\n- tag: 'v2.1.0'\n+ tag: 'v2.4.0'\ncsiNodeDriver:\nrepository: 'keppel.$REGION.cloud.sap/ccloud/csi-node-driver-registrar'\ntag: 'v1.3.0'\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
Bump csi liveness probe
|
596,227 |
26.01.2022 10:34:30
| -3,600 |
2f7d08efd10dbd438a8ee1f6cacd1d52330a4e75
|
Fix audit log file permission
Kubernetes 1.22 changed the default audit log file permissions to 600,
which causes problems with the non-root fluentd sidecar. The log file is
now created in an init container with the correct permissions.
|
[
{
"change_type": "MODIFY",
"old_path": "charts/kube-master/templates/api.yaml",
"new_path": "charts/kube-master/templates/api.yaml",
"diff": "@@ -149,6 +149,18 @@ spec:\nreadOnly: true\n{{- else }}\n- until etcdctl --total-timeout=4s --endpoints http://{{ include \"etcd.fullname\" . }}:2379 cluster-health; do sleep 5; done;\n+{{- end }}\n+{{- if and (semverCompare \">= 1.22\" .Values.version.kubernetes) .Values.audit }}\n+ - name: auditlog-permission-fix\n+ image: \"{{ include \"fluentd.image\" . }}\"\n+ command:\n+ - sh\n+ - -c\n+ args:\n+ - touch /var/log/audit.log && chmod 644 /var/log/audit.log\n+ volumeMounts:\n+ - mountPath: /var/log\n+ name: logs\n{{- end}}\ncontainers:\n- name: apiserver\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
Fix audit log file permission (#636)
Kubernetes 1.22 changed the default audit log file permissions to 600,
which causes problems with the non-root fluentd sidecar. The log file is
now created in an init container with the correct permissions.
|
596,227 |
26.01.2022 14:44:34
| -3,600 |
66565e8df107121564a4d8e730af04f116a494fc
|
Suppress fluentd config dump
|
[
{
"change_type": "MODIFY",
"old_path": "charts/kube-master/templates/api.yaml",
"new_path": "charts/kube-master/templates/api.yaml",
"diff": "@@ -361,6 +361,7 @@ spec:\nresources:\nrequests:\nephemeral-storage: \"2Gi\"\n+ args: [\"--suppress-config-dump\"]\nenv:\n{{- if eq .Values.audit \"elasticsearch\" }}\n- name: ELK_USERNAME\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
Suppress fluentd config dump
|
596,226 |
10.02.2022 11:10:03
| -3,600 |
1bfa11efff05161d34186cc4991dc846a03e9362
|
Fix chart lock file
|
[
{
"change_type": "MODIFY",
"old_path": "charts/kubernikus/requirements.lock",
"new_path": "charts/kubernikus/requirements.lock",
"diff": "@@ -5,5 +5,5 @@ dependencies:\n- name: rbac\nrepository: file://vendor/rbac\nversion: 0.1.4\n-digest: sha256:a4510ed53ee02c40005d48e96caa7d87e8e53fc855b482bdbf7933e7a0fdd5e4\n-generated: \"2022-01-26T14:51:06.206173+02:00\"\n+digest: sha256:a289ead7bbcf7deecda78b9b2fb4124b9021df4355dcf80686baefee13f88764\n+generated: 2022-02-10T11:09:36.29028573+01:00\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
Fix chart lock file
|
596,227 |
14.02.2022 10:00:28
| -3,600 |
1a239c4d2f5ab3f0de70c3996f399539d6aeb68c
|
Adjust audit-policy.yaml
Previously, maintenance-controller leases were not drop due to the
missing "coordination.k8s.io" group. Also role and rolebinding
request bodies need to be logged.
|
[
{
"change_type": "MODIFY",
"old_path": "charts/kube-master/templates/configmap.yaml",
"new_path": "charts/kube-master/templates/configmap.yaml",
"diff": "@@ -94,7 +94,9 @@ data:\nresources:\n- group: \"\"\nresources: [\"secrets\", \"serviceaccounts\"]\n- - level: Metadata\n+ - group: \"\"\n+ resources: [\"pods/log\"]\n+ - level: Request\nresources:\n- group: \"rbac.authorization.k8s.io\"\nresources: [\"clusterroles\", \"clusterrolebindings\", \"roles\", \"rolebindings\"]\n@@ -114,7 +116,10 @@ data:\n- level: None\nresources:\n- group: \"\"\n- resources: [\"configmaps\", \"leases\"]\n+ resources: [\"configmaps\"]\n+ resourceNames: [\"maintenance-controller-leader-election.cloud.sap\"]\n+ - group: \"coordination.k8s.io\"\n+ resources: [\"leases\"]\nresourceNames: [\"maintenance-controller-leader-election.cloud.sap\"]\n- level: None\nnonResourceURLs:\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
Adjust audit-policy.yaml (#638)
Previously, maintenance-controller leases were not drop due to the
missing "coordination.k8s.io" group. Also role and rolebinding
request bodies need to be logged.
|
596,228 |
14.03.2022 15:01:51
| -3,600 |
1f403f19956fad8a08458aafaccc50d25e792686
|
add support for OIDC authentication with kubernikusctl
|
[
{
"change_type": "MODIFY",
"old_path": "pkg/cmd/kubernikusctl/auth/init.go",
"new_path": "pkg/cmd/kubernikusctl/auth/init.go",
"diff": "@@ -23,6 +23,7 @@ type InitOptions struct {\nurl *url.URL\nname string\nkubeconfigPath string\n+ authType string\nopenstack *common.OpenstackClient\nkubernikus *common.KubernikusClient\n@@ -58,6 +59,7 @@ func (o *InitOptions) BindFlags(flags *pflag.FlagSet) {\nflags.StringVar(&o._url, \"url\", o._url, \"URL for Kubernikus API\")\nflags.StringVar(&o.name, \"name\", o.name, \"Cluster Name\")\nflags.StringVar(&o.kubeconfigPath, \"kubeconfig\", o.kubeconfigPath, \"Overwrites kubeconfig auto-detection with explicit path\")\n+ flags.StringVar(&o.authType, \"auth-type\", o.authType, \"Authentication type\")\n}\nfunc (o *InitOptions) Validate(c *cobra.Command, args []string) (err error) {\n@@ -108,10 +110,17 @@ func (o *InitOptions) Run(c *cobra.Command) (err error) {\n}\n}\n+ var kubeconfig string\n+ var errCredentials error\n+ if o.authType == \"oidc\" {\n+ fmt.Printf(\"Fetching OIDC credentials for %v from %v\\n\", o.name, o.url)\n+ kubeconfig, errCredentials = o.kubernikus.GetCredentialsOIDC(o.name)\n+ } else {\nfmt.Printf(\"Fetching credentials for %v from %v\\n\", o.name, o.url)\n- kubeconfig, err := o.kubernikus.GetCredentials(o.name)\n- if err != nil {\n- return errors.Wrap(err, \"Couldn't fetch credentials from Kubernikus API\")\n+ kubeconfig, errCredentials = o.kubernikus.GetCredentials(o.name)\n+ }\n+ if errCredentials != nil {\n+ return errors.Wrap(errCredentials, \"Couldn't fetch credentials from Kubernikus API\")\n}\nif storePasswordInKeyRing {\n"
},
{
"change_type": "MODIFY",
"old_path": "pkg/cmd/kubernikusctl/common/kubernikus.go",
"new_path": "pkg/cmd/kubernikusctl/common/kubernikus.go",
"diff": "@@ -56,6 +56,25 @@ func (k *KubernikusClient) GetCredentials(name string) (string, error) {\nreturn ok.Payload.Kubeconfig, nil\n}\n+func (k *KubernikusClient) GetCredentialsOIDC(name string) (string, error) {\n+ ok, err := k.client.Operations.GetClusterCredentialsOIDC(\n+ operations.NewGetClusterCredentialsOIDCParams().WithName(name),\n+ k.authFunc())\n+\n+ switch err.(type) {\n+ case *operations.GetClusterCredentialsOIDCDefault:\n+ result := err.(*operations.GetClusterCredentialsOIDCDefault)\n+ if result.Code() == 404 {\n+ return \"\", errors.Errorf(\"Cluster %v not found\", name)\n+ }\n+ return \"\", errors.Errorf(result.Payload.Message)\n+ case error:\n+ return \"\", errors.Wrapf(err, \"A generic error occurred\")\n+ }\n+\n+ return ok.Payload.Kubeconfig, nil\n+}\n+\nfunc (k *KubernikusClient) CreateCluster(cluster *models.Kluster) error {\nparams := operations.NewCreateClusterParams().WithBody(cluster)\n_, err := k.client.Operations.CreateCluster(params, k.authFunc())\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
add support for OIDC authentication with kubernikusctl (#627)
|
596,227 |
24.03.2022 13:09:38
| -3,600 |
da4f333b99a7439949c786e7dac31b86fc580815
|
Fix issue in kube-master chart
Corrects some small issues, which wil cause issue with a future Helm3
migration
|
[
{
"change_type": "MODIFY",
"old_path": "charts/kube-master/Chart.yaml",
"new_path": "charts/kube-master/Chart.yaml",
"diff": "apiVersion: v1\ndescription: A Helm chart for Kubernetes\nname: kube-master\n-version: 2.0.6\n+version: 2.0.7\n"
},
{
"change_type": "MODIFY",
"old_path": "charts/kube-master/templates/dex.yaml",
"new_path": "charts/kube-master/templates/dex.yaml",
"diff": "@@ -138,9 +138,9 @@ spec:\n{{- end }}\nvolumes:\n- name: config\n- defaultMode: 420\nconfigMap:\nname: {{ include \"master.fullname\" . }}-dex\n+ defaultMode: 0420\nitems:\n- key: config.yaml\npath: config.yaml\n"
},
{
"change_type": "MODIFY",
"old_path": "charts/kube-master/values.yaml",
"new_path": "charts/kube-master/values.yaml",
"diff": "@@ -6,7 +6,7 @@ image:\npullPolicy: IfNotPresent\n# Settings for the openstack cloudprovider\n-openstack:\n+openstack: {}\n#authURL:\n#username:\n#password:\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
Fix issue in kube-master chart (#647)
Corrects some small issues, which wil cause issue with a future Helm3
migration
|
596,226 |
05.05.2022 12:30:41
| -7,200 |
115e480b52accd8f1c940599fb9a0530de20e1fa
|
Add 1.23 to e2e test
|
[
{
"change_type": "MODIFY",
"old_path": "ci/pipeline.yaml",
"new_path": "ci/pipeline.yaml",
"diff": "@@ -1507,7 +1507,7 @@ jobs:\n- put: all-versions.trigger\n- - name: e2e-1.22.4\n+ - name: e2e-1.23.6\nserial: true\nplan:\n- in_parallel:\n@@ -1524,14 +1524,14 @@ jobs:\n<<: *task_e2e_tests\ntimeout: 2h\nparams:\n- KLUSTER_VERSION: \"1.22.4\"\n+ KLUSTER_VERSION: \"1.23.6\"\nKLUSTER_CIDR: \"100.100.0.0/16\"\nISOLATE_TEST: \"true\"\nKLUSTER_OS_IMAGES: flatcar-stable-amd64,flatcar-beta-amd64\n<<: *auth_e2e_master\n<<: *slack\n- - name: e2e-1.21.5\n+ - name: e2e-1.22.4\nserial: true\nplan:\n- in_parallel:\n@@ -1548,14 +1548,14 @@ jobs:\n<<: *task_e2e_tests\ntimeout: 2h\nparams:\n- KLUSTER_VERSION: \"1.21.5\"\n+ KLUSTER_VERSION: \"1.22.4\"\nKLUSTER_CIDR: \"100.101.0.0/16\"\nISOLATE_TEST: \"true\"\nKLUSTER_OS_IMAGES: flatcar-stable-amd64,flatcar-beta-amd64\n<<: *auth_e2e_master\n<<: *slack\n- - name: e2e-1.20.8\n+ - name: e2e-1.21.5\nserial: true\nplan:\n- in_parallel:\n@@ -1572,14 +1572,14 @@ jobs:\n<<: *task_e2e_tests\ntimeout: 2h\nparams:\n- KLUSTER_VERSION: \"1.20.8\"\n+ KLUSTER_VERSION: \"1.21.5\"\nKLUSTER_CIDR: \"100.102.0.0/16\"\nISOLATE_TEST: \"true\"\nKLUSTER_OS_IMAGES: flatcar-stable-amd64,flatcar-beta-amd64\n<<: *auth_e2e_master\n<<: *slack\n- - name: e2e-1.19.11\n+ - name: e2e-1.20.8\nserial: true\nplan:\n- in_parallel:\n@@ -1596,14 +1596,14 @@ jobs:\n<<: *task_e2e_tests\ntimeout: 2h\nparams:\n- KLUSTER_VERSION: \"1.19.11\"\n+ KLUSTER_VERSION: \"1.20.8\"\nKLUSTER_CIDR: \"100.103.0.0/16\"\nISOLATE_TEST: \"true\"\nKLUSTER_OS_IMAGES: flatcar-stable-amd64,flatcar-beta-amd64\n<<: *auth_e2e_master\n<<: *slack\n- - name: e2e-1.18.10\n+ - name: e2e-1.19.11\nserial: true\nplan:\n- in_parallel:\n@@ -1620,14 +1620,14 @@ jobs:\n<<: *task_e2e_tests\ntimeout: 2h\nparams:\n- KLUSTER_VERSION: \"1.18.10\"\n+ KLUSTER_VERSION: \"1.19.11\"\nKLUSTER_CIDR: \"100.104.0.0/16\"\nISOLATE_TEST: \"true\"\nKLUSTER_OS_IMAGES: flatcar-stable-amd64,flatcar-beta-amd64\n<<: *auth_e2e_master\n<<: *slack\n- - name: e2e-1.17.13\n+ - name: e2e-1.18.10\nserial: true\nplan:\n- in_parallel:\n@@ -1644,14 +1644,14 @@ jobs:\n<<: *task_e2e_tests\ntimeout: 2h\nparams:\n- KLUSTER_VERSION: \"1.17.13\"\n+ KLUSTER_VERSION: \"1.18.10\"\nKLUSTER_CIDR: \"100.105.0.0/16\"\nISOLATE_TEST: \"true\"\nKLUSTER_OS_IMAGES: flatcar-stable-amd64,flatcar-beta-amd64\n<<: *auth_e2e_master\n<<: *slack\n- - name: e2e-1.16.14\n+ - name: e2e-1.17.13\nserial: true\nplan:\n- in_parallel:\n@@ -1668,14 +1668,14 @@ jobs:\n<<: *task_e2e_tests\ntimeout: 2h\nparams:\n- KLUSTER_VERSION: \"1.16.14\"\n+ KLUSTER_VERSION: \"1.17.13\"\nKLUSTER_CIDR: \"100.106.0.0/16\"\nISOLATE_TEST: \"true\"\nKLUSTER_OS_IMAGES: flatcar-stable-amd64,flatcar-beta-amd64\n<<: *auth_e2e_master\n<<: *slack\n- - name: e2e-1.15.9\n+ - name: e2e-1.16.14\nserial: true\nplan:\n- in_parallel:\n@@ -1692,14 +1692,14 @@ jobs:\n<<: *task_e2e_tests\ntimeout: 2h\nparams:\n- KLUSTER_VERSION: \"1.15.9\"\n+ KLUSTER_VERSION: \"1.16.14\"\nKLUSTER_CIDR: \"100.107.0.0/16\"\nISOLATE_TEST: \"true\"\nKLUSTER_OS_IMAGES: flatcar-stable-amd64,flatcar-beta-amd64\n<<: *auth_e2e_master\n<<: *slack\n- - name: e2e-1.14.5\n+ - name: e2e-1.15.9\nserial: true\nplan:\n- in_parallel:\n@@ -1716,14 +1716,14 @@ jobs:\n<<: *task_e2e_tests\ntimeout: 2h\nparams:\n- KLUSTER_VERSION: \"1.14.5\"\n+ KLUSTER_VERSION: \"1.15.9\"\nKLUSTER_CIDR: \"100.108.0.0/16\"\nISOLATE_TEST: \"true\"\nKLUSTER_OS_IMAGES: flatcar-stable-amd64,flatcar-beta-amd64\n<<: *auth_e2e_master\n<<: *slack\n- - name: e2e-1.13.9\n+ - name: e2e-1.14.5\nserial: true\nplan:\n- in_parallel:\n@@ -1740,14 +1740,14 @@ jobs:\n<<: *task_e2e_tests\ntimeout: 2h\nparams:\n- KLUSTER_VERSION: \"1.13.9\"\n+ KLUSTER_VERSION: \"1.14.5\"\nKLUSTER_CIDR: \"100.109.0.0/16\"\nISOLATE_TEST: \"true\"\nKLUSTER_OS_IMAGES: flatcar-stable-amd64,flatcar-beta-amd64\n<<: *auth_e2e_master\n<<: *slack\n- - name: e2e-1.12.10\n+ - name: e2e-1.13.9\nserial: true\nplan:\n- in_parallel:\n@@ -1764,7 +1764,7 @@ jobs:\n<<: *task_e2e_tests\ntimeout: 2h\nparams:\n- KLUSTER_VERSION: \"1.12.10\"\n+ KLUSTER_VERSION: \"1.13.9\"\nKLUSTER_CIDR: \"100.110.0.0/16\"\nISOLATE_TEST: \"true\"\nKLUSTER_OS_IMAGES: flatcar-stable-amd64,flatcar-beta-amd64\n@@ -1777,7 +1777,7 @@ jobs:\nplan:\n- get: kubernikus.builds\nresource: kubernikus.git\n- passed: [e2e-1.22.4]\n+ passed: [e2e-1.23.6]\n- task: changes\nconfig:\n<<: *task_github_compare_url\n@@ -3730,6 +3730,8 @@ groups:\n- tiller_master\n- all-versions\n+ - e2e-1.23.6\n+\n- e2e-1.22.4\n- e2e-1.21.5\n@@ -3750,8 +3752,6 @@ groups:\n- e2e-1.13.9\n- - e2e-1.12.10\n-\n- prod\n- ASIA\n@@ -3909,6 +3909,8 @@ groups:\n- tiller_master\n- all-versions\n+ - e2e-1.23.6\n+\n- e2e-1.22.4\n- e2e-1.21.5\n@@ -3929,8 +3931,6 @@ groups:\n- e2e-1.13.9\n- - e2e-1.12.10\n-\n- prod\n- ASIA\n"
},
{
"change_type": "MODIFY",
"old_path": "ci/pipeline.yaml.erb",
"new_path": "ci/pipeline.yaml.erb",
"diff": "<%\nVERSIONS = [\n+ '1.23.6',\n'1.22.4',\n'1.21.5',\n'1.20.8',\n@@ -9,8 +10,7 @@ VERSIONS = [\n'1.16.14',\n'1.15.9',\n'1.14.5',\n- '1.13.9',\n- '1.12.10'\n+ '1.13.9'\n]\nREGIONS = {\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
Add 1.23 to e2e test
|
596,226 |
05.05.2022 13:33:15
| -7,200 |
5d6a362febf8f9a3830d433af629ac7516ec36ff
|
Use 1.23 in regional e2e test
|
[
{
"change_type": "MODIFY",
"old_path": "test/e2e/setup_test.go",
"new_path": "test/e2e/setup_test.go",
"diff": "@@ -40,7 +40,7 @@ func (s *SetupTests) Run(t *testing.T) {\n}\nfunc (s *SetupTests) CreateCluster(t *testing.T) {\n- version := \"1.22.4\"\n+ version := \"1.23.6\"\nif v := os.Getenv(\"KLUSTER_VERSION\"); v != \"\" {\nversion = v\n}\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
Use 1.23 in regional e2e test
|
596,226 |
10.05.2022 10:50:16
| -7,200 |
21dca60dbf99d5bbede29a9a7e99869175ed17b1
|
Fix flatcar release feed url
|
[
{
"change_type": "MODIFY",
"old_path": "test/e2e/node_test.go",
"new_path": "test/e2e/node_test.go",
"diff": "@@ -173,7 +173,7 @@ func (k NodeTests) currentFlatcarVersion(channel string) (string, error) {\n} `json:\"current\"`\n}\n- feed_url := fmt.Sprintf(\"https://kinvolk.io/flatcar-container-linux/releases-json/releases-%s.json\", channel)\n+ feed_url := fmt.Sprintf(\"https://www.flatcar.org/releases-json/releases-%s.json\", channel)\nresp, err := http.Get(feed_url)\nif err != nil {\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
Fix flatcar release feed url
|
596,226 |
20.05.2022 10:01:38
| -7,200 |
7d79688ece0b55838440c5fe9b9a52089a05c672
|
Fix ingress api versions
|
[
{
"change_type": "MODIFY",
"old_path": "charts/kubernikus/Chart.yaml",
"new_path": "charts/kubernikus/Chart.yaml",
"diff": "apiVersion: v1\ndescription: A Helm chart for Kubernetes\nname: kubernikus\n-version: 0.3.8\n+version: 0.3.9\n"
},
{
"change_type": "MODIFY",
"old_path": "charts/kubernikus/templates/dex.yaml",
"new_path": "charts/kubernikus/templates/dex.yaml",
"diff": "@@ -98,12 +98,19 @@ spec:\nhttp:\npaths:\n- path: /\n+{{- if .Capabilities.APIVersions.Has \"networking.k8s.io/v1/Ingress\" }}\npathType: Prefix\n+{{- end }}\nbackend:\n+{{- if .Capabilities.APIVersions.Has \"networking.k8s.io/v1/Ingress\" }}\nservice:\nname: kubernikus-api-dex\nport:\nnumber: 80\n+{{- else }}\n+ serviceName: kubernikus-api-dex\n+ servicePort: 80\n+{{- end }}\ntls:\n- hosts:\n- {{ include \"oidc.issuer\" . | quote }}\n"
},
{
"change_type": "MODIFY",
"old_path": "charts/kubernikus/templates/ingress.yaml",
"new_path": "charts/kubernikus/templates/ingress.yaml",
"diff": "@@ -23,9 +23,16 @@ spec:\nhttp:\npaths:\n- path: /\n+{{- if .Capabilities.APIVersions.Has \"networking.k8s.io/v1/Ingress\" }}\npathType: Prefix\n+{{- end }}\nbackend:\n+{{- if .Capabilities.APIVersions.Has \"networking.k8s.io/v1/Ingress\" }}\nservice:\nname: kubernikus-api\nport:\nnumber: {{ .Values.api.port }}\n+{{- else }}\n+ serviceName: kubernikus-api\n+ servicePort: {{ .Values.api.port }}\n+{{- end }}\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
Fix ingress api versions
|
596,227 |
21.06.2022 14:37:32
| -7,200 |
caee272332bc14f44c1f3f6d34f483c7c1966f0a
|
fix images.yaml path in helm3 migration
|
[
{
"change_type": "MODIFY",
"old_path": "pkg/migration/21_helm_2to3.go",
"new_path": "pkg/migration/21_helm_2to3.go",
"diff": "@@ -2,7 +2,9 @@ package migration\nimport (\n\"context\"\n+ \"errors\"\n\"fmt\"\n+ \"os\"\n\"path\"\n\"strconv\"\n\"strings\"\n@@ -38,7 +40,11 @@ func migrateHelmReleases(kluster *v1.Kluster, clients config.Clients) error {\nif strings.HasPrefix(pullRegion, \"qa-de\") {\npullRegion = \"eu-de-1\"\n}\n- imageRegistry, err := version.NewImageRegistry(path.Join(\"charts\", \"images.yaml\"), pullRegion)\n+ chartsPath := path.Join(\"charts\", \"images.yaml\")\n+ if _, err := os.Stat(chartsPath); errors.Is(err, os.ErrNotExist) {\n+ chartsPath = \"/etc/kubernikus/charts/images.yaml\"\n+ }\n+ imageRegistry, err := version.NewImageRegistry(chartsPath, pullRegion)\nif err != nil {\nreturn err\n}\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
fix images.yaml path in helm3 migration (#663)
|
596,227 |
21.06.2022 15:47:26
| -7,200 |
d6d0318b8fa08510c2dc28e52a1971c1336b6f17
|
Don't create releases for all old version
|
[
{
"change_type": "MODIFY",
"old_path": "pkg/migration/21_helm_2to3.go",
"new_path": "pkg/migration/21_helm_2to3.go",
"diff": "@@ -6,6 +6,7 @@ import (\n\"fmt\"\n\"os\"\n\"path\"\n+ \"sort\"\n\"strconv\"\n\"strings\"\n@@ -54,10 +55,11 @@ func migrateHelmReleases(kluster *v1.Kluster, clients config.Clients) error {\nif err != nil {\nreturn err\n}\n+ sort.Ints(versions2)\nclient2 := clients.Helm\nclient3 := clients.Helm3\n- for _, version2 := range versions2 {\n- rsp, err := client2.ReleaseContent(kluster.Name, helm.ContentReleaseVersion(int32(version2)))\n+ latestVersion2 := versions2[len(versions2)-1]\n+ rsp, err := client2.ReleaseContent(kluster.Name, helm.ContentReleaseVersion(int32(latestVersion2)))\nif err != nil {\nreturn err\n}\n@@ -78,7 +80,6 @@ func migrateHelmReleases(kluster *v1.Kluster, clients config.Clients) error {\nif err != nil {\nreturn err\n}\n- }\nreturn nil\n}\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
Don't create releases for all old version (#664)
|
596,227 |
21.06.2022 17:18:46
| -7,200 |
c294286fa2662e18f478b72d95e64ed6807e71be
|
Skip Helm3 migration if Helm3 release exists
|
[
{
"change_type": "MODIFY",
"old_path": "pkg/migration/21_helm_2to3.go",
"new_path": "pkg/migration/21_helm_2to3.go",
"diff": "@@ -12,6 +12,7 @@ import (\nhelm_2to3 \"github.com/helm/helm-2to3/pkg/v3\"\n\"helm.sh/helm/v3/pkg/action\"\n+ \"helm.sh/helm/v3/pkg/storage/driver\"\nmeta_v1 \"k8s.io/apimachinery/pkg/apis/meta/v1\"\n\"k8s.io/helm/pkg/helm\"\n@@ -67,6 +68,18 @@ func migrateHelmReleases(kluster *v1.Kluster, clients config.Clients) error {\nif err != nil {\nreturn err\n}\n+ _, err = client3.Releases.Last(release3.Name)\n+ switch {\n+ // Helm3 release not found => so migrate\n+ case errors.Is(err, driver.ErrReleaseNotFound):\n+ break\n+ // Return any other error\n+ case err != nil:\n+ return err\n+ // No error => Helm3 release exists => do nothing\n+ case err == nil:\n+ return nil\n+ }\nerr = client3.Releases.Create(release3)\nif err != nil {\nreturn err\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
Skip Helm3 migration if Helm3 release exists (#665)
|
596,227 |
22.06.2022 13:44:39
| -7,200 |
2d4a653464d420d2168c15eb09354b3aac5de72c
|
Set Helm3 release namespace on install
|
[
{
"change_type": "MODIFY",
"old_path": "pkg/controller/ground.go",
"new_path": "pkg/controller/ground.go",
"diff": "@@ -629,6 +629,7 @@ func (op *GroundControl) createKluster(kluster *v1.Kluster) error {\n}\ninstall := action.NewInstall(op.Helm3)\ninstall.ReleaseName = kluster.GetName()\n+ install.Namespace = kluster.GetNamespace()\n_, err = install.Run(chart, helmValues)\nreturn err\n}\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
Set Helm3 release namespace on install (#666)
|
596,226 |
30.06.2022 11:32:45
| -7,200 |
3a4f0cd3ef1bfb93217127b34e30bdc6ebb3549f
|
Use helm3 in kubernikus/kubernikus-dex
|
[
{
"change_type": "MODIFY",
"old_path": "ci/pipeline.yaml",
"new_path": "ci/pipeline.yaml",
"diff": "@@ -913,6 +913,7 @@ task_helm_kubernikus-dex: &task_helm_kubernikus-dex\nvault-injector interpolate secrets.git/$REGION/values/keystone.yaml secrets.git/$REGION/values/globals.yaml /tmp/kubernikus-dex.yaml\nhelm diff upgrade kubernikus-dex kubernikus-copy.builds/charts/kubernikus-dex/ \\\n+ --namespace=kubernikus \\\n--values secrets.git/$REGION/values/keystone.yaml \\\n--values secrets.git/$REGION/values/globals.yaml \\\n--values /tmp/kubernikus-dex.yaml \\\n@@ -932,7 +933,7 @@ task_helm_kubernikus-dex: &task_helm_kubernikus-dex\nKUBELOGON_PASSWORD:\nKUBERNIKUS_NAME:\nKUBERNIKUS_URL:\n- HELM_VERSION: 2.11.0\n+ HELM_VERSION: 3.2.4\nVAULT_ADDR: https://vault.global.cloud.sap # DO NOT CHANGE\nVAULT_KV_ENGINE: secrets # DO NOT CHANGE\nVAULT_ROLE_ID: # (required) set this to ((auth.role_id)) to receive credentials automatically from Concourse\n@@ -970,6 +971,7 @@ task_helm_kubernikus: &task_helm_kubernikus\nvault-injector interpolate secrets.git/global/values/kubernikus.yaml secrets.git/kubernikus/$KUBERNIKUS_NAME/values/kubernikus.yaml\nhelm diff upgrade kubernikus kubernikus.builds/charts/kubernikus/ \\\n+ --namespace=kubernikus-system \\\n--suppress-secrets --allow-unreleased --context 5 \\\n--values secrets.git/global/values/kubernikus.yaml \\\n--values secrets.git/kubernikus/$KUBERNIKUS_NAME/values/kubernikus.yaml \\\n@@ -994,7 +996,7 @@ task_helm_kubernikus: &task_helm_kubernikus\nKUBELOGON_PASSWORD:\nKUBERNIKUS_NAME:\nKUBERNIKUS_URL:\n- HELM_VERSION: 2.11.0\n+ HELM_VERSION: 3.2.4\nVAULT_ADDR: https://vault.global.cloud.sap # DO NOT CHANGE\nVAULT_KV_ENGINE: secrets # DO NOT CHANGE\nVAULT_ROLE_ID: # (required) set this to ((auth.role_id)) to receive credentials automatically from Concourse\n"
},
{
"change_type": "MODIFY",
"old_path": "ci/task_helm_kubernikus-dex.yaml",
"new_path": "ci/task_helm_kubernikus-dex.yaml",
"diff": "@@ -33,6 +33,7 @@ run:\nvault-injector interpolate secrets.git/$REGION/values/keystone.yaml secrets.git/$REGION/values/globals.yaml /tmp/kubernikus-dex.yaml\nhelm diff upgrade kubernikus-dex kubernikus-copy.builds/charts/kubernikus-dex/ \\\n+ --namespace=kubernikus \\\n--values secrets.git/$REGION/values/keystone.yaml \\\n--values secrets.git/$REGION/values/globals.yaml \\\n--values /tmp/kubernikus-dex.yaml \\\n@@ -52,7 +53,7 @@ params:\nKUBELOGON_PASSWORD:\nKUBERNIKUS_NAME:\nKUBERNIKUS_URL:\n- HELM_VERSION: 2.11.0\n+ HELM_VERSION: 3.2.4\nVAULT_ADDR: https://vault.global.cloud.sap # DO NOT CHANGE\nVAULT_KV_ENGINE: secrets # DO NOT CHANGE\nVAULT_ROLE_ID: # (required) set this to ((auth.role_id)) to receive credentials automatically from Concourse\n"
},
{
"change_type": "MODIFY",
"old_path": "ci/task_helm_kubernikus.yaml",
"new_path": "ci/task_helm_kubernikus.yaml",
"diff": "@@ -28,6 +28,7 @@ run:\nvault-injector interpolate secrets.git/global/values/kubernikus.yaml secrets.git/kubernikus/$KUBERNIKUS_NAME/values/kubernikus.yaml\nhelm diff upgrade kubernikus kubernikus.builds/charts/kubernikus/ \\\n+ --namespace=kubernikus-system \\\n--suppress-secrets --allow-unreleased --context 5 \\\n--values secrets.git/global/values/kubernikus.yaml \\\n--values secrets.git/kubernikus/$KUBERNIKUS_NAME/values/kubernikus.yaml \\\n@@ -52,7 +53,7 @@ params:\nKUBELOGON_PASSWORD:\nKUBERNIKUS_NAME:\nKUBERNIKUS_URL:\n- HELM_VERSION: 2.11.0\n+ HELM_VERSION: 3.2.4\nVAULT_ADDR: https://vault.global.cloud.sap # DO NOT CHANGE\nVAULT_KV_ENGINE: secrets # DO NOT CHANGE\nVAULT_ROLE_ID: # (required) set this to ((auth.role_id)) to receive credentials automatically from Concourse\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
Use helm3 in kubernikus/kubernikus-dex
|
596,227 |
30.06.2022 16:23:25
| -7,200 |
2767287d5d5f8062190150093e755eccff653f74
|
Filter output of kubernikusctl again
|
[
{
"change_type": "MODIFY",
"old_path": "pkg/cmd/kubernikusctl/common/log.go",
"new_path": "pkg/cmd/kubernikusctl/common/log.go",
"diff": "@@ -21,8 +21,7 @@ func BindLogFlags(flags *pflag.FlagSet) {\nfunc SetupLogger() {\nlogger := kitLog.NewLogfmtLogger(kitLog.NewSyncWriter(os.Stderr))\nlogger = log.NewTrailingNilFilter(logger)\n- //logger = log.NewLevelFilter(logLevel, logger)\nlogger = kitLog.With(logger, \"ts\", kitLog.DefaultTimestampUTC, \"caller\", log.Caller(4))\n+ klog.ClampLevel(klog.Level(logLevel))\nklog.SetLogger(logger)\n-\n}\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
Filter output of kubernikusctl again (#651)
Co-authored-by: Fabian Ruff <fabian.ruff@sap.com>
|
596,227 |
05.07.2022 15:48:52
| -7,200 |
a8ca2282e84f43610c0b1f4cade883573fa1c15e
|
Fix Helm3 migration pv access mode
|
[
{
"change_type": "MODIFY",
"old_path": "pkg/migration/21_helm_2to3.go",
"new_path": "pkg/migration/21_helm_2to3.go",
"diff": "@@ -32,7 +32,7 @@ func migrateHelmReleases(kluster *v1.Kluster, clients config.Clients) error {\nif err != nil {\nreturn err\n}\n- accessMode, err := util.PVAccessMode(clients.Kubernetes, nil)\n+ accessMode, err := util.PVAccessMode(clients.Kubernetes, kluster)\nif err != nil {\nreturn err\n}\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
Fix Helm3 migration pv access mode (#670)
|
596,226 |
06.07.2022 10:58:11
| -7,200 |
eaf31ba529a6f490a28bdb891e6b6f84745ec395
|
Remove tiller from pipeline
|
[
{
"change_type": "MODIFY",
"old_path": "ci/pipeline.yaml",
"new_path": "ci/pipeline.yaml",
"diff": "@@ -1074,46 +1074,6 @@ task_oci_build: &task_oci_build\nBUILD_ARGS_FILE:\n-task_tiller: &task_tiller\n- platform: 'linux'\n-\n- image_resource:\n- type: registry-image\n- source:\n- repository: keppel.eu-de-1.cloud.sap/ccloud-dockerhub-mirror/sapcc/unified-kubernetes-toolbox\n- tag: 'latest'\n-\n- inputs:\n- - name: tiller-version.git\n-\n- run:\n- path: /bin/sh\n- args:\n- - -c\n- - |\n- set -exo pipefail\n- tag=$(cat tiller-version.git/kubernikus/tiller.version)\n- if [ -f tiller-version.git/kubernikus/$KUBERNIKUS_NAME/tiller.version ]; then\n- tag=$(cat tiller-version.git/kubernikus/$KUBERNIKUS_NAME/tiller.version )\n- fi\n- if kubectl get deployment tiller-deploy --namespace=kube-system >/dev/null ; then\n- kubectl set image --namespace=kube-system deployment/tiller-deploy tiller=sapcc/tiller:$tag\n- else\n- kubectl create sa tiller --namespace kube-system\n- kubectl create clusterrolebinding tiller --clusterrole cluster-admin --serviceaccount=kube-system:tiller\n- helm init --service-account tiller --history-max 5 --tiller-image sapcc/tiller:$tag --upgrade --wait\n- fi\n-\n- params:\n- REGION:\n- CONTEXT:\n- KUBELOGON_USER:\n- KUBELOGON_PASSWORD:\n- KUBERNIKUS_NAME:\n- KUBERNIKUS_URL:\n- HELM_VERSION: 2.11.0\n-\n-\ntask_whitesource: &task_whitesource\nplatform: 'linux'\n@@ -1284,18 +1244,6 @@ resources:\nbranch: master\ndepth: 1\n- - name: tiller-version.git\n- icon: key-variant\n- type: git-proxy\n- webhook_token: aldhjalkdhahdjkahdjkhjadhjadhkjadlkjhAAdd\n- check_every: 1h\n- source:\n- uri: git@github.wdf.sap.corp:cc/secrets.git\n- private_key: ((github-private-key/cc-secrets))\n- branch: master\n- depth: 1\n- paths: [ kubernikus/tiller.version, kubernikus/*/tiller.version ]\n-\n- name: helm-charts.git\ntype: git\nicon: github-circle\n@@ -1425,21 +1373,6 @@ jobs:\nimage: image/image.tar\nadditional_tags: kubernikus.git/.git/ref\n- - name: tiller_master\n- plan:\n- - in_parallel:\n- - get: tiller-version.git\n- trigger: true\n- - in_parallel:\n- - task: master\n- config:\n- <<: *task_tiller\n- timeout: 10m\n- params:\n- REGION: \"k-master\"\n- <<: *auth_master\n- <<: *slack\n-\n- name: master\nserial: true\nplan:\n@@ -1824,32 +1757,13 @@ jobs:\n<<: *auth_ap-ae-1\n<<: *slack\n- - name: tiller_ap-ae-1\n- plan:\n- - in_parallel:\n- - get: silver.tag\n- trigger: true\n- passed: [seed_ap-ae-1]\n- - get: tiller-version.git\n- trigger: true\n- - in_parallel:\n- - task: k-ap-ae-1\n- config:\n- <<: *task_tiller\n- timeout: 10m\n- params:\n- REGION: ap-ae-1\n- CONTEXT: k-ap-ae-1\n- <<: *auth_ap-ae-1\n- <<: *slack\n-\n- name: deploy_ap-ae-1\nserial: true\nplan:\n- in_parallel:\n- get: silver.tag\ntrigger: true\n- passed: [tiller_ap-ae-1]\n+ passed: [seed_ap-ae-1]\n- get: secrets.git\npassed: [seed_ap-ae-1]\n- get: kubernikus.builds\n@@ -1934,32 +1848,13 @@ jobs:\n<<: *auth_ap-au-1\n<<: *slack\n- - name: tiller_ap-au-1\n- plan:\n- - in_parallel:\n- - get: silver.tag\n- trigger: true\n- passed: [seed_ap-au-1]\n- - get: tiller-version.git\n- trigger: true\n- - in_parallel:\n- - task: k-ap-au-1\n- config:\n- <<: *task_tiller\n- timeout: 10m\n- params:\n- REGION: ap-au-1\n- CONTEXT: k-ap-au-1\n- <<: *auth_ap-au-1\n- <<: *slack\n-\n- name: deploy_ap-au-1\nserial: true\nplan:\n- in_parallel:\n- get: silver.tag\ntrigger: true\n- passed: [tiller_ap-au-1]\n+ passed: [seed_ap-au-1]\n- get: secrets.git\npassed: [seed_ap-au-1]\n- get: kubernikus.builds\n@@ -2044,32 +1939,13 @@ jobs:\n<<: *auth_ap-cn-1\n<<: *slack\n- - name: tiller_ap-cn-1\n- plan:\n- - in_parallel:\n- - get: silver.tag\n- trigger: true\n- passed: [seed_ap-cn-1]\n- - get: tiller-version.git\n- trigger: true\n- - in_parallel:\n- - task: k-ap-cn-1\n- config:\n- <<: *task_tiller\n- timeout: 10m\n- params:\n- REGION: ap-cn-1\n- CONTEXT: k-ap-cn-1\n- <<: *auth_ap-cn-1\n- <<: *slack\n-\n- name: deploy_ap-cn-1\nserial: true\nplan:\n- in_parallel:\n- get: silver.tag\ntrigger: true\n- passed: [tiller_ap-cn-1]\n+ passed: [seed_ap-cn-1]\n- get: secrets.git\npassed: [seed_ap-cn-1]\n- get: kubernikus.builds\n@@ -2154,32 +2030,13 @@ jobs:\n<<: *auth_ap-jp-1\n<<: *slack\n- - name: tiller_ap-jp-1\n- plan:\n- - in_parallel:\n- - get: silver.tag\n- trigger: true\n- passed: [seed_ap-jp-1]\n- - get: tiller-version.git\n- trigger: true\n- - in_parallel:\n- - task: k-ap-jp-1\n- config:\n- <<: *task_tiller\n- timeout: 10m\n- params:\n- REGION: ap-jp-1\n- CONTEXT: k-ap-jp-1\n- <<: *auth_ap-jp-1\n- <<: *slack\n-\n- name: deploy_ap-jp-1\nserial: true\nplan:\n- in_parallel:\n- get: silver.tag\ntrigger: true\n- passed: [tiller_ap-jp-1]\n+ passed: [seed_ap-jp-1]\n- get: secrets.git\npassed: [seed_ap-jp-1]\n- get: kubernikus.builds\n@@ -2264,32 +2121,13 @@ jobs:\n<<: *auth_ap-sa-1\n<<: *slack\n- - name: tiller_ap-sa-1\n- plan:\n- - in_parallel:\n- - get: silver.tag\n- trigger: true\n- passed: [seed_ap-sa-1]\n- - get: tiller-version.git\n- trigger: true\n- - in_parallel:\n- - task: k-ap-sa-1\n- config:\n- <<: *task_tiller\n- timeout: 10m\n- params:\n- REGION: ap-sa-1\n- CONTEXT: k-ap-sa-1\n- <<: *auth_ap-sa-1\n- <<: *slack\n-\n- name: deploy_ap-sa-1\nserial: true\nplan:\n- in_parallel:\n- get: silver.tag\ntrigger: true\n- passed: [tiller_ap-sa-1]\n+ passed: [seed_ap-sa-1]\n- get: secrets.git\npassed: [seed_ap-sa-1]\n- get: kubernikus.builds\n@@ -2374,32 +2212,13 @@ jobs:\n<<: *auth_eu-ru-1\n<<: *slack\n- - name: tiller_eu-ru-1\n- plan:\n- - in_parallel:\n- - get: silver.tag\n- trigger: true\n- passed: [seed_eu-ru-1]\n- - get: tiller-version.git\n- trigger: true\n- - in_parallel:\n- - task: k-eu-ru-1\n- config:\n- <<: *task_tiller\n- timeout: 10m\n- params:\n- REGION: eu-ru-1\n- CONTEXT: k-eu-ru-1\n- <<: *auth_eu-ru-1\n- <<: *slack\n-\n- name: deploy_eu-ru-1\nserial: true\nplan:\n- in_parallel:\n- get: silver.tag\ntrigger: true\n- passed: [tiller_eu-ru-1]\n+ passed: [seed_eu-ru-1]\n- get: secrets.git\npassed: [seed_eu-ru-1]\n- get: kubernikus.builds\n@@ -2484,32 +2303,13 @@ jobs:\n<<: *auth_la-br-1\n<<: *slack\n- - name: tiller_la-br-1\n- plan:\n- - in_parallel:\n- - get: silver.tag\n- trigger: true\n- passed: [seed_la-br-1]\n- - get: tiller-version.git\n- trigger: true\n- - in_parallel:\n- - task: k-la-br-1\n- config:\n- <<: *task_tiller\n- timeout: 10m\n- params:\n- REGION: la-br-1\n- CONTEXT: k-la-br-1\n- <<: *auth_la-br-1\n- <<: *slack\n-\n- name: deploy_la-br-1\nserial: true\nplan:\n- in_parallel:\n- get: silver.tag\ntrigger: true\n- passed: [tiller_la-br-1]\n+ passed: [seed_la-br-1]\n- get: secrets.git\npassed: [seed_la-br-1]\n- get: kubernikus.builds\n@@ -2609,32 +2409,13 @@ jobs:\n<<: *auth_ap-jp-2\n<<: *slack\n- - name: tiller_ap-jp-2\n- plan:\n- - in_parallel:\n- - get: bronze.tag\n- trigger: true\n- passed: [seed_ap-jp-2]\n- - get: tiller-version.git\n- trigger: true\n- - in_parallel:\n- - task: k-ap-jp-2\n- config:\n- <<: *task_tiller\n- timeout: 10m\n- params:\n- REGION: ap-jp-2\n- CONTEXT: k-ap-jp-2\n- <<: *auth_ap-jp-2\n- <<: *slack\n-\n- name: deploy_ap-jp-2\nserial: true\nplan:\n- in_parallel:\n- get: bronze.tag\ntrigger: true\n- passed: [tiller_ap-jp-2]\n+ passed: [seed_ap-jp-2]\n- get: secrets.git\npassed: [seed_ap-jp-2]\n- get: kubernikus.builds\n@@ -2719,32 +2500,13 @@ jobs:\n<<: *auth_ap-sa-2\n<<: *slack\n- - name: tiller_ap-sa-2\n- plan:\n- - in_parallel:\n- - get: bronze.tag\n- trigger: true\n- passed: [seed_ap-sa-2]\n- - get: tiller-version.git\n- trigger: true\n- - in_parallel:\n- - task: k-ap-sa-2\n- config:\n- <<: *task_tiller\n- timeout: 10m\n- params:\n- REGION: ap-sa-2\n- CONTEXT: k-ap-sa-2\n- <<: *auth_ap-sa-2\n- <<: *slack\n-\n- name: deploy_ap-sa-2\nserial: true\nplan:\n- in_parallel:\n- get: bronze.tag\ntrigger: true\n- passed: [tiller_ap-sa-2]\n+ passed: [seed_ap-sa-2]\n- get: secrets.git\npassed: [seed_ap-sa-2]\n- get: kubernikus.builds\n@@ -2829,32 +2591,13 @@ jobs:\n<<: *auth_na-ca-1\n<<: *slack\n- - name: tiller_na-ca-1\n- plan:\n- - in_parallel:\n- - get: bronze.tag\n- trigger: true\n- passed: [seed_na-ca-1]\n- - get: tiller-version.git\n- trigger: true\n- - in_parallel:\n- - task: k-na-ca-1\n- config:\n- <<: *task_tiller\n- timeout: 10m\n- params:\n- REGION: na-ca-1\n- CONTEXT: k-na-ca-1\n- <<: *auth_na-ca-1\n- <<: *slack\n-\n- name: deploy_na-ca-1\nserial: true\nplan:\n- in_parallel:\n- get: bronze.tag\ntrigger: true\n- passed: [tiller_na-ca-1]\n+ passed: [seed_na-ca-1]\n- get: secrets.git\npassed: [seed_na-ca-1]\n- get: kubernikus.builds\n@@ -2939,32 +2682,13 @@ jobs:\n<<: *auth_na-us-3\n<<: *slack\n- - name: tiller_na-us-3\n- plan:\n- - in_parallel:\n- - get: bronze.tag\n- trigger: true\n- passed: [seed_na-us-3]\n- - get: tiller-version.git\n- trigger: true\n- - in_parallel:\n- - task: k-na-us-3\n- config:\n- <<: *task_tiller\n- timeout: 10m\n- params:\n- REGION: na-us-3\n- CONTEXT: k-na-us-3\n- <<: *auth_na-us-3\n- <<: *slack\n-\n- name: deploy_na-us-3\nserial: true\nplan:\n- in_parallel:\n- get: bronze.tag\ntrigger: true\n- passed: [tiller_na-us-3]\n+ passed: [seed_na-us-3]\n- get: secrets.git\npassed: [seed_na-us-3]\n- get: kubernikus.builds\n@@ -3064,32 +2788,13 @@ jobs:\n<<: *auth_eu-de-1\n<<: *slack\n- - name: tiller_eu-de-1\n- plan:\n- - in_parallel:\n- - get: gold.tag\n- trigger: true\n- passed: [seed_eu-de-1]\n- - get: tiller-version.git\n- trigger: true\n- - in_parallel:\n- - task: k-eu-de-1\n- config:\n- <<: *task_tiller\n- timeout: 10m\n- params:\n- REGION: eu-de-1\n- CONTEXT: k-eu-de-1\n- <<: *auth_eu-de-1\n- <<: *slack\n-\n- name: deploy_eu-de-1\nserial: true\nplan:\n- in_parallel:\n- get: gold.tag\ntrigger: true\n- passed: [tiller_eu-de-1]\n+ passed: [seed_eu-de-1]\n- get: secrets.git\npassed: [seed_eu-de-1]\n- get: kubernikus.builds\n@@ -3174,32 +2879,13 @@ jobs:\n<<: *auth_eu-de-2\n<<: *slack\n- - name: tiller_eu-de-2\n- plan:\n- - in_parallel:\n- - get: gold.tag\n- trigger: true\n- passed: [seed_eu-de-2]\n- - get: tiller-version.git\n- trigger: true\n- - in_parallel:\n- - task: k-eu-de-2\n- config:\n- <<: *task_tiller\n- timeout: 10m\n- params:\n- REGION: eu-de-2\n- CONTEXT: k-eu-de-2\n- <<: *auth_eu-de-2\n- <<: *slack\n-\n- name: deploy_eu-de-2\nserial: true\nplan:\n- in_parallel:\n- get: gold.tag\ntrigger: true\n- passed: [tiller_eu-de-2]\n+ passed: [seed_eu-de-2]\n- get: secrets.git\npassed: [seed_eu-de-2]\n- get: kubernikus.builds\n@@ -3284,32 +2970,13 @@ jobs:\n<<: *auth_eu-nl-1\n<<: *slack\n- - name: tiller_eu-nl-1\n- plan:\n- - in_parallel:\n- - get: gold.tag\n- trigger: true\n- passed: [seed_eu-nl-1]\n- - get: tiller-version.git\n- trigger: true\n- - in_parallel:\n- - task: k-eu-nl-1\n- config:\n- <<: *task_tiller\n- timeout: 10m\n- params:\n- REGION: eu-nl-1\n- CONTEXT: k-eu-nl-1\n- <<: *auth_eu-nl-1\n- <<: *slack\n-\n- name: deploy_eu-nl-1\nserial: true\nplan:\n- in_parallel:\n- get: gold.tag\ntrigger: true\n- passed: [tiller_eu-nl-1]\n+ passed: [seed_eu-nl-1]\n- get: secrets.git\npassed: [seed_eu-nl-1]\n- get: kubernikus.builds\n@@ -3394,32 +3061,13 @@ jobs:\n<<: *auth_na-us-1\n<<: *slack\n- - name: tiller_na-us-1\n- plan:\n- - in_parallel:\n- - get: gold.tag\n- trigger: true\n- passed: [seed_na-us-1]\n- - get: tiller-version.git\n- trigger: true\n- - in_parallel:\n- - task: k-na-us-1\n- config:\n- <<: *task_tiller\n- timeout: 10m\n- params:\n- REGION: na-us-1\n- CONTEXT: k-na-us-1\n- <<: *auth_na-us-1\n- <<: *slack\n-\n- name: deploy_na-us-1\nserial: true\nplan:\n- in_parallel:\n- get: gold.tag\ntrigger: true\n- passed: [tiller_na-us-1]\n+ passed: [seed_na-us-1]\n- get: secrets.git\npassed: [seed_na-us-1]\n- get: kubernikus.builds\n@@ -3504,32 +3152,13 @@ jobs:\n<<: *auth_na-us-2\n<<: *slack\n- - name: tiller_na-us-2\n- plan:\n- - in_parallel:\n- - get: gold.tag\n- trigger: true\n- passed: [seed_na-us-2]\n- - get: tiller-version.git\n- trigger: true\n- - in_parallel:\n- - task: k-na-us-2\n- config:\n- <<: *task_tiller\n- timeout: 10m\n- params:\n- REGION: na-us-2\n- CONTEXT: k-na-us-2\n- <<: *auth_na-us-2\n- <<: *slack\n-\n- name: deploy_na-us-2\nserial: true\nplan:\n- in_parallel:\n- get: gold.tag\ntrigger: true\n- passed: [tiller_na-us-2]\n+ passed: [seed_na-us-2]\n- get: secrets.git\npassed: [seed_na-us-2]\n- get: kubernikus.builds\n@@ -3629,32 +3258,13 @@ jobs:\n<<: *auth_qa-de-1\n<<: *slack\n- - name: tiller_qa-de-1\n- plan:\n- - in_parallel:\n- - get: infra.tag\n- trigger: true\n- passed: [seed_qa-de-1]\n- - get: tiller-version.git\n- trigger: true\n- - in_parallel:\n- - task: k-qa-de-1\n- config:\n- <<: *task_tiller\n- timeout: 10m\n- params:\n- REGION: qa-de-1\n- CONTEXT: k-qa-de-1\n- <<: *auth_qa-de-1\n- <<: *slack\n-\n- name: deploy_qa-de-1\nserial: true\nplan:\n- in_parallel:\n- get: infra.tag\ntrigger: true\n- passed: [tiller_qa-de-1]\n+ passed: [seed_qa-de-1]\n- get: secrets.git\npassed: [seed_qa-de-1]\n- get: kubernikus.builds\n@@ -3729,7 +3339,6 @@ groups:\n- master\n- cli\n- whitesource\n- - tiller_master\n- all-versions\n- e2e-1.23.6\n@@ -3770,7 +3379,6 @@ groups:\njobs:\n- silver\n- seed_ap-ae-1\n- - tiller_ap-ae-1\n- deploy_ap-ae-1\n- soak_ap-ae-1\n@@ -3778,7 +3386,6 @@ groups:\njobs:\n- silver\n- seed_ap-au-1\n- - tiller_ap-au-1\n- deploy_ap-au-1\n- soak_ap-au-1\n@@ -3786,7 +3393,6 @@ groups:\njobs:\n- silver\n- seed_ap-cn-1\n- - tiller_ap-cn-1\n- deploy_ap-cn-1\n- soak_ap-cn-1\n@@ -3794,7 +3400,6 @@ groups:\njobs:\n- silver\n- seed_ap-jp-1\n- - tiller_ap-jp-1\n- deploy_ap-jp-1\n- soak_ap-jp-1\n@@ -3802,7 +3407,6 @@ groups:\njobs:\n- bronze\n- seed_ap-jp-2\n- - tiller_ap-jp-2\n- deploy_ap-jp-2\n- soak_ap-jp-2\n@@ -3810,7 +3414,6 @@ groups:\njobs:\n- silver\n- seed_ap-sa-1\n- - tiller_ap-sa-1\n- deploy_ap-sa-1\n- soak_ap-sa-1\n@@ -3818,7 +3421,6 @@ groups:\njobs:\n- bronze\n- seed_ap-sa-2\n- - tiller_ap-sa-2\n- deploy_ap-sa-2\n- soak_ap-sa-2\n@@ -3826,7 +3428,6 @@ groups:\njobs:\n- gold\n- seed_eu-de-1\n- - tiller_eu-de-1\n- deploy_eu-de-1\n- soak_eu-de-1\n@@ -3834,7 +3435,6 @@ groups:\njobs:\n- gold\n- seed_eu-de-2\n- - tiller_eu-de-2\n- deploy_eu-de-2\n- soak_eu-de-2\n@@ -3842,7 +3442,6 @@ groups:\njobs:\n- gold\n- seed_eu-nl-1\n- - tiller_eu-nl-1\n- deploy_eu-nl-1\n- soak_eu-nl-1\n@@ -3850,7 +3449,6 @@ groups:\njobs:\n- silver\n- seed_eu-ru-1\n- - tiller_eu-ru-1\n- deploy_eu-ru-1\n- soak_eu-ru-1\n@@ -3858,7 +3456,6 @@ groups:\njobs:\n- silver\n- seed_la-br-1\n- - tiller_la-br-1\n- deploy_la-br-1\n- soak_la-br-1\n@@ -3866,7 +3463,6 @@ groups:\njobs:\n- bronze\n- seed_na-ca-1\n- - tiller_na-ca-1\n- deploy_na-ca-1\n- soak_na-ca-1\n@@ -3874,7 +3470,6 @@ groups:\njobs:\n- gold\n- seed_na-us-1\n- - tiller_na-us-1\n- deploy_na-us-1\n- soak_na-us-1\n@@ -3882,7 +3477,6 @@ groups:\njobs:\n- gold\n- seed_na-us-2\n- - tiller_na-us-2\n- deploy_na-us-2\n- soak_na-us-2\n@@ -3890,7 +3484,6 @@ groups:\njobs:\n- bronze\n- seed_na-us-3\n- - tiller_na-us-3\n- deploy_na-us-3\n- soak_na-us-3\n@@ -3898,7 +3491,6 @@ groups:\njobs:\n- infra\n- seed_qa-de-1\n- - tiller_qa-de-1\n- deploy_qa-de-1\n- soak_qa-de-1\n@@ -3908,7 +3500,6 @@ groups:\n- master\n- cli\n- whitesource\n- - tiller_master\n- all-versions\n- e2e-1.23.6\n@@ -3945,87 +3536,70 @@ groups:\n- seed_ap-ae-1\n- - tiller_ap-ae-1\n- deploy_ap-ae-1\n- soak_ap-ae-1\n- seed_ap-au-1\n- - tiller_ap-au-1\n- deploy_ap-au-1\n- soak_ap-au-1\n- seed_ap-cn-1\n- - tiller_ap-cn-1\n- deploy_ap-cn-1\n- soak_ap-cn-1\n- seed_ap-jp-1\n- - tiller_ap-jp-1\n- deploy_ap-jp-1\n- soak_ap-jp-1\n- seed_ap-jp-2\n- - tiller_ap-jp-2\n- deploy_ap-jp-2\n- soak_ap-jp-2\n- seed_ap-sa-1\n- - tiller_ap-sa-1\n- deploy_ap-sa-1\n- soak_ap-sa-1\n- seed_ap-sa-2\n- - tiller_ap-sa-2\n- deploy_ap-sa-2\n- soak_ap-sa-2\n- seed_eu-de-1\n- - tiller_eu-de-1\n- deploy_eu-de-1\n- soak_eu-de-1\n- seed_eu-de-2\n- - tiller_eu-de-2\n- deploy_eu-de-2\n- soak_eu-de-2\n- seed_eu-nl-1\n- - tiller_eu-nl-1\n- deploy_eu-nl-1\n- soak_eu-nl-1\n- seed_eu-ru-1\n- - tiller_eu-ru-1\n- deploy_eu-ru-1\n- soak_eu-ru-1\n- seed_la-br-1\n- - tiller_la-br-1\n- deploy_la-br-1\n- soak_la-br-1\n- seed_na-ca-1\n- - tiller_na-ca-1\n- deploy_na-ca-1\n- soak_na-ca-1\n- seed_na-us-1\n- - tiller_na-us-1\n- deploy_na-us-1\n- soak_na-us-1\n- seed_na-us-2\n- - tiller_na-us-2\n- deploy_na-us-2\n- soak_na-us-2\n- seed_na-us-3\n- - tiller_na-us-3\n- deploy_na-us-3\n- soak_na-us-3\n- seed_qa-de-1\n- - tiller_qa-de-1\n- deploy_qa-de-1\n- soak_qa-de-1\n"
},
{
"change_type": "MODIFY",
"old_path": "ci/pipeline.yaml.erb",
"new_path": "ci/pipeline.yaml.erb",
"diff": "@@ -162,18 +162,6 @@ resources:\nbranch: master\ndepth: 1\n- - name: tiller-version.git\n- icon: key-variant\n- type: git-proxy\n- webhook_token: aldhjalkdhahdjkahdjkhjadhjadhkjadlkjhAAdd\n- check_every: 1h\n- source:\n- uri: git@github.wdf.sap.corp:cc/secrets.git\n- private_key: ((github-private-key/cc-secrets))\n- branch: master\n- depth: 1\n- paths: [ kubernikus/tiller.version, kubernikus/*/tiller.version ]\n-\n- name: helm-charts.git\ntype: git\nicon: github-circle\n@@ -282,21 +270,6 @@ jobs:\nimage: image/image.tar\nadditional_tags: kubernikus.git/.git/ref\n- - name: tiller_master\n- plan:\n- - in_parallel:\n- - get: tiller-version.git\n- trigger: true\n- - in_parallel:\n- - task: master\n- config:\n- <<: *task_tiller\n- timeout: 10m\n- params:\n- REGION: \"k-master\"\n- <<: *auth_master\n- <<: *slack\n-\n- name: master\nserial: true\nplan:\n@@ -441,32 +414,13 @@ jobs:\n<<: *auth_<%= region %>\n<<: *slack\n- - name: tiller_<%= region %>\n- plan:\n- - in_parallel:\n- - get: <%= group %>.tag\n- trigger: true\n- passed: [seed_<%= region %>]\n- - get: tiller-version.git\n- trigger: true\n- - in_parallel:\n- - task: k-<%= region %>\n- config:\n- <<: *task_tiller\n- timeout: 10m\n- params:\n- REGION: <%= region %>\n- CONTEXT: k-<%= region %>\n- <<: *auth_<%= region %>\n- <<: *slack\n-\n- name: deploy_<%= region %>\nserial: true\nplan:\n- in_parallel:\n- get: <%= group %>.tag\ntrigger: true\n- passed: [tiller_<%= region %>]\n+ passed: [seed_<%= region %>]\n- get: secrets.git\npassed: [seed_<%= region %>]\n- get: kubernikus.builds\n@@ -539,7 +493,6 @@ groups:\n- master\n- cli\n- whitesource\n- - tiller_master\n- all-versions\n<% VERSIONS.each do |version| %>\n- e2e-<%= version %>\n@@ -554,7 +507,6 @@ groups:\njobs:\n- <%= meta.fetch(:continent) %>\n- seed_<%= region %>\n- - tiller_<%= region %>\n- deploy_<%= region %>\n- soak_<%= region %>\n<% end %>\n@@ -564,7 +516,6 @@ groups:\n- master\n- cli\n- whitesource\n- - tiller_master\n- all-versions\n<% VERSIONS.each do |version| %>\n- e2e-<%= version %>\n@@ -575,7 +526,6 @@ groups:\n<% end %>\n<% DEPLOYABLE_REGIONS.each do |region, meta| %>\n- seed_<%= region %>\n- - tiller_<%= region %>\n- deploy_<%= region %>\n- soak_<%= region %>\n<% end %>\n"
},
{
"change_type": "DELETE",
"old_path": "ci/task_tiller.yaml",
"new_path": null,
"diff": "-platform: 'linux'\n-\n-image_resource:\n- type: registry-image\n- source:\n- repository: keppel.eu-de-1.cloud.sap/ccloud-dockerhub-mirror/sapcc/unified-kubernetes-toolbox\n- tag: 'latest'\n-\n-inputs:\n- - name: tiller-version.git\n-\n-run:\n- path: /bin/sh\n- args:\n- - -c\n- - |\n- set -exo pipefail\n- tag=$(cat tiller-version.git/kubernikus/tiller.version)\n- if [ -f tiller-version.git/kubernikus/$KUBERNIKUS_NAME/tiller.version ]; then\n- tag=$(cat tiller-version.git/kubernikus/$KUBERNIKUS_NAME/tiller.version )\n- fi\n- if kubectl get deployment tiller-deploy --namespace=kube-system >/dev/null ; then\n- kubectl set image --namespace=kube-system deployment/tiller-deploy tiller=sapcc/tiller:$tag\n- else\n- kubectl create sa tiller --namespace kube-system\n- kubectl create clusterrolebinding tiller --clusterrole cluster-admin --serviceaccount=kube-system:tiller\n- helm init --service-account tiller --history-max 5 --tiller-image sapcc/tiller:$tag --upgrade --wait\n- fi\n-\n-params:\n- REGION:\n- CONTEXT:\n- KUBELOGON_USER:\n- KUBELOGON_PASSWORD:\n- KUBERNIKUS_NAME:\n- KUBERNIKUS_URL:\n- HELM_VERSION: 2.11.0\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
Remove tiller from pipeline
|
596,226 |
06.07.2022 14:42:43
| -7,200 |
892771160706d36f30229b0ccf6b2c4c3aa2dbd7
|
Fix regression in cloud controller manager 1.23
|
[
{
"change_type": "MODIFY",
"old_path": "charts/images.yaml",
"new_path": "charts/images.yaml",
"diff": "@@ -23,8 +23,8 @@ imagesForVersion:\nrepository: 'keppel.$REGION.cloud.sap/ccloud/kube-proxy'\ntag: 'v1.23.6'\ncloudControllerManager:\n- repository: 'keppel.$REGION.cloud.sap/ccloud/openstack-cloud-controller-manager'\n- tag: 'v1.23.1'\n+ repository: 'keppel.$REGION.cloud.sap/ccloud/openstack-cloud-controller-manager-amd64'\n+ tag: '1.23.2-sap.1'\ndex:\nrepository: 'keppel.$REGION.cloud.sap/ccloud/dex'\ntag: '38f4f8ea8d487470a1dd5b83d66b428d8b502f81'\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
Fix regression in cloud controller manager 1.23
|
596,226 |
07.07.2022 12:07:54
| -7,200 |
99972fa9e5dcda321143927e6d750246bdc7bd69
|
Bump etcd backup sidecar in 1.23 to fix deadlock
|
[
{
"change_type": "MODIFY",
"old_path": "charts/images.yaml",
"new_path": "charts/images.yaml",
"diff": "@@ -15,7 +15,7 @@ imagesForVersion:\ntag: 'v3.4.13-bootstrap-3'\netcdBackup:\nrepository: 'keppel.$REGION.cloud.sap/ccloud/etcdbrctl'\n- tag: 'v0.15.1'\n+ tag: 'v0.15.4'\nkubelet:\nrepository: 'keppel.$REGION.cloud.sap/ccloud/kubelet'\ntag: 'v1.23.6'\n"
},
{
"change_type": "MODIFY",
"old_path": "contrib/all/Makefile",
"new_path": "contrib/all/Makefile",
"diff": "@@ -20,7 +20,7 @@ CSI_NODE_DRIVER_VERSION=v2.5.0\nCSI_PLUGIN_VERSION?=v1.23.0\nETCD_VERSION=v3.4.13-bootstrap-3\n-ETCDBRCTL_VERSION=v0.15.1\n+ETCDBRCTL_VERSION=v0.15.4\nFLANNEL_VERSION=v0.17.0\nCOREDNS_VERSION=1.9.1\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
Bump etcd backup sidecar in 1.23 to fix deadlock
|
596,226 |
07.07.2022 16:37:31
| -7,200 |
b86795f535d4d0f196892314aabe5ebd1516067e
|
Use helm 3.9.0 in pipeline
|
[
{
"change_type": "MODIFY",
"old_path": "ci/pipeline.yaml",
"new_path": "ci/pipeline.yaml",
"diff": "@@ -933,7 +933,7 @@ task_helm_kubernikus-dex: &task_helm_kubernikus-dex\nKUBELOGON_PASSWORD:\nKUBERNIKUS_NAME:\nKUBERNIKUS_URL:\n- HELM_VERSION: 3.2.4\n+ HELM_VERSION: 3.9.0\nVAULT_ADDR: https://vault.global.cloud.sap # DO NOT CHANGE\nVAULT_KV_ENGINE: secrets # DO NOT CHANGE\nVAULT_ROLE_ID: # (required) set this to ((auth.role_id)) to receive credentials automatically from Concourse\n@@ -996,7 +996,7 @@ task_helm_kubernikus: &task_helm_kubernikus\nKUBELOGON_PASSWORD:\nKUBERNIKUS_NAME:\nKUBERNIKUS_URL:\n- HELM_VERSION: 3.2.4\n+ HELM_VERSION: 3.9.0\nVAULT_ADDR: https://vault.global.cloud.sap # DO NOT CHANGE\nVAULT_KV_ENGINE: secrets # DO NOT CHANGE\nVAULT_ROLE_ID: # (required) set this to ((auth.role_id)) to receive credentials automatically from Concourse\n@@ -1034,7 +1034,7 @@ task_helm_seed: &task_helm_seed\nKUBELOGON_PASSWORD:\nNAMESPACE: monsoon3\nRELEASE: kubernikus\n- HELM_VERSION: 3.2.4\n+ HELM_VERSION: 3.9.0\nVAULT_ADDR: https://vault.global.cloud.sap # DO NOT CHANGE\nVAULT_KV_ENGINE: secrets # DO NOT CHANGE\nVAULT_ROLE_ID: # (required) set this to ((auth.role_id)) to receive credentials automatically from Concourse\n"
},
{
"change_type": "MODIFY",
"old_path": "ci/task_helm_kubernikus-dex.yaml",
"new_path": "ci/task_helm_kubernikus-dex.yaml",
"diff": "@@ -53,7 +53,7 @@ params:\nKUBELOGON_PASSWORD:\nKUBERNIKUS_NAME:\nKUBERNIKUS_URL:\n- HELM_VERSION: 3.2.4\n+ HELM_VERSION: 3.9.0\nVAULT_ADDR: https://vault.global.cloud.sap # DO NOT CHANGE\nVAULT_KV_ENGINE: secrets # DO NOT CHANGE\nVAULT_ROLE_ID: # (required) set this to ((auth.role_id)) to receive credentials automatically from Concourse\n"
},
{
"change_type": "MODIFY",
"old_path": "ci/task_helm_kubernikus.yaml",
"new_path": "ci/task_helm_kubernikus.yaml",
"diff": "@@ -53,7 +53,7 @@ params:\nKUBELOGON_PASSWORD:\nKUBERNIKUS_NAME:\nKUBERNIKUS_URL:\n- HELM_VERSION: 3.2.4\n+ HELM_VERSION: 3.9.0\nVAULT_ADDR: https://vault.global.cloud.sap # DO NOT CHANGE\nVAULT_KV_ENGINE: secrets # DO NOT CHANGE\nVAULT_ROLE_ID: # (required) set this to ((auth.role_id)) to receive credentials automatically from Concourse\n"
},
{
"change_type": "MODIFY",
"old_path": "ci/task_helm_seed.yaml",
"new_path": "ci/task_helm_seed.yaml",
"diff": "@@ -28,7 +28,7 @@ params:\nKUBELOGON_PASSWORD:\nNAMESPACE: monsoon3\nRELEASE: kubernikus\n- HELM_VERSION: 3.2.4\n+ HELM_VERSION: 3.9.0\nVAULT_ADDR: https://vault.global.cloud.sap # DO NOT CHANGE\nVAULT_KV_ENGINE: secrets # DO NOT CHANGE\nVAULT_ROLE_ID: # (required) set this to ((auth.role_id)) to receive credentials automatically from Concourse\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
Use helm 3.9.0 in pipeline
|
596,226 |
21.07.2022 10:27:45
| -7,200 |
c34033719eb90ab4648769cb43eb7885ff0537c6
|
Remove eu-ru-1 from pipeline
|
[
{
"change_type": "MODIFY",
"old_path": "ci/pipeline.yaml",
"new_path": "ci/pipeline.yaml",
"diff": "@@ -381,44 +381,6 @@ auth_e2e_eu-nl-1: &auth_e2e_eu-nl-1\n-auth_eu-ru-1: &auth_eu-ru-1\n- KUBELOGON_USER: ((unified-kubernetes-auth/default.username))\n- KUBELOGON_PASSWORD: ((unified-kubernetes-auth/default.password))\n- KUBERNIKUS_NAME: k-eu-ru-1\n- KUBERNIKUS_URL: https://kubernikus.admin.eu-ru-1.cloud.sap\n- VAULT_ROLE_ID: ((auth.role_id))\n- VAULT_SECRET_ID: ((auth.secret_id))\n-\n-auth_e2e_eu-ru-1: &auth_e2e_eu-ru-1\n- CP_KLUSTER: k-eu-ru-1\n- CP_KUBERNIKUS_URL: https://kubernikus.admin.eu-ru-1.cloud.sap\n- CP_OS_PROJECT_DOMAIN_NAME: ccadmin\n- VAULT_ROLE_ID: ((auth.role_id))\n- VAULT_SECRET_ID: ((auth.secret_id))\n-\n- CP_OIDC_AUTH_URL: https://kubernikus.admin.eu-ru-1.cloud.sap/auth/login\n- CP_OIDC_CONNECTOR_ID: eb31733cfb9d4ab283bb19e00863b780\n- CP_OIDC_USERNAME: ((unified-kubernetes-auth/default.username))\n- CP_OIDC_PASSWORD: ((unified-kubernetes-auth/default.password))\n-\n- KUBERNIKUS_NAME: e2e\n- OS_PROJECT_DOMAIN_NAME: ccadmin\n-\n- NODEPOOL_AVZ: eu-ru-1a\n-\n-\n- OS_AUTH_URL: https://identity-3.eu-ru-1.cloud.sap/v3\n- OS_USERNAME: kubernikus-pipeline\n- OS_PASSWORD: ((keystone-user-password/kubernikus-pipeline))\n- OS_USER_DOMAIN_NAME: Default\n- OS_PROJECT_NAME: kubernikus_e2e\n- OS_REGION_NAME: eu-ru-1\n- KUBERNIKUS_URL: https://kubernikus.eu-ru-1.cloud.sap\n- CP_OS_PROJECT_NAME: kubernikus\n-\n-\n-\n-\nauth_la-br-1: &auth_la-br-1\nKUBELOGON_USER: ((unified-kubernetes-auth/default.username))\nKUBELOGON_PASSWORD: ((unified-kubernetes-auth/default.password))\n@@ -2190,97 +2152,6 @@ jobs:\n- - name: seed_eu-ru-1\n- serial: true\n- plan:\n- - in_parallel:\n- - get: silver.tag\n- passed: [silver]\n- trigger: true\n- - get: secrets.git\n- passed: [silver]\n- - get: helm-charts.git\n- passed: [silver]\n- - task: seed\n- config:\n- <<: *task_helm_seed\n- input_mapping:\n- charts.git: helm-charts.git\n- params:\n- REGION: eu-ru-1\n- CONTEXT: eu-ru-1\n- <<: *auth_eu-ru-1\n- <<: *slack\n-\n- - name: deploy_eu-ru-1\n- serial: true\n- plan:\n- - in_parallel:\n- - get: silver.tag\n- trigger: true\n- passed: [seed_eu-ru-1]\n- - get: secrets.git\n- passed: [seed_eu-ru-1]\n- - get: kubernikus.builds\n- resource: kubernikus.git\n- passed: [silver]\n- - in_parallel:\n- - task: kubernikus_eu-ru-1\n- config:\n- <<: *task_helm_kubernikus\n- params:\n- REGION: eu-ru-1\n- CONTEXT: k-eu-ru-1\n- <<: *auth_eu-ru-1\n- - task: kubernikus-dex_eu-ru-1\n- config:\n- <<: *task_helm_kubernikus-dex\n- params:\n- REGION: eu-ru-1\n- CONTEXT: k-eu-ru-1\n- <<: *auth_eu-ru-1\n- <<: *slack\n-\n- - name: soak_eu-ru-1\n- serial: true\n- build_logs_to_retain: 168\n- plan:\n- - in_parallel:\n- - get: silver.tag\n- trigger: true\n- passed: [deploy_eu-ru-1]\n- - get: kubernikus.builds\n- resource: kubernikus.git\n- passed: [deploy_eu-ru-1]\n- - get: hourly\n- trigger: true\n- - task: e2e_tests\n- config:\n- <<: *task_e2e_tests\n- timeout: 45m\n- params:\n- <<: *auth_e2e_eu-ru-1\n-\n- on_failure:\n- put: slack-notification\n- params:\n- channel: '#kubernikus-ci'\n- text: Soak tests in eu-ru-1 failed\n- text_file: test-result/content\n- attachments:\n- - text: $TEXT_FILE_CONTENT\n- color: \"#d00000\"\n- footer: $ATC_EXTERNAL_URL/teams/$BUILD_TEAM_NAME/pipelines/$BUILD_PIPELINE_NAME/jobs/$BUILD_JOB_NAME/builds/$BUILD_NAME\n- footer_icon: https://ci.concourse-ci.org/public/images/favicon-failed.png\n-\n- on_success: { put: slack-alert, inputs: [], params: {alert_type: fixed } }\n- on_abort: { put: slack-alert, inputs: [], params: {alert_type: broke } }\n- on_error: { put: slack-alert, inputs: [], params: {alert_type: broke } }\n-\n-\n-\n-\n-\n- name: seed_la-br-1\nserial: true\nplan:\n@@ -3445,13 +3316,6 @@ groups:\n- deploy_eu-nl-1\n- soak_eu-nl-1\n- - name: eu-ru-1\n- jobs:\n- - silver\n- - seed_eu-ru-1\n- - deploy_eu-ru-1\n- - soak_eu-ru-1\n-\n- name: la-br-1\njobs:\n- silver\n@@ -3575,10 +3439,6 @@ groups:\n- deploy_eu-nl-1\n- soak_eu-nl-1\n- - seed_eu-ru-1\n- - deploy_eu-ru-1\n- - soak_eu-ru-1\n-\n- seed_la-br-1\n- deploy_la-br-1\n- soak_la-br-1\n"
},
{
"change_type": "MODIFY",
"old_path": "ci/pipeline.yaml.erb",
"new_path": "ci/pipeline.yaml.erb",
"diff": "@@ -24,7 +24,6 @@ REGIONS = {\n\"eu-de-1\": { continent: 'gold', install: 'manual' , e2e: 'manual', nodepool_avz: 'eu-de-1d', oidc: 'd7df5ce6c37643e49b3c93528c29818b' },\n\"eu-de-2\": { continent: 'gold', install: 'manual' , e2e: 'manual', nodepool_avz: 'eu-de-2b', oidc: '57da32d8d50d49a9a9fbbe6b890e9e13' },\n\"eu-nl-1\": { continent: 'gold', install: 'manual' , e2e: 'terraform', nodepool_avz: 'eu-nl-1a', oidc: 'd9b056a6762649f684a8a02b30f38324' },\n- \"eu-ru-1\": { continent: 'silver', install: 'terraform', e2e: 'terraform', nodepool_avz: 'eu-ru-1a', oidc: 'eb31733cfb9d4ab283bb19e00863b780' },\n\"la-br-1\": { continent: 'silver', install: 'terraform', e2e: 'terraform', nodepool_avz: 'la-br-1b', oidc: 'b33ac00cb0004d50bb34eb4e4cbbeb54' },\n\"na-ca-1\": { continent: 'bronze', install: 'terraform', e2e: 'terraform', nodepool_avz: 'na-ca-1b', oidc: 'b7269ea1dde54a6383453ab69f25cce4' },\n\"na-us-1\": { continent: 'gold', install: 'manual' , e2e: 'manual', nodepool_avz: 'na-us-1b', oidc: '040ed45cb5d8403790cf341a26c50601' },\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
Remove eu-ru-1 from pipeline
|
596,227 |
01.08.2022 15:17:02
| -7,200 |
0f0196ab7a40459b85fce0dcdb09515623cc6f4b
|
Ignore missing release when uninstalling
|
[
{
"change_type": "MODIFY",
"old_path": "pkg/controller/ground.go",
"new_path": "pkg/controller/ground.go",
"diff": "@@ -13,7 +13,6 @@ import (\n\"github.com/go-kit/kit/log\"\n\"github.com/go-openapi/swag\"\n\"github.com/pkg/errors\"\n- \"google.golang.org/grpc\"\n\"helm.sh/helm/v3/pkg/action\"\n\"helm.sh/helm/v3/pkg/chart/loader\"\napi_v1 \"k8s.io/api/core/v1\"\n@@ -742,11 +741,8 @@ func (op *GroundControl) terminateKluster(kluster *v1.Kluster) error {\nuninstall := action.NewUninstall(op.Helm3)\n_, err := uninstall.Run(kluster.GetName())\n- if err != nil {\n- return err\n- }\n- if err != nil && !strings.Contains(grpc.ErrorDesc(err), fmt.Sprintf(`%s: release: not found`, kluster.GetName())) { //nolint:staticcheck\n+ if err != nil && !strings.Contains(err.Error(), fmt.Sprintf(`%s: release: not found`, kluster.GetName())) { //nolint:staticcheck\nreturn err\n}\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
Ignore missing release when uninstalling (#676)
|
596,226 |
25.08.2022 13:46:11
| -7,200 |
c4be40db13e3a7ee7b815a0d557b2632369a7928
|
Use rbac v1 in kubernikus chart
|
[
{
"change_type": "MODIFY",
"old_path": "charts/kubernikus/Chart.yaml",
"new_path": "charts/kubernikus/Chart.yaml",
"diff": "@@ -2,7 +2,7 @@ apiVersion: v2\ndescription: A Helm chart for Kubernetes\nname: kubernikus\ntype: application\n-version: 0.3.14\n+version: 0.3.15\ndependencies:\n- name: k8sniff\nrepository: file://../k8sniff\n"
},
{
"change_type": "MODIFY",
"old_path": "charts/kubernikus/templates/clusterrolebinding.yaml",
"new_path": "charts/kubernikus/templates/clusterrolebinding.yaml",
"diff": "{{- if .Values.standalone}}\n-apiVersion: rbac.authorization.k8s.io/v1beta1\n+apiVersion: rbac.authorization.k8s.io/v1\nkind: ClusterRoleBinding\nmetadata:\nname: kubernikus-default\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
Use rbac v1 in kubernikus chart
|
596,226 |
08.09.2022 13:19:30
| -7,200 |
68d74ec74b3c3940eae0b22b03224960e3d68d7a
|
Disable servicing reboot
* Revert "Disable servicing controller"
This reverts commit
* Only disable reboot in servicing
|
[
{
"change_type": "MODIFY",
"old_path": "pkg/controller/operator.go",
"new_path": "pkg/controller/operator.go",
"diff": "@@ -29,6 +29,7 @@ import (\n\"github.com/sapcc/kubernikus/pkg/controller/migration\"\n\"github.com/sapcc/kubernikus/pkg/controller/nodeobservatory\"\n\"github.com/sapcc/kubernikus/pkg/controller/routegc\"\n+ \"github.com/sapcc/kubernikus/pkg/controller/servicing\"\nkubernikus_informers \"github.com/sapcc/kubernikus/pkg/generated/informers/externalversions\"\n_ \"github.com/sapcc/kubernikus/pkg/util/workqueue/prometheus\"\n\"github.com/sapcc/kubernikus/pkg/version\"\n@@ -194,8 +195,8 @@ func NewKubernikusOperator(options *KubernikusOperatorOptions, logger log.Logger\no.Config.Kubernikus.Controllers[\"migration\"] = migration.NewController(3, o.Factories, o.Clients, recorder, logger)\ncase \"hammertime\":\no.Config.Kubernikus.Controllers[\"hammertime\"] = hammertime.New(10*time.Second, 20*time.Second, o.Factories, o.Clients, recorder, logger)\n- //case \"servicing\":\n- //o.Config.Kubernikus.Controllers[\"servicing\"] = servicing.NewController(10, o.Factories, o.Clients, recorder, options.NodeUpdateHoldoff, logger)\n+ case \"servicing\":\n+ o.Config.Kubernikus.Controllers[\"servicing\"] = servicing.NewController(10, o.Factories, o.Clients, recorder, options.NodeUpdateHoldoff, logger)\ncase \"certs\":\no.Config.Kubernikus.Controllers[\"certs\"] = certs.New(12*time.Hour, o.Factories, o.Config, o.Clients, logger)\n}\n"
},
{
"change_type": "MODIFY",
"old_path": "pkg/controller/servicing/reconciler.go",
"new_path": "pkg/controller/servicing/reconciler.go",
"diff": "@@ -5,6 +5,7 @@ import (\n\"github.com/go-kit/kit/log\"\n\"github.com/pkg/errors\"\n+ core_v1 \"k8s.io/api/core/v1\"\n\"k8s.io/client-go/tools/record\"\n\"github.com/sapcc/kubernikus/pkg/api/models\"\n@@ -154,7 +155,8 @@ func (r *KlusterReconciler) Do() error {\nupdate := r.Lister.Updating()\nreplace := r.Lister.Replace()\n- reboot := r.Lister.Reboot()\n+ //reboot := r.Lister.Reboot()\n+ reboot := make([]*core_v1.Node, 0)\n// The following block retires already updating nodes\nif len(update) > 0 {\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
Disable servicing reboot (#684)
* Revert "Disable servicing controller"
This reverts commit 0e995ef426ceb7629ac392978893bd2668b8cb77.
* Only disable reboot in servicing
|
596,226 |
08.09.2022 13:53:48
| -7,200 |
36c357e51fe1d2d2c91f843a35bc0a1321380416
|
Disable servicing reboot unit test
|
[
{
"change_type": "MODIFY",
"old_path": "pkg/controller/servicing/controller_test.go",
"new_path": "pkg/controller/servicing/controller_test.go",
"diff": "@@ -241,6 +241,7 @@ func TestServicingControllerReconcile(t *testing.T) {\nexpectedReboot: false,\nexpectedReplace: true,\n},\n+ /*\n{\nmessage: \"Nodes with outdate OS should be rebooted\",\noptions: &FakeKlusterOptions{\n@@ -261,6 +262,7 @@ func TestServicingControllerReconcile(t *testing.T) {\nexpectedReboot: true,\nexpectedReplace: false,\n},\n+ */\n{\nmessage: \"Up-to-date Node should neither be rebooted nor be replaced\",\noptions: &FakeKlusterOptions{\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
Disable servicing reboot unit test
|
596,226 |
08.09.2022 14:17:47
| -7,200 |
7785fa8954ee012a804eb3ce94799f267332b904
|
Disable flatcar node version e2e test
|
[
{
"change_type": "MODIFY",
"old_path": "test/e2e/node_test.go",
"new_path": "test/e2e/node_test.go",
"diff": "@@ -47,7 +47,7 @@ func (k *NodeTests) Run(t *testing.T) {\n_ = t.Run(\"Created\", k.StateRunning) &&\nt.Run(\"Tagged\", k.Tagged) &&\nt.Run(\"Registered\", k.Registered) &&\n- t.Run(\"LatestContainerLinux\", k.LatestContainerLinux) &&\n+ //t.Run(\"LatestContainerLinux\", k.LatestContainerLinux) &&\nt.Run(\"Schedulable\", k.StateSchedulable) &&\nt.Run(\"NetworkUnavailable\", k.ConditionNetworkUnavailable) &&\nt.Run(\"Healthy\", k.StateHealthy) &&\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
Disable flatcar node version e2e test
|
596,226 |
22.09.2022 16:37:56
| -7,200 |
b9396bdc0dc425151c5106da8f1b48a0fb9cd637
|
Bump versions for e2e tests
|
[
{
"change_type": "MODIFY",
"old_path": "ci/pipeline.yaml",
"new_path": "ci/pipeline.yaml",
"diff": "@@ -752,44 +752,6 @@ task_cli: &task_cli\nparams:\n-task_cli2: &task_cli2\n- platform: 'linux'\n-\n- image_resource:\n- type: registry-image\n- source:\n- repository: keppel.eu-de-1.cloud.sap/ccloud-dockerhub-mirror/library/golang\n- tag: 1.17-alpine\n- inputs:\n- - name: kubernikus.builds\n- path: build\n- outputs:\n- - name: binaries\n-\n- run:\n- path: /bin/sh\n- args:\n- - -exc\n- - |\n- apk add --no-cache upx git curl gpg gpg-agent\n- echo \"$SIGNING_KEY\" > gpg.key\n- gpg --import gpg.key\n-\n- curl -fLO https://github.com/goreleaser/goreleaser/releases/download/v1.10.3/goreleaser_1.10.3_x86_64.apk\n- apk add --allow-untrusted goreleaser_1.10.3_x86_64.apk\n-\n- cd build\n- git config --global --add safe.directory $PWD #not sure this is ony neede for fly execute but YOLO\n- git status\n- git clean -f\n- git tag v1.0.0+$(git rev-parse HEAD)\n- goreleaser release\n-\n- params:\n- SIGNING_KEY:\n- GITHUB_TOKEN:\n-\n-\ntask_e2e_tests: &task_e2e_tests\nplatform: 'linux'\n@@ -1606,7 +1568,7 @@ jobs:\n- put: all-versions.trigger\n- - name: e2e-1.24.5\n+ - name: e2e-1.24.6\nserial: true\nplan:\n- in_parallel:\n@@ -1623,14 +1585,14 @@ jobs:\n<<: *task_e2e_tests\ntimeout: 2h\nparams:\n- KLUSTER_VERSION: \"1.24.5\"\n+ KLUSTER_VERSION: \"1.24.6\"\nKLUSTER_CIDR: \"100.100.0.0/16\"\nISOLATE_TEST: \"true\"\nKLUSTER_OS_IMAGES: flatcar-stable-amd64,flatcar-beta-amd64\n<<: *auth_e2e_master\n<<: *slack\n- - name: e2e-1.23.11\n+ - name: e2e-1.23.12\nserial: true\nplan:\n- in_parallel:\n@@ -1647,14 +1609,14 @@ jobs:\n<<: *task_e2e_tests\ntimeout: 2h\nparams:\n- KLUSTER_VERSION: \"1.23.11\"\n+ KLUSTER_VERSION: \"1.23.12\"\nKLUSTER_CIDR: \"100.101.0.0/16\"\nISOLATE_TEST: \"true\"\nKLUSTER_OS_IMAGES: flatcar-stable-amd64,flatcar-beta-amd64\n<<: *auth_e2e_master\n<<: *slack\n- - name: e2e-1.22.14\n+ - name: e2e-1.22.15\nserial: true\nplan:\n- in_parallel:\n@@ -1671,7 +1633,7 @@ jobs:\n<<: *task_e2e_tests\ntimeout: 2h\nparams:\n- KLUSTER_VERSION: \"1.22.14\"\n+ KLUSTER_VERSION: \"1.22.15\"\nKLUSTER_CIDR: \"100.102.0.0/16\"\nISOLATE_TEST: \"true\"\nKLUSTER_OS_IMAGES: flatcar-stable-amd64,flatcar-beta-amd64\n@@ -1900,7 +1862,7 @@ jobs:\nplan:\n- get: kubernikus.builds\nresource: kubernikus.git\n- passed: [e2e-1.24.5]\n+ passed: [e2e-1.24.6]\n- task: changes\nconfig:\n<<: *task_github_compare_url\n@@ -3470,11 +3432,11 @@ groups:\n- whitesource\n- all-versions\n- - e2e-1.24.5\n+ - e2e-1.24.6\n- - e2e-1.23.11\n+ - e2e-1.23.12\n- - e2e-1.22.14\n+ - e2e-1.22.15\n- e2e-1.21.5\n@@ -3626,11 +3588,11 @@ groups:\n- whitesource\n- all-versions\n- - e2e-1.24.5\n+ - e2e-1.24.6\n- - e2e-1.23.11\n+ - e2e-1.23.12\n- - e2e-1.22.14\n+ - e2e-1.22.15\n- e2e-1.21.5\n"
},
{
"change_type": "MODIFY",
"old_path": "ci/pipeline.yaml.erb",
"new_path": "ci/pipeline.yaml.erb",
"diff": "<%\nVERSIONS = [\n- '1.24.5',\n- '1.23.11',\n- '1.22.14',\n+ '1.24.6',\n+ '1.23.12',\n+ '1.22.15',\n'1.21.5',\n'1.20.8',\n'1.19.11',\n"
},
{
"change_type": "MODIFY",
"old_path": "test/e2e/setup_test.go",
"new_path": "test/e2e/setup_test.go",
"diff": "@@ -40,7 +40,7 @@ func (s *SetupTests) Run(t *testing.T) {\n}\nfunc (s *SetupTests) CreateCluster(t *testing.T) {\n- version := \"1.23.11\"\n+ version := \"1.23.12\"\nif v := os.Getenv(\"KLUSTER_VERSION\"); v != \"\" {\nversion = v\n}\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
Bump versions for e2e tests
|
596,230 |
29.09.2022 15:44:13
| -7,200 |
34716b6a301ac36aff345da80cb173b7999eab84
|
include scheduler config into extra vars
|
[
{
"change_type": "MODIFY",
"old_path": "charts/kube-master/templates/scheduler.yaml",
"new_path": "charts/kube-master/templates/scheduler.yaml",
"diff": "@@ -44,6 +44,11 @@ spec:\n- name: config\nconfigMap:\nname: {{ include \"master.fullname\" . }}\n+{{- if .Values.scheduler.config }}\n+ - name: scheduler-config\n+ configMap:\n+ name: {{ include \"master.fullname\" }}-scheduler\n+{{- end }}\ncontainers:\n- name: scheduler\n{{- if (semverCompare \">= 1.12\" .Values.version.kubernetes) }}\n@@ -72,6 +77,9 @@ spec:\n- --authorization-kubeconfig=/etc/kubernetes/config/kubeconfig\n{{- end }}\n- --leader-elect=false\n+{{- if .Values.scheduler.config }}\n+ - --config=/etc/kubernetes/scheduler/scheduler.config\n+{{- end }}\nlivenessProbe:\nhttpGet:\npath: /healthz\n@@ -91,5 +99,24 @@ spec:\n- mountPath: /etc/kubernetes/config\nname: config\nreadOnly: true\n+{{- if .Values.scheduler.config }}\n+ - mountPath: /etc/kubernetes/scheduler\n+ name: scheduler-config\n+ readOnly: true\n+{{- end}}\nresources:\n{{ toYaml .Values.scheduler.resources | indent 12 }}\n+{{- if .Values.scheduler.config }}\n+---\n+apiVersion: v1\n+kind: ConfigMap\n+metadata:\n+ name: {{ include \"master.fullname\" }}-scheduler\n+ labels:\n+ app: {{ include \"master.fullname\" }}-scheduler\n+ chart: {{ .Chart.Name }}-{{ .Chart.Version | replace \"+\" \"_\" }}\n+ release: {{ .Release.Name }}\n+data:\n+ scheduler.config: |-\n+{{ toYaml .Values.scheduler.config | indent 4 }}\n+{{- end}}\n\\ No newline at end of file\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
include scheduler config into extra vars (#673)
Co-authored-by: Fabian Ruff <fabian.ruff@sap.com>
|
596,226 |
29.09.2022 16:43:31
| -7,200 |
b4add70edd2f71be3f8477bab1dd3495f239a1f5
|
Use k8s 1.24.6 as default in e2e test
|
[
{
"change_type": "MODIFY",
"old_path": "test/e2e/setup_test.go",
"new_path": "test/e2e/setup_test.go",
"diff": "@@ -40,7 +40,7 @@ func (s *SetupTests) Run(t *testing.T) {\n}\nfunc (s *SetupTests) CreateCluster(t *testing.T) {\n- version := \"1.23.12\"\n+ version := \"1.24.6\"\nif v := os.Getenv(\"KLUSTER_VERSION\"); v != \"\" {\nversion = v\n}\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
Use k8s 1.24.6 as default in e2e test
|
596,226 |
06.10.2022 17:05:01
| -7,200 |
93f3d3bc01f436388503cfe623437be231940883
|
Use go 1.18 in ci
|
[
{
"change_type": "MODIFY",
"old_path": "ci/pipeline.yaml",
"new_path": "ci/pipeline.yaml",
"diff": "@@ -655,7 +655,7 @@ task_build_args: &task_build_args\ntype: registry-image\nsource:\nrepository: keppel.eu-de-1.cloud.sap/ccloud-dockerhub-mirror/library/alpine\n- tag: 3.13\n+ tag: latest\ninputs:\n- name: docs-builder.image\n- name: kubernikus.git\n@@ -679,7 +679,7 @@ task_checksum: &task_checksum\ntype: registry-image\nsource:\nrepository: keppel.eu-de-1.cloud.sap/ccloud-dockerhub-mirror/library/golang\n- tag: 1.16-alpine3.13\n+ tag: 1.18-alpine\ninputs:\n- name: kubernikus.builds\n@@ -715,7 +715,7 @@ task_cli: &task_cli\ntype: registry-image\nsource:\nrepository: keppel.eu-de-1.cloud.sap/ccloud-dockerhub-mirror/library/golang\n- tag: 1.17-alpine\n+ tag: 1.18-alpine\ninputs:\n- name: kubernikus.builds\npath: gopath/src/github.com/sapcc/kubernikus\n@@ -759,7 +759,7 @@ task_e2e_tests: &task_e2e_tests\ntype: registry-image\nsource:\nrepository: keppel.eu-de-1.cloud.sap/ccloud-dockerhub-mirror/library/golang\n- tag: 1.16-alpine3.13\n+ tag: 1.18-alpine\ninputs:\n- name: kubernikus.builds\n@@ -822,7 +822,7 @@ task_github_compare_url: &task_github_compare_url\ntype: registry-image\nsource:\nrepository: keppel.eu-de-1.cloud.sap/ccloud-dockerhub-mirror/library/alpine\n- tag: 3.13\n+ tag: latest\ninputs:\n- name: kubernikus.builds\noutputs:\n"
},
{
"change_type": "MODIFY",
"old_path": "ci/task_build_args.yaml",
"new_path": "ci/task_build_args.yaml",
"diff": "@@ -3,7 +3,7 @@ image_resource:\ntype: registry-image\nsource:\nrepository: keppel.eu-de-1.cloud.sap/ccloud-dockerhub-mirror/library/alpine\n- tag: 3.13\n+ tag: latest\ninputs:\n- name: docs-builder.image\n- name: kubernikus.git\n"
},
{
"change_type": "MODIFY",
"old_path": "ci/task_checksum.yaml",
"new_path": "ci/task_checksum.yaml",
"diff": "@@ -4,7 +4,7 @@ image_resource:\ntype: registry-image\nsource:\nrepository: keppel.eu-de-1.cloud.sap/ccloud-dockerhub-mirror/library/golang\n- tag: 1.16-alpine3.13\n+ tag: 1.18-alpine\ninputs:\n- name: kubernikus.builds\n"
},
{
"change_type": "MODIFY",
"old_path": "ci/task_cli.yaml",
"new_path": "ci/task_cli.yaml",
"diff": "@@ -4,7 +4,7 @@ image_resource:\ntype: registry-image\nsource:\nrepository: keppel.eu-de-1.cloud.sap/ccloud-dockerhub-mirror/library/golang\n- tag: 1.17-alpine\n+ tag: 1.18-alpine\ninputs:\n- name: kubernikus.builds\npath: gopath/src/github.com/sapcc/kubernikus\n"
},
{
"change_type": "MODIFY",
"old_path": "ci/task_e2e_tests.yaml",
"new_path": "ci/task_e2e_tests.yaml",
"diff": "@@ -4,7 +4,7 @@ image_resource:\ntype: registry-image\nsource:\nrepository: keppel.eu-de-1.cloud.sap/ccloud-dockerhub-mirror/library/golang\n- tag: 1.16-alpine3.13\n+ tag: 1.18-alpine\ninputs:\n- name: kubernikus.builds\n"
},
{
"change_type": "MODIFY",
"old_path": "ci/task_github_compare_url.yaml",
"new_path": "ci/task_github_compare_url.yaml",
"diff": "@@ -4,7 +4,7 @@ image_resource:\ntype: registry-image\nsource:\nrepository: keppel.eu-de-1.cloud.sap/ccloud-dockerhub-mirror/library/alpine\n- tag: 3.13\n+ tag: latest\ninputs:\n- name: kubernikus.builds\noutputs:\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
Use go 1.18 in ci
|
596,226 |
11.10.2022 09:55:51
| -7,200 |
4e0c43e7844d43469e4ea8be36d92c15fc430b2b
|
Use la-br-1a in e2e test
|
[
{
"change_type": "MODIFY",
"old_path": "ci/pipeline.yaml",
"new_path": "ci/pipeline.yaml",
"diff": "@@ -404,7 +404,7 @@ auth_e2e_la-br-1: &auth_e2e_la-br-1\nKUBERNIKUS_NAME: e2e\nOS_PROJECT_DOMAIN_NAME: ccadmin\n- NODEPOOL_AVZ: la-br-1b\n+ NODEPOOL_AVZ: la-br-1a\nOS_AUTH_URL: https://identity-3.la-br-1.cloud.sap/v3\n"
},
{
"change_type": "MODIFY",
"old_path": "ci/pipeline.yaml.erb",
"new_path": "ci/pipeline.yaml.erb",
"diff": "@@ -16,7 +16,7 @@ REGIONS = {\n\"eu-de-1\": { continent: 'gold', install: 'manual' , e2e: 'manual', nodepool_avz: 'eu-de-1d', oidc: 'd7df5ce6c37643e49b3c93528c29818b' },\n\"eu-de-2\": { continent: 'gold', install: 'manual' , e2e: 'manual', nodepool_avz: 'eu-de-2b', oidc: '57da32d8d50d49a9a9fbbe6b890e9e13' },\n\"eu-nl-1\": { continent: 'gold', install: 'manual' , e2e: 'terraform', nodepool_avz: 'eu-nl-1a', oidc: 'd9b056a6762649f684a8a02b30f38324' },\n- \"la-br-1\": { continent: 'silver', install: 'terraform', e2e: 'terraform', nodepool_avz: 'la-br-1b', oidc: 'b33ac00cb0004d50bb34eb4e4cbbeb54' },\n+ \"la-br-1\": { continent: 'silver', install: 'terraform', e2e: 'terraform', nodepool_avz: 'la-br-1a', oidc: 'b33ac00cb0004d50bb34eb4e4cbbeb54' },\n\"na-ca-1\": { continent: 'bronze', install: 'terraform', e2e: 'terraform', nodepool_avz: 'na-ca-1b', oidc: 'b7269ea1dde54a6383453ab69f25cce4' },\n\"na-us-1\": { continent: 'gold', install: 'manual' , e2e: 'manual', nodepool_avz: 'na-us-1b', oidc: '040ed45cb5d8403790cf341a26c50601' },\n\"na-us-2\": { continent: 'gold', install: 'terraform', e2e: 'terraform', nodepool_avz: 'na-us-2b', oidc: 'fc0f2b8450e1444197609800c8b5c4cb' },\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
Use la-br-1a in e2e test
|
596,229 |
14.10.2022 10:11:31
| -7,200 |
77b5f01045a24a419dd450f3060d8f098b39195d
|
Increase proxy-buffer-size for Kubernetes dashboard
|
[
{
"change_type": "MODIFY",
"old_path": "charts/kube-master/templates/dashboard.yaml",
"new_path": "charts/kube-master/templates/dashboard.yaml",
"diff": "@@ -158,6 +158,8 @@ apiVersion: networking.k8s.io/v1beta1\n{{- end }}\nkind: Ingress\nmetadata:\n+ annotations:\n+ ingress.kubernetes.io/proxy-buffer-size: 8k\nlabels:\napp: {{ include \"master.fullname\" . }}-dashboard\nchart: {{ .Chart.Name }}-{{ .Chart.Version | replace \"+\" \"_\" }}\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
Increase proxy-buffer-size for Kubernetes dashboard (#704)
|
596,226 |
17.10.2022 15:13:58
| -7,200 |
3ed6dcd9e47d9e077774636e37c25f18c4a30563
|
Fix csi snapshot controller rbac
|
[
{
"change_type": "MODIFY",
"old_path": "pkg/controller/ground/bootstrap/csi/manifest.go",
"new_path": "pkg/controller/ground/bootstrap/csi/manifest.go",
"diff": "@@ -1122,13 +1122,16 @@ rules:\nverbs: [\"get\", \"list\", \"watch\"]\n- apiGroups: [\"snapshot.storage.k8s.io\"]\nresources: [\"volumesnapshotcontents\"]\n- verbs: [\"create\", \"get\", \"list\", \"watch\", \"update\", \"delete\"]\n+ verbs: [\"create\", \"get\", \"list\", \"watch\", \"update\", \"delete\", \"patch\"]\n+ - apiGroups: [\"snapshot.storage.k8s.io\"]\n+ resources: [\"volumesnapshotcontents/status\"]\n+ verbs: [\"patch\"]\n- apiGroups: [\"snapshot.storage.k8s.io\"]\nresources: [\"volumesnapshots\"]\n- verbs: [\"get\", \"list\", \"watch\", \"update\"]\n+ verbs: [\"get\", \"list\", \"watch\", \"update\", \"patch\"]\n- apiGroups: [\"snapshot.storage.k8s.io\"]\nresources: [\"volumesnapshots/status\"]\n- verbs: [\"update\"]\n+ verbs: [\"update\", \"patch\"]\n`\nCSISnapshotControllerClusterRoleBinding = `\nkind: ClusterRoleBinding\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
Fix csi snapshot controller rbac
|
596,226 |
20.10.2022 10:58:45
| -7,200 |
df2f3a04acd18283930a9ce8f83b637bd330f79f
|
Pin concourse-slack-alert-resource to version v0.16.0
|
[
{
"change_type": "MODIFY",
"old_path": "ci/pipeline.yaml",
"new_path": "ci/pipeline.yaml",
"diff": "@@ -1430,6 +1430,7 @@ resource_types:\ncheck_every: 24h\nsource:\nrepository: keppel.eu-de-2.cloud.sap/ccloud-dockerhub-mirror/arbourd/concourse-slack-alert-resource\n+ tag: v0.16.0\n- name: slack-notification\ntype: docker-image\n"
},
{
"change_type": "MODIFY",
"old_path": "ci/pipeline.yaml.erb",
"new_path": "ci/pipeline.yaml.erb",
"diff": "@@ -209,6 +209,7 @@ resource_types:\ncheck_every: 24h\nsource:\nrepository: keppel.eu-de-2.cloud.sap/ccloud-dockerhub-mirror/arbourd/concourse-slack-alert-resource\n+ tag: v0.16.0\n- name: slack-notification\ntype: docker-image\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
Pin concourse-slack-alert-resource to version v0.16.0
|
596,227 |
20.10.2022 16:15:26
| -7,200 |
74c12b886620a3d5f2815dc0dfbb9895600f3484
|
Use Patch to apply a migration
|
[
{
"change_type": "MODIFY",
"old_path": "go.mod",
"new_path": "go.mod",
"diff": "@@ -17,6 +17,7 @@ require (\ngithub.com/databus23/guttle v0.0.0-20210623071842-89102dbdfc85\ngithub.com/databus23/keystone v0.0.0-20180111110916-350fd0e663cd\ngithub.com/databus23/requestutil v0.0.0-20160108082528-5ff8e981f38f\n+ github.com/evanphx/json-patch v5.6.0+incompatible\ngithub.com/ghodss/yaml v1.0.0\ngithub.com/go-kit/kit v0.10.0\ngithub.com/go-openapi/errors v0.20.2\n"
},
{
"change_type": "MODIFY",
"old_path": "go.sum",
"new_path": "go.sum",
"diff": "@@ -461,8 +461,9 @@ github.com/envoyproxy/protoc-gen-validate v0.1.0/go.mod h1:iSmxcyjqTsJpI2R4NaDN7\ngithub.com/evanphx/json-patch v4.2.0+incompatible/go.mod h1:50XU6AFN0ol/bzJsmQLiYLvXMP4fmwYFNcr97nuDLSk=\ngithub.com/evanphx/json-patch v4.9.0+incompatible/go.mod h1:50XU6AFN0ol/bzJsmQLiYLvXMP4fmwYFNcr97nuDLSk=\ngithub.com/evanphx/json-patch v4.11.0+incompatible/go.mod h1:50XU6AFN0ol/bzJsmQLiYLvXMP4fmwYFNcr97nuDLSk=\n-github.com/evanphx/json-patch v4.12.0+incompatible h1:4onqiflcdA9EOZ4RxV643DvftH5pOlLGNtQ5lPWQu84=\ngithub.com/evanphx/json-patch v4.12.0+incompatible/go.mod h1:50XU6AFN0ol/bzJsmQLiYLvXMP4fmwYFNcr97nuDLSk=\n+github.com/evanphx/json-patch v5.6.0+incompatible h1:jBYDEEiFBPxA0v50tFdvOzQQTCvpL6mnFh5mB2/l16U=\n+github.com/evanphx/json-patch v5.6.0+incompatible/go.mod h1:50XU6AFN0ol/bzJsmQLiYLvXMP4fmwYFNcr97nuDLSk=\ngithub.com/exponent-io/jsonpath v0.0.0-20151013193312-d6023ce2651d h1:105gxyaGwCFad8crR9dcMQWvV9Hvulu6hwUh4tWPJnM=\ngithub.com/exponent-io/jsonpath v0.0.0-20151013193312-d6023ce2651d/go.mod h1:ZZMPRZwes7CROmyNKgQzC3XPs6L/G2EJLHddWejkmf4=\ngithub.com/fatih/camelcase v1.0.0/go.mod h1:yN2Sb0lFhZJUdVvtELVWefmrXpuZESvPmqwoZc+/fpc=\n"
},
{
"change_type": "MODIFY",
"old_path": "pkg/migration/migration.go",
"new_path": "pkg/migration/migration.go",
"diff": "@@ -2,9 +2,12 @@ package migration\nimport (\n\"context\"\n+ \"encoding/json\"\n\"fmt\"\n+ jsonpatch \"github.com/evanphx/json-patch\"\nmetav1 \"k8s.io/apimachinery/pkg/apis/meta/v1\"\n+ \"k8s.io/apimachinery/pkg/types\"\nv1 \"github.com/sapcc/kubernikus/pkg/apis/kubernikus/v1\"\n\"github.com/sapcc/kubernikus/pkg/controller/config\"\n@@ -75,6 +78,8 @@ func migrateKluster(kluster *v1.Kluster, version int, migration Migration, clien\nvar rawData []byte\nvar err error\n+ original := kluster.DeepCopy()\n+\n//TODO: Don't import fake pkg outside of test code\nif _, ok := clients.Kubernikus.(*kubernikusfake.Clientset); !ok {\nrequest := clients.Kubernikus.KubernikusV1().RESTClient().Get().Namespace(kluster.Namespace).Resource(\"klusters\").Name(kluster.Name)\n@@ -87,5 +92,19 @@ func migrateKluster(kluster *v1.Kluster, version int, migration Migration, clien\nreturn nil, err\n}\nkluster.Status.SpecVersion = int64(version)\n- return clients.Kubernikus.KubernikusV1().Klusters(kluster.Namespace).Update(context.TODO(), kluster, metav1.UpdateOptions{})\n+\n+ originalJSON, err := json.Marshal(original)\n+ if err != nil {\n+ return nil, err\n+ }\n+ klusterJSON, err := json.Marshal(kluster)\n+ if err != nil {\n+ return nil, err\n+ }\n+ patchJSON, err := jsonpatch.CreateMergePatch(originalJSON, klusterJSON)\n+ if err != nil {\n+ return nil, err\n+ }\n+\n+ return clients.Kubernikus.KubernikusV1().Klusters(kluster.Namespace).Patch(context.TODO(), kluster.Name, types.MergePatchType, patchJSON, metav1.PatchOptions{})\n}\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
Use Patch to apply a migration (#698)
|
596,226 |
20.10.2022 16:20:57
| -7,200 |
9ca281956e2c739e22b1028fc3f930aa75fcacd5
|
Changes pipeline
|
[
{
"change_type": "MODIFY",
"old_path": "ci/pipeline.yaml",
"new_path": "ci/pipeline.yaml",
"diff": "@@ -1569,7 +1569,7 @@ jobs:\n- put: all-versions.trigger\n- - name: e2e-1.24.6\n+ - name: e2e-1.24.7\nserial: true\nplan:\n- in_parallel:\n@@ -1586,7 +1586,7 @@ jobs:\n<<: *task_e2e_tests\ntimeout: 2h\nparams:\n- KLUSTER_VERSION: \"1.24.6\"\n+ KLUSTER_VERSION: \"1.24.7\"\nKLUSTER_CIDR: \"100.100.0.0/16\"\nISOLATE_TEST: \"true\"\nKLUSTER_OS_IMAGES: flatcar-stable-amd64,flatcar-beta-amd64\n@@ -1647,7 +1647,7 @@ jobs:\nplan:\n- get: kubernikus.builds\nresource: kubernikus.git\n- passed: [e2e-1.24.6]\n+ passed: [e2e-1.24.7]\n- task: changes\nconfig:\n<<: *task_github_compare_url\n@@ -3217,7 +3217,7 @@ groups:\n- whitesource\n- all-versions\n- - e2e-1.24.6\n+ - e2e-1.24.7\n- e2e-1.23.12\n@@ -3355,7 +3355,7 @@ groups:\n- whitesource\n- all-versions\n- - e2e-1.24.6\n+ - e2e-1.24.7\n- e2e-1.23.12\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
Changes pipeline
|
596,226 |
21.10.2022 11:41:14
| -7,200 |
b7fd23d8e952279656e45829f0d1c7cad032adfd
|
Bootstrap ccm roles
|
[
{
"change_type": "MODIFY",
"old_path": "pkg/controller/ground.go",
"new_path": "pkg/controller/ground.go",
"diff": "@@ -28,6 +28,7 @@ import (\nv1 \"github.com/sapcc/kubernikus/pkg/apis/kubernikus/v1\"\n\"github.com/sapcc/kubernikus/pkg/controller/config\"\n\"github.com/sapcc/kubernikus/pkg/controller/ground\"\n+ \"github.com/sapcc/kubernikus/pkg/controller/ground/bootstrap/ccm\"\n\"github.com/sapcc/kubernikus/pkg/controller/ground/bootstrap/csi\"\n\"github.com/sapcc/kubernikus/pkg/controller/ground/bootstrap/network\"\n\"github.com/sapcc/kubernikus/pkg/controller/metrics\"\n@@ -690,6 +691,7 @@ func (op *GroundControl) upgradeKluster(kluster *v1.Kluster, toVersion string) e\nreturn errors.Wrap(err, \"seed cinder CSI roles on upgrade\")\n}\n}\n+\nif !kluster.Spec.NoCloud && strings.HasPrefix(toVersion, \"1.24\") && strings.HasPrefix(kluster.Status.ApiserverVersion, \"1.23\") {\nkubernetes, err := op.Clients.Satellites.ClientFor(kluster)\nif err != nil {\n@@ -701,6 +703,17 @@ func (op *GroundControl) upgradeKluster(kluster *v1.Kluster, toVersion string) e\n}\n}\n+ if !kluster.Spec.NoCloud && strings.HasPrefix(toVersion, \"1.25\") && strings.HasPrefix(kluster.Status.ApiserverVersion, \"1.24\") {\n+ kubernetes, err := op.Clients.Satellites.ClientFor(kluster)\n+ if err != nil {\n+ return errors.Wrap(err, \"client\")\n+ }\n+\n+ if err := ccm.SeedCloudControllerManagerRoles(kubernetes); err != nil {\n+ return errors.Wrap(err, \"seed CCM roles\")\n+ }\n+ }\n+\naccessMode, err := util.PVAccessMode(op.Clients.Kubernetes, kluster)\nif err != nil {\nreturn fmt.Errorf(\"Couldn't determine access mode for pvc: %s\", err)\n"
},
{
"change_type": "MODIFY",
"old_path": "pkg/controller/ground/bootstrap.go",
"new_path": "pkg/controller/ground/bootstrap.go",
"diff": "@@ -15,6 +15,7 @@ import (\nopenstack_project \"github.com/sapcc/kubernikus/pkg/client/openstack/project\"\n\"github.com/sapcc/kubernikus/pkg/controller/config\"\n\"github.com/sapcc/kubernikus/pkg/controller/ground/bootstrap\"\n+ \"github.com/sapcc/kubernikus/pkg/controller/ground/bootstrap/ccm\"\n\"github.com/sapcc/kubernikus/pkg/controller/ground/bootstrap/csi\"\n\"github.com/sapcc/kubernikus/pkg/controller/ground/bootstrap/dns\"\n\"github.com/sapcc/kubernikus/pkg/controller/ground/bootstrap/gpu\"\n@@ -107,6 +108,13 @@ func SeedKluster(clients config.Clients, factories config.Factories, images vers\n}\n}\n}\n+\n+ if ok, _ := util.KlusterVersionConstraint(kluster, \">= 1.25\"); ok {\n+ if err := ccm.SeedCloudControllerManagerRoles(kubernetes); err != nil {\n+ return errors.Wrap(err, \"seed CCM roles\")\n+ }\n+ }\n+\nif !kluster.Spec.NoCloud {\nif ok, _ := util.KlusterVersionConstraint(kluster, \">= 1.24\"); ok {\nif err := network.SeedNetwork(kubernetes, images.Versions[kluster.Spec.Version], *kluster.Spec.ClusterCIDR, kluster.Status.Apiserver, kluster.Spec.AdvertiseAddress, kluster.Spec.AdvertisePort); err != nil {\n"
},
{
"change_type": "ADD",
"old_path": null,
"new_path": "pkg/controller/ground/bootstrap/ccm/ccm.go",
"diff": "+package ccm\n+\n+import (\n+ \"github.com/pkg/errors\"\n+ rbac \"k8s.io/api/rbac/v1\"\n+ \"k8s.io/apimachinery/pkg/runtime/serializer\"\n+ clientset \"k8s.io/client-go/kubernetes\"\n+ clientsetscheme \"k8s.io/client-go/kubernetes/scheme\"\n+\n+ \"github.com/sapcc/kubernikus/pkg/controller/ground/bootstrap\"\n+)\n+\n+func SeedCloudControllerManagerRoles(client clientset.Interface) error {\n+ err := createClusterRole(client, CCMClusterRole)\n+ if err != nil {\n+ return errors.Wrap(err, \"CCMClusterRole\")\n+ }\n+\n+ err = createClusterRole(client, CCMClusterRoleNode)\n+ if err != nil {\n+ return errors.Wrap(err, \"CCMClusterRoleNode\")\n+ }\n+\n+ err = createClusterRoleBinding(client, CCMClusterRoleBinding)\n+ if err != nil {\n+ return errors.Wrap(err, \"CCMClusterRoleBinding\")\n+ }\n+\n+ err = createClusterRoleBinding(client, CCMClusterRoleBindingNode)\n+ if err != nil {\n+ return errors.Wrap(err, \"CCMClusterRoleBindingNode\")\n+ }\n+\n+ return nil\n+}\n+\n+func createClusterRoleBinding(client clientset.Interface, manifest string) error {\n+ template, err := bootstrap.RenderManifest(manifest, nil)\n+ if err != nil {\n+ return err\n+ }\n+\n+ clusterRoleBinding, _, err := serializer.NewCodecFactory(clientsetscheme.Scheme).UniversalDeserializer().Decode(template, nil, &rbac.ClusterRoleBinding{})\n+ if err != nil {\n+ return err\n+ }\n+\n+ if err := bootstrap.CreateOrUpdateClusterRoleBindingV1(client, clusterRoleBinding.(*rbac.ClusterRoleBinding)); err != nil {\n+ return err\n+ }\n+\n+ return nil\n+}\n+\n+func createClusterRole(client clientset.Interface, manifest string) error {\n+ template, err := bootstrap.RenderManifest(manifest, nil)\n+ if err != nil {\n+ return err\n+ }\n+\n+ clusterRole, _, err := serializer.NewCodecFactory(clientsetscheme.Scheme).UniversalDeserializer().Decode(template, nil, &rbac.ClusterRole{})\n+ if err != nil {\n+ return err\n+ }\n+\n+ if err := bootstrap.CreateOrUpdateClusterRoleV1(client, clusterRole.(*rbac.ClusterRole)); err != nil {\n+ return err\n+ }\n+\n+ return nil\n+}\n"
},
{
"change_type": "ADD",
"old_path": null,
"new_path": "pkg/controller/ground/bootstrap/ccm/manifest.go",
"diff": "+package ccm\n+\n+const (\n+ CCMClusterRole = `\n+ apiVersion: rbac.authorization.k8s.io/v1\n+ kind: ClusterRole\n+ metadata:\n+ name: system:cloud-controller-manager\n+ rules:\n+ - apiGroups:\n+ - coordination.k8s.io\n+ resources:\n+ - leases\n+ verbs:\n+ - get\n+ - create\n+ - update\n+ - apiGroups:\n+ - \"\"\n+ resources:\n+ - events\n+ verbs:\n+ - create\n+ - patch\n+ - update\n+ - apiGroups:\n+ - \"\"\n+ resources:\n+ - nodes\n+ verbs:\n+ - '*'\n+ - apiGroups:\n+ - \"\"\n+ resources:\n+ - nodes/status\n+ verbs:\n+ - patch\n+ - apiGroups:\n+ - \"\"\n+ resources:\n+ - services\n+ verbs:\n+ - list\n+ - patch\n+ - update\n+ - watch\n+ - apiGroups:\n+ - \"\"\n+ resources:\n+ - services/status\n+ verbs:\n+ - patch\n+ - apiGroups:\n+ - \"\"\n+ resources:\n+ - serviceaccounts\n+ verbs:\n+ - create\n+ - get\n+ - apiGroups:\n+ - \"\"\n+ resources:\n+ - serviceaccounts/token\n+ verbs:\n+ - create\n+ - apiGroups:\n+ - \"\"\n+ resources:\n+ - persistentvolumes\n+ verbs:\n+ - '*'\n+ - apiGroups:\n+ - \"\"\n+ resources:\n+ - endpoints\n+ verbs:\n+ - create\n+ - get\n+ - list\n+ - watch\n+ - update\n+ - apiGroups:\n+ - \"\"\n+ resources:\n+ - configmaps\n+ verbs:\n+ - get\n+ - list\n+ - watch\n+ - apiGroups:\n+ - \"\"\n+ resources:\n+ - secrets\n+ verbs:\n+ - list\n+ - get\n+ - watch\n+`\n+ CCMClusterRoleNode = `\n+ apiVersion: rbac.authorization.k8s.io/v1\n+ kind: ClusterRole\n+ metadata:\n+ name: system:cloud-node-controller\n+ rules:\n+ - apiGroups:\n+ - \"\"\n+ resources:\n+ - nodes\n+ verbs:\n+ - '*'\n+ - apiGroups:\n+ - \"\"\n+ resources:\n+ - nodes/status\n+ verbs:\n+ - patch\n+ - apiGroups:\n+ - \"\"\n+ resources:\n+ - events\n+ verbs:\n+ - create\n+ - patch\n+ - update\n+`\n+\n+ CCMClusterRoleBinding = `\n+ apiVersion: rbac.authorization.k8s.io/v1\n+ kind: ClusterRoleBinding\n+ metadata:\n+ name: system:cloud-controller-manager\n+ roleRef:\n+ apiGroup: rbac.authorization.k8s.io\n+ kind: ClusterRole\n+ name: system:cloud-controller-manager\n+ subjects:\n+ - kind: ServiceAccount\n+ name: cloud-controller-manager\n+ namespace: kube-system\n+`\n+\n+ CCMClusterRoleBindingNode = `\n+apiVersion: rbac.authorization.k8s.io/v1\n+kind: ClusterRoleBinding\n+metadata:\n+ name: system:cloud-node-controller\n+roleRef:\n+ apiGroup: rbac.authorization.k8s.io\n+ kind: ClusterRole\n+ name: system:cloud-node-controller\n+subjects:\n+- kind: ServiceAccount\n+ name: cloud-node-controller\n+ namespace: kube-system\n+`\n+)\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
Bootstrap ccm roles
|
596,227 |
24.10.2022 16:33:48
| -7,200 |
7089634a8ec100eb912b3b535edfee198ff839ab
|
Make kluster.Spec writable
|
[
{
"change_type": "MODIFY",
"old_path": "pkg/api/handlers/update_cluster.go",
"new_path": "pkg/api/handlers/update_cluster.go",
"diff": "@@ -24,6 +24,11 @@ type updateCluster struct {\n}\nfunc (d *updateCluster) Handle(params operations.UpdateClusterParams, principal *models.Principal) middleware.Responder {\n+ // validate spec.name has some value that makes sense\n+ // as it is read-only on a semantic level.\n+ if !(params.Body.Spec.Name == \"\" || params.Body.Spec.Name == params.Name) {\n+ return NewErrorResponse(&operations.UpdateClusterDefault{}, 500, \"spec.name needs to be removed, an empty string or the clusters name\")\n+ }\nkluster, err := editCluster(d.Kubernikus.KubernikusV1().Klusters(d.Namespace), principal, params.Name, func(kluster *v1.Kluster) error {\n// ensure audit value reaches the spec so it\n"
},
{
"change_type": "MODIFY",
"old_path": "pkg/api/models/kluster_spec.go",
"new_path": "pkg/api/models/kluster_spec.go",
"diff": "@@ -53,7 +53,6 @@ type KlusterSpec struct {\nDNSDomain string `json:\"dnsDomain,omitempty\"`\n// name\n- // Read Only: true\nName string `json:\"name,omitempty\"`\n// no cloud\n@@ -319,10 +318,6 @@ func (m *KlusterSpec) validateVersion(formats strfmt.Registry) error {\nfunc (m *KlusterSpec) ContextValidate(ctx context.Context, formats strfmt.Registry) error {\nvar res []error\n- if err := m.contextValidateName(ctx, formats); err != nil {\n- res = append(res, err)\n- }\n-\nif err := m.contextValidateNodePools(ctx, formats); err != nil {\nres = append(res, err)\n}\n@@ -337,15 +332,6 @@ func (m *KlusterSpec) ContextValidate(ctx context.Context, formats strfmt.Regist\nreturn nil\n}\n-func (m *KlusterSpec) contextValidateName(ctx context.Context, formats strfmt.Registry) error {\n-\n- if err := validate.ReadOnly(ctx, \"name\", \"body\", string(m.Name)); err != nil {\n- return err\n- }\n-\n- return nil\n-}\n-\nfunc (m *KlusterSpec) contextValidateNodePools(ctx context.Context, formats strfmt.Registry) error {\nfor i := 0; i < len(m.NodePools); i++ {\n"
},
{
"change_type": "MODIFY",
"old_path": "pkg/api/spec/embedded_spec.go",
"new_path": "pkg/api/spec/embedded_spec.go",
"diff": "@@ -640,8 +640,7 @@ func init() {\n\"x-nullable\": false\n},\n\"name\": {\n- \"type\": \"string\",\n- \"readOnly\": true\n+ \"type\": \"string\"\n},\n\"noCloud\": {\n\"type\": \"boolean\"\n@@ -1688,8 +1687,7 @@ func init() {\n\"x-nullable\": false\n},\n\"name\": {\n- \"type\": \"string\",\n- \"readOnly\": true\n+ \"type\": \"string\"\n},\n\"noCloud\": {\n\"type\": \"boolean\"\n"
},
{
"change_type": "MODIFY",
"old_path": "swagger.yml",
"new_path": "swagger.yml",
"diff": "@@ -524,7 +524,11 @@ definitions:\ntype: string\nname:\ntype: string\n- readOnly: true\n+ # name is on a semantic level read-only.\n+ # go-swagger did validate that for a long time\n+ # so people depend on setting a value here,\n+ # which equals the cluster name.\n+ # readOnly: true\nbackup:\ntype: string\nx-nullable: false\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
Make kluster.Spec writable
|
596,226 |
17.11.2022 16:03:03
| -3,600 |
173d59827307d9f3479ca5bb0d53bb71c40aa875
|
Remove support for k8s 1.22
|
[
{
"change_type": "MODIFY",
"old_path": "charts/images.yaml",
"new_path": "charts/images.yaml",
"diff": "@@ -880,7 +880,6 @@ imagesForVersion:\nscheduler:\nrepository: keppel.$REGION.cloud.sap/ccloud/kube-scheduler\ntag: v1.22.15\n- supported: true\nwormhole:\nrepository: keppel.$REGION.cloud.sap/ccloud/kubernikus\ntag: changeme\n"
},
{
"change_type": "MODIFY",
"old_path": "ci/pipeline.yaml",
"new_path": "ci/pipeline.yaml",
"diff": "@@ -1641,30 +1641,6 @@ jobs:\n<<: *auth_e2e_master\n<<: *slack\n- - name: e2e-1.22.15\n- serial: true\n- plan:\n- - in_parallel:\n- - get: secrets.git\n- - get: kubernikus.builds\n- resource: kubernikus.git\n- passed: [master]\n- trigger: true\n- - get: all-versions.trigger\n- trigger: true\n- passed: [all-versions]\n- - task: e2e_tests\n- config:\n- <<: *task_e2e_tests\n- timeout: 2h\n- params:\n- KLUSTER_VERSION: \"1.22.15\"\n- KLUSTER_CIDR: \"100.103.0.0/16\"\n- ISOLATE_TEST: \"true\"\n- KLUSTER_OS_IMAGES: flatcar-stable-amd64,flatcar-beta-amd64\n- <<: *auth_e2e_master\n- <<: *slack\n-\n- name: prod\nserial: true\n@@ -3247,8 +3223,6 @@ groups:\n- e2e-1.23.12\n- - e2e-1.22.15\n-\n- prod\n- silver\n@@ -3387,8 +3361,6 @@ groups:\n- e2e-1.23.12\n- - e2e-1.22.15\n-\n- prod\n- silver\n"
},
{
"change_type": "MODIFY",
"old_path": "ci/pipeline.yaml.erb",
"new_path": "ci/pipeline.yaml.erb",
"diff": "VERSIONS = [\n'1.25.3',\n'1.24.7',\n- '1.23.12',\n- '1.22.15'\n+ '1.23.12'\n]\nREGIONS = {\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
Remove support for k8s 1.22 (#712)
|
596,227 |
24.11.2022 14:05:01
| -3,600 |
7c0809ac8c5964e2b698b68be843b67b24f56fc9
|
Unbuffer e2e test output
|
[
{
"change_type": "MODIFY",
"old_path": "ci/task_e2e_tests.yaml",
"new_path": "ci/task_e2e_tests.yaml",
"diff": "@@ -24,10 +24,10 @@ run:\nexport GOPATH=$PWD/gopath\ncd gopath/src/github.com/sapcc/kubernikus\n- apk add --no-progress --no-cache make bash git\n+ apk add --no-progress --no-cache coreutils make bash git\nexport RUN_PARALLEL=false\nexport CGO_ENABLED=0\n- make test-e2e | tee test.output\n+ stdbuf --output=L make test-e2e | tee test.output\nrc=$?\n#Get the longest uniq (!) whitespace prefix of --- FAIL: lines\nif [ $rc -ne 0 ] ; then\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
Unbuffer e2e test output (#722)
|
596,227 |
24.11.2022 15:57:09
| -3,600 |
054a982bd03dcb10a30ca89fd38e34a2258be759
|
Add kubernikusctl darwin arm64 build
|
[
{
"change_type": "MODIFY",
"old_path": "ci/task_cli.yaml",
"new_path": "ci/task_cli.yaml",
"diff": "@@ -24,6 +24,7 @@ run:\napk add --no-cache upx bash make git\nmake bin/darwin/kubernikusctl\n+ make bin/darwin/kubernikusctl.arm64 GOOARCH=arm64\nmake bin/linux/kubernikusctl\nmake bin/windows/kubernikusctl.exe\nupx bin/linux/*\n@@ -35,6 +36,7 @@ run:\necho \"1.0.0+$SHA\" > $BINARIES/tag\ncp bin/darwin/kubernikusctl $BINARIES/kubernikusctl_darwin_amd64\n+ cp bin/darwin/kubernikusctl.arm64 $BINARIES/kubernikusctl_darwin_arm64\ncp bin/linux/kubernikusctl $BINARIES/kubernikusctl_linux_amd64\ncp bin/windows/kubernikusctl.exe $BINARIES/kubernikusctl_windows_amd64.exe\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
Add kubernikusctl darwin arm64 build (#721)
Co-authored-by: Fabian Ruff <fabian.ruff@sap.com>
|
596,226 |
09.12.2022 11:13:08
| -3,600 |
414593e5bed33a4db7b5eb96668cfc41b85dd425
|
Raise kube dns e2e test timeout
|
[
{
"change_type": "MODIFY",
"old_path": "test/e2e/network_test.go",
"new_path": "test/e2e/network_test.go",
"diff": "@@ -22,7 +22,7 @@ import (\nconst (\nTestWaitForPodsRunningTimeout = 5 * time.Minute\n- TestWaitForKubeDNSRunningTimeout = 2 * time.Minute\n+ TestWaitForKubeDNSRunningTimeout = 5 * time.Minute\nTestWaitForServiceEndpointsTimeout = 5 * time.Minute\nTestPodTimeout = 1 * time.Minute\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
Raise kube dns e2e test timeout
|
596,226 |
15.12.2022 16:00:42
| -3,600 |
b5dcaeb9ef7e2d83c1e01524196d2d7f16310768
|
Add general label to csi pods
|
[
{
"change_type": "MODIFY",
"old_path": "charts/kube-master/templates/csi-driver-controller.yaml",
"new_path": "charts/kube-master/templates/csi-driver-controller.yaml",
"diff": "@@ -30,6 +30,7 @@ spec:\n{{- end }}\nlabels:\napp: {{ include \"master.fullname\" . }}-csi\n+ component: csi\nrole: controller\nrelease: {{ .Release.Name }}\nspec:\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
Add general label to csi pods (#726)
|
596,226 |
15.12.2022 16:01:45
| -3,600 |
5bda578db5d57db2d898c976ee38cd92b18f13f5
|
Add apiserver local domain to certificate
|
[
{
"change_type": "MODIFY",
"old_path": "pkg/util/certificates.go",
"new_path": "pkg/util/certificates.go",
"diff": "@@ -278,7 +278,7 @@ func (cf *CertificateFactory) Ensure() ([]CertUpdates, error) {\nreturn nil, err\n}\n- apiServerDNSNames := []string{\"kubernetes\", \"kubernetes.default\", \"kubernetes.default.svc\", \"apiserver\", cf.kluster.Name, fmt.Sprintf(\"%s.%s\", cf.kluster.Name, cf.kluster.Namespace), fmt.Sprintf(\"%v.%v\", cf.kluster.Name, cf.domain)}\n+ apiServerDNSNames := []string{\"kubernetes\", \"kubernetes.default\", \"kubernetes.default.svc\", \"kubernetes.default.svc.cluster.local\", \"apiserver\", cf.kluster.Name, fmt.Sprintf(\"%s.%s\", cf.kluster.Name, cf.kluster.Namespace), fmt.Sprintf(\"%v.%v\", cf.kluster.Name, cf.domain)}\napiServerIPs := []net.IP{net.IPv4(127, 0, 0, 1), apiServiceIP, apiIP}\nif ann := cf.kluster.Annotations[AdditionalApiserverSANsAnnotation]; ann != \"\" {\ndnsNames, ips, err := addtionalSANsFromAnnotation(ann)\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
Add apiserver local domain to certificate (#724)
|
596,226 |
15.12.2022 16:31:49
| -3,600 |
f0fce83908a2efd288f8fe2e5e7059d0820b9656
|
Change k8s version in e2e test
|
[
{
"change_type": "MODIFY",
"old_path": "test/e2e/setup_test.go",
"new_path": "test/e2e/setup_test.go",
"diff": "@@ -40,7 +40,7 @@ func (s *SetupTests) Run(t *testing.T) {\n}\nfunc (s *SetupTests) CreateCluster(t *testing.T) {\n- version := \"1.25.4\"\n+ version := \"1.25.5\"\nif v := os.Getenv(\"KLUSTER_VERSION\"); v != \"\" {\nversion = v\n}\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
Change k8s version in e2e test
|
596,227 |
19.12.2022 16:04:25
| -3,600 |
0bc83f66dd673086db9b44106532f4ecae6102f7
|
More keppel.global
|
[
{
"change_type": "MODIFY",
"old_path": "charts/images.yaml",
"new_path": "charts/images.yaml",
"diff": "@@ -4,19 +4,19 @@ imagesForVersion:\nrepository: keppel.global.cloud.sap/ccloud/dex\ntag: 38f4f8ea8d487470a1dd5b83d66b428d8b502f81\netcd:\n- repository: keppel.$REGION.cloud.sap/ccloud-dockerhub-mirror/sapcc/etcd\n+ repository: keppel.global.cloud.sap/ccloud-dockerhub-mirror/sapcc/etcd\ntag: v3.3.14\netcdBackup:\n- repository: keppel.$REGION.cloud.sap/ccloud-dockerhub-mirror/sapcc/etcdbrctl\n+ repository: keppel.global.cloud.sap/ccloud-dockerhub-mirror/sapcc/etcdbrctl\ntag: 0.5.2\nflannel:\n- repository: keppel.$REGION.cloud.sap/ccloud-quay-mirror/coreos/flannel\n+ repository: keppel.global.cloud.sap/ccloud-quay-mirror/coreos/flannel\ntag: v0.12.0\nfluentd:\nrepository: keppel.global.cloud.sap/ccloud/kubernikus-fluentd\ntag: v1.14-1\nhyperkube:\n- repository: keppel.$REGION.cloud.sap/ccloud-dockerhub-mirror/sapcc/hyperkube\n+ repository: keppel.global.cloud.sap/ccloud-dockerhub-mirror/sapcc/hyperkube\ntag: v1.10.1\nrecycler:\nrepository: keppel.global.cloud.sap/ccloud-k8sgcr-mirror/debian-base\n@@ -29,19 +29,19 @@ imagesForVersion:\nrepository: keppel.global.cloud.sap/ccloud/dex\ntag: 38f4f8ea8d487470a1dd5b83d66b428d8b502f81\netcd:\n- repository: keppel.$REGION.cloud.sap/ccloud-dockerhub-mirror/sapcc/etcd\n+ repository: keppel.global.cloud.sap/ccloud-dockerhub-mirror/sapcc/etcd\ntag: v3.3.14\netcdBackup:\n- repository: keppel.$REGION.cloud.sap/ccloud-dockerhub-mirror/sapcc/etcdbrctl\n+ repository: keppel.global.cloud.sap/ccloud-dockerhub-mirror/sapcc/etcdbrctl\ntag: 0.5.2\nflannel:\n- repository: keppel.$REGION.cloud.sap/ccloud-quay-mirror/coreos/flannel\n+ repository: keppel.global.cloud.sap/ccloud-quay-mirror/coreos/flannel\ntag: v0.12.0\nfluentd:\nrepository: keppel.global.cloud.sap/ccloud/kubernikus-fluentd\ntag: v1.14-1\nhyperkube:\n- repository: keppel.$REGION.cloud.sap/ccloud-dockerhub-mirror/sapcc/hyperkube\n+ repository: keppel.global.cloud.sap/ccloud-dockerhub-mirror/sapcc/hyperkube\ntag: v1.10.11\nrecycler:\nrepository: keppel.global.cloud.sap/ccloud-k8sgcr-mirror/debian-base\n@@ -54,19 +54,19 @@ imagesForVersion:\nrepository: keppel.global.cloud.sap/ccloud/dex\ntag: 38f4f8ea8d487470a1dd5b83d66b428d8b502f81\netcd:\n- repository: keppel.$REGION.cloud.sap/ccloud-dockerhub-mirror/sapcc/etcd\n+ repository: keppel.global.cloud.sap/ccloud-dockerhub-mirror/sapcc/etcd\ntag: v3.3.14\netcdBackup:\n- repository: keppel.$REGION.cloud.sap/ccloud-dockerhub-mirror/sapcc/etcdbrctl\n+ repository: keppel.global.cloud.sap/ccloud-dockerhub-mirror/sapcc/etcdbrctl\ntag: 0.5.2\nflannel:\n- repository: keppel.$REGION.cloud.sap/ccloud-quay-mirror/coreos/flannel\n+ repository: keppel.global.cloud.sap/ccloud-quay-mirror/coreos/flannel\ntag: v0.12.0\nfluentd:\nrepository: keppel.global.cloud.sap/ccloud/kubernikus-fluentd\ntag: v1.14-1\nhyperkube:\n- repository: keppel.$REGION.cloud.sap/ccloud-dockerhub-mirror/sapcc/hyperkube\n+ repository: keppel.global.cloud.sap/ccloud-dockerhub-mirror/sapcc/hyperkube\ntag: v1.10.7\nrecycler:\nrepository: keppel.global.cloud.sap/ccloud-k8sgcr-mirror/debian-base\n@@ -85,19 +85,19 @@ imagesForVersion:\nrepository: keppel.global.cloud.sap/ccloud/dex\ntag: 38f4f8ea8d487470a1dd5b83d66b428d8b502f81\netcd:\n- repository: keppel.$REGION.cloud.sap/ccloud-dockerhub-mirror/sapcc/etcd\n+ repository: keppel.global.cloud.sap/ccloud-dockerhub-mirror/sapcc/etcd\ntag: v3.3.14\netcdBackup:\n- repository: keppel.$REGION.cloud.sap/ccloud-dockerhub-mirror/sapcc/etcdbrctl\n+ repository: keppel.global.cloud.sap/ccloud-dockerhub-mirror/sapcc/etcdbrctl\ntag: 0.5.2\nflannel:\n- repository: keppel.$REGION.cloud.sap/ccloud-quay-mirror/coreos/flannel\n+ repository: keppel.global.cloud.sap/ccloud-quay-mirror/coreos/flannel\ntag: v0.12.0\nfluentd:\nrepository: keppel.global.cloud.sap/ccloud/kubernikus-fluentd\ntag: v1.14-1\nhyperkube:\n- repository: keppel.$REGION.cloud.sap/ccloud-dockerhub-mirror/sapcc/hyperkube\n+ repository: keppel.global.cloud.sap/ccloud-dockerhub-mirror/sapcc/hyperkube\ntag: v1.11.9\nrecycler:\nrepository: keppel.global.cloud.sap/ccloud-k8sgcr-mirror/debian-base\n@@ -116,19 +116,19 @@ imagesForVersion:\nrepository: keppel.global.cloud.sap/ccloud/dex\ntag: 38f4f8ea8d487470a1dd5b83d66b428d8b502f81\netcd:\n- repository: keppel.$REGION.cloud.sap/ccloud-dockerhub-mirror/sapcc/etcd\n+ repository: keppel.global.cloud.sap/ccloud-dockerhub-mirror/sapcc/etcd\ntag: v3.3.14\netcdBackup:\n- repository: keppel.$REGION.cloud.sap/ccloud-dockerhub-mirror/sapcc/etcdbrctl\n+ repository: keppel.global.cloud.sap/ccloud-dockerhub-mirror/sapcc/etcdbrctl\ntag: 0.5.2\nflannel:\n- repository: keppel.$REGION.cloud.sap/ccloud-quay-mirror/coreos/flannel\n+ repository: keppel.global.cloud.sap/ccloud-quay-mirror/coreos/flannel\ntag: v0.12.0\nfluentd:\nrepository: keppel.global.cloud.sap/ccloud/kubernikus-fluentd\ntag: v1.14-1\nhyperkube:\n- repository: keppel.$REGION.cloud.sap/ccloud-dockerhub-mirror/sapcc/hyperkube\n+ repository: keppel.global.cloud.sap/ccloud-dockerhub-mirror/sapcc/hyperkube\ntag: v1.12.10\npause:\nrepository: keppel.global.cloud.sap/ccloud-dockerhub-mirror/sapcc/pause-amd64\n@@ -141,7 +141,7 @@ imagesForVersion:\ntag: changeme\n1.13.9:\ncloudControllerManager:\n- repository: keppel.$REGION.cloud.sap/ccloud-dockerhub-mirror/sapcc/openstack-cloud-controller-manager\n+ repository: keppel.global.cloud.sap/ccloud-dockerhub-mirror/sapcc/openstack-cloud-controller-manager\ntag: v1.13.1\ncoreDNS:\nrepository: keppel.global.cloud.sap/ccloud-dockerhub-mirror/coredns/coredns\n@@ -156,19 +156,19 @@ imagesForVersion:\nrepository: keppel.global.cloud.sap/ccloud/dex\ntag: 38f4f8ea8d487470a1dd5b83d66b428d8b502f81\netcd:\n- repository: keppel.$REGION.cloud.sap/ccloud-dockerhub-mirror/sapcc/etcd\n+ repository: keppel.global.cloud.sap/ccloud-dockerhub-mirror/sapcc/etcd\ntag: v3.3.14\netcdBackup:\n- repository: keppel.$REGION.cloud.sap/ccloud-dockerhub-mirror/sapcc/etcdbrctl\n+ repository: keppel.global.cloud.sap/ccloud-dockerhub-mirror/sapcc/etcdbrctl\ntag: 0.5.2\nflannel:\n- repository: keppel.$REGION.cloud.sap/ccloud-quay-mirror/coreos/flannel\n+ repository: keppel.global.cloud.sap/ccloud-quay-mirror/coreos/flannel\ntag: v0.12.0\nfluentd:\nrepository: keppel.global.cloud.sap/ccloud/kubernikus-fluentd\ntag: v1.14-1\nhyperkube:\n- repository: keppel.$REGION.cloud.sap/ccloud-dockerhub-mirror/sapcc/hyperkube\n+ repository: keppel.global.cloud.sap/ccloud-dockerhub-mirror/sapcc/hyperkube\ntag: v1.13.9\npause:\nrepository: keppel.global.cloud.sap/ccloud-dockerhub-mirror/sapcc/pause-amd64\n@@ -181,7 +181,7 @@ imagesForVersion:\ntag: changeme\n1.14.5:\ncloudControllerManager:\n- repository: keppel.$REGION.cloud.sap/ccloud-dockerhub-mirror/sapcc/openstack-cloud-controller-manager\n+ repository: keppel.global.cloud.sap/ccloud-dockerhub-mirror/sapcc/openstack-cloud-controller-manager\ntag: v1.14.0-sap.0\ncoreDNS:\nrepository: keppel.global.cloud.sap/ccloud-dockerhub-mirror/coredns/coredns\n@@ -196,19 +196,19 @@ imagesForVersion:\nrepository: keppel.global.cloud.sap/ccloud/dex\ntag: 38f4f8ea8d487470a1dd5b83d66b428d8b502f81\netcd:\n- repository: keppel.$REGION.cloud.sap/ccloud-dockerhub-mirror/sapcc/etcd\n+ repository: keppel.global.cloud.sap/ccloud-dockerhub-mirror/sapcc/etcd\ntag: v3.3.14\netcdBackup:\n- repository: keppel.$REGION.cloud.sap/ccloud-dockerhub-mirror/sapcc/etcdbrctl\n+ repository: keppel.global.cloud.sap/ccloud-dockerhub-mirror/sapcc/etcdbrctl\ntag: 0.5.2\nflannel:\n- repository: keppel.$REGION.cloud.sap/ccloud-quay-mirror/coreos/flannel\n+ repository: keppel.global.cloud.sap/ccloud-quay-mirror/coreos/flannel\ntag: v0.12.0\nfluentd:\nrepository: keppel.global.cloud.sap/ccloud/kubernikus-fluentd\ntag: v1.14-1\nhyperkube:\n- repository: keppel.$REGION.cloud.sap/ccloud-dockerhub-mirror/sapcc/hyperkube\n+ repository: keppel.global.cloud.sap/ccloud-dockerhub-mirror/sapcc/hyperkube\ntag: v1.14.5\npause:\nrepository: keppel.global.cloud.sap/ccloud-dockerhub-mirror/sapcc/pause-amd64\n@@ -221,7 +221,7 @@ imagesForVersion:\ntag: changeme\n1.15.2:\ncloudControllerManager:\n- repository: keppel.$REGION.cloud.sap/ccloud-dockerhub-mirror/sapcc/openstack-cloud-controller-manager\n+ repository: keppel.global.cloud.sap/ccloud-dockerhub-mirror/sapcc/openstack-cloud-controller-manager\ntag: v1.15.0-sap.2\ncoreDNS:\nrepository: keppel.global.cloud.sap/ccloud-dockerhub-mirror/coredns/coredns\n@@ -236,19 +236,19 @@ imagesForVersion:\nrepository: keppel.global.cloud.sap/ccloud/dex\ntag: 38f4f8ea8d487470a1dd5b83d66b428d8b502f81\netcd:\n- repository: keppel.$REGION.cloud.sap/ccloud-dockerhub-mirror/sapcc/etcd\n+ repository: keppel.global.cloud.sap/ccloud-dockerhub-mirror/sapcc/etcd\ntag: v3.3.14\netcdBackup:\n- repository: keppel.$REGION.cloud.sap/ccloud-dockerhub-mirror/sapcc/etcdbrctl\n+ repository: keppel.global.cloud.sap/ccloud-dockerhub-mirror/sapcc/etcdbrctl\ntag: 0.5.2\nflannel:\n- repository: keppel.$REGION.cloud.sap/ccloud-quay-mirror/coreos/flannel\n+ repository: keppel.global.cloud.sap/ccloud-quay-mirror/coreos/flannel\ntag: v0.12.0\nfluentd:\nrepository: keppel.global.cloud.sap/ccloud/kubernikus-fluentd\ntag: v1.14-1\nhyperkube:\n- repository: keppel.$REGION.cloud.sap/ccloud-dockerhub-mirror/sapcc/hyperkube\n+ repository: keppel.global.cloud.sap/ccloud-dockerhub-mirror/sapcc/hyperkube\ntag: v1.15.2\npause:\nrepository: keppel.global.cloud.sap/ccloud-dockerhub-mirror/sapcc/pause-amd64\n@@ -261,7 +261,7 @@ imagesForVersion:\ntag: changeme\n1.15.9:\ncloudControllerManager:\n- repository: keppel.$REGION.cloud.sap/ccloud-dockerhub-mirror/sapcc/openstack-cloud-controller-manager\n+ repository: keppel.global.cloud.sap/ccloud-dockerhub-mirror/sapcc/openstack-cloud-controller-manager\ntag: v1.15.0-sap.2\ncoreDNS:\nrepository: keppel.global.cloud.sap/ccloud-dockerhub-mirror/coredns/coredns\n@@ -276,19 +276,19 @@ imagesForVersion:\nrepository: keppel.global.cloud.sap/ccloud/dex\ntag: 38f4f8ea8d487470a1dd5b83d66b428d8b502f81\netcd:\n- repository: keppel.$REGION.cloud.sap/ccloud-dockerhub-mirror/sapcc/etcd\n+ repository: keppel.global.cloud.sap/ccloud-dockerhub-mirror/sapcc/etcd\ntag: v3.3.14\netcdBackup:\n- repository: keppel.$REGION.cloud.sap/ccloud-dockerhub-mirror/sapcc/etcdbrctl\n+ repository: keppel.global.cloud.sap/ccloud-dockerhub-mirror/sapcc/etcdbrctl\ntag: 0.5.2\nflannel:\n- repository: keppel.$REGION.cloud.sap/ccloud-quay-mirror/coreos/flannel\n+ repository: keppel.global.cloud.sap/ccloud-quay-mirror/coreos/flannel\ntag: v0.12.0\nfluentd:\nrepository: keppel.global.cloud.sap/ccloud/kubernikus-fluentd\ntag: v1.14-1\nhyperkube:\n- repository: keppel.$REGION.cloud.sap/ccloud-dockerhub-mirror/sapcc/hyperkube\n+ repository: keppel.global.cloud.sap/ccloud-dockerhub-mirror/sapcc/hyperkube\ntag: v1.15.9\npause:\nrepository: keppel.global.cloud.sap/ccloud-dockerhub-mirror/sapcc/pause-amd64\n@@ -301,7 +301,7 @@ imagesForVersion:\ntag: changeme\n1.16.14:\ncloudControllerManager:\n- repository: keppel.$REGION.cloud.sap/ccloud-dockerhub-mirror/sapcc/openstack-cloud-controller-manager\n+ repository: keppel.global.cloud.sap/ccloud-dockerhub-mirror/sapcc/openstack-cloud-controller-manager\ntag: v1.16.0\ncoreDNS:\nrepository: keppel.global.cloud.sap/ccloud-dockerhub-mirror/coredns/coredns\n@@ -316,19 +316,19 @@ imagesForVersion:\nrepository: keppel.global.cloud.sap/ccloud/dex\ntag: 38f4f8ea8d487470a1dd5b83d66b428d8b502f81\netcd:\n- repository: keppel.$REGION.cloud.sap/ccloud-dockerhub-mirror/sapcc/etcd\n+ repository: keppel.global.cloud.sap/ccloud-dockerhub-mirror/sapcc/etcd\ntag: v3.3.14\netcdBackup:\n- repository: keppel.$REGION.cloud.sap/ccloud-dockerhub-mirror/sapcc/etcdbrctl\n+ repository: keppel.global.cloud.sap/ccloud-dockerhub-mirror/sapcc/etcdbrctl\ntag: 0.5.2\nflannel:\n- repository: keppel.$REGION.cloud.sap/ccloud-quay-mirror/coreos/flannel\n+ repository: keppel.global.cloud.sap/ccloud-quay-mirror/coreos/flannel\ntag: v0.12.0\nfluentd:\nrepository: keppel.global.cloud.sap/ccloud/kubernikus-fluentd\ntag: v1.14-1\nhyperkube:\n- repository: keppel.$REGION.cloud.sap/ccloud-dockerhub-mirror/sapcc/hyperkube\n+ repository: keppel.global.cloud.sap/ccloud-dockerhub-mirror/sapcc/hyperkube\ntag: v1.16.14\npause:\nrepository: keppel.global.cloud.sap/ccloud-dockerhub-mirror/sapcc/pause-amd64\n@@ -341,7 +341,7 @@ imagesForVersion:\ntag: changeme\n1.16.9:\ncloudControllerManager:\n- repository: keppel.$REGION.cloud.sap/ccloud-dockerhub-mirror/sapcc/openstack-cloud-controller-manager\n+ repository: keppel.global.cloud.sap/ccloud-dockerhub-mirror/sapcc/openstack-cloud-controller-manager\ntag: v1.16.0\ncoreDNS:\nrepository: keppel.global.cloud.sap/ccloud-dockerhub-mirror/coredns/coredns\n@@ -356,19 +356,19 @@ imagesForVersion:\nrepository: keppel.global.cloud.sap/ccloud/dex\ntag: 38f4f8ea8d487470a1dd5b83d66b428d8b502f81\netcd:\n- repository: keppel.$REGION.cloud.sap/ccloud-dockerhub-mirror/sapcc/etcd\n+ repository: keppel.global.cloud.sap/ccloud-dockerhub-mirror/sapcc/etcd\ntag: v3.3.14\netcdBackup:\n- repository: keppel.$REGION.cloud.sap/ccloud-dockerhub-mirror/sapcc/etcdbrctl\n+ repository: keppel.global.cloud.sap/ccloud-dockerhub-mirror/sapcc/etcdbrctl\ntag: 0.5.2\nflannel:\n- repository: keppel.$REGION.cloud.sap/ccloud-quay-mirror/coreos/flannel\n+ repository: keppel.global.cloud.sap/ccloud-quay-mirror/coreos/flannel\ntag: v0.12.0\nfluentd:\nrepository: keppel.global.cloud.sap/ccloud/kubernikus-fluentd\ntag: v1.14-1\nhyperkube:\n- repository: keppel.$REGION.cloud.sap/ccloud-dockerhub-mirror/sapcc/hyperkube\n+ repository: keppel.global.cloud.sap/ccloud-dockerhub-mirror/sapcc/hyperkube\ntag: v1.16.9\npause:\nrepository: keppel.global.cloud.sap/ccloud-dockerhub-mirror/sapcc/pause-amd64\n@@ -396,19 +396,19 @@ imagesForVersion:\nrepository: keppel.global.cloud.sap/ccloud/dex\ntag: 38f4f8ea8d487470a1dd5b83d66b428d8b502f81\netcd:\n- repository: keppel.$REGION.cloud.sap/ccloud-dockerhub-mirror/sapcc/etcd\n+ repository: keppel.global.cloud.sap/ccloud-dockerhub-mirror/sapcc/etcd\ntag: v3.3.14\netcdBackup:\n- repository: keppel.$REGION.cloud.sap/ccloud-dockerhub-mirror/sapcc/etcdbrctl\n+ repository: keppel.global.cloud.sap/ccloud-dockerhub-mirror/sapcc/etcdbrctl\ntag: 0.5.2\nflannel:\n- repository: keppel.$REGION.cloud.sap/ccloud-quay-mirror/coreos/flannel\n+ repository: keppel.global.cloud.sap/ccloud-quay-mirror/coreos/flannel\ntag: v0.12.0\nfluentd:\nrepository: keppel.global.cloud.sap/ccloud/kubernikus-fluentd\ntag: v1.14-1\nhyperkube:\n- repository: keppel.$REGION.cloud.sap/ccloud-dockerhub-mirror/sapcc/hyperkube\n+ repository: keppel.global.cloud.sap/ccloud-dockerhub-mirror/sapcc/hyperkube\ntag: v1.17.13\npause:\nrepository: keppel.global.cloud.sap/ccloud-dockerhub-mirror/sapcc/pause-amd64\n@@ -436,19 +436,19 @@ imagesForVersion:\nrepository: keppel.global.cloud.sap/ccloud/dex\ntag: 38f4f8ea8d487470a1dd5b83d66b428d8b502f81\netcd:\n- repository: keppel.$REGION.cloud.sap/ccloud-dockerhub-mirror/sapcc/etcd\n+ repository: keppel.global.cloud.sap/ccloud-dockerhub-mirror/sapcc/etcd\ntag: v3.3.14\netcdBackup:\n- repository: keppel.$REGION.cloud.sap/ccloud-dockerhub-mirror/sapcc/etcdbrctl\n+ repository: keppel.global.cloud.sap/ccloud-dockerhub-mirror/sapcc/etcdbrctl\ntag: 0.5.2\nflannel:\n- repository: keppel.$REGION.cloud.sap/ccloud-quay-mirror/coreos/flannel\n+ repository: keppel.global.cloud.sap/ccloud-quay-mirror/coreos/flannel\ntag: v0.12.0\nfluentd:\nrepository: keppel.global.cloud.sap/ccloud/kubernikus-fluentd\ntag: v1.14-1\nhyperkube:\n- repository: keppel.$REGION.cloud.sap/ccloud-dockerhub-mirror/sapcc/hyperkube\n+ repository: keppel.global.cloud.sap/ccloud-dockerhub-mirror/sapcc/hyperkube\ntag: v1.18.10\npause:\nrepository: keppel.global.cloud.sap/ccloud-dockerhub-mirror/sapcc/pause-amd64\n@@ -482,13 +482,13 @@ imagesForVersion:\nrepository: keppel.global.cloud.sap/ccloud/dex\ntag: 38f4f8ea8d487470a1dd5b83d66b428d8b502f81\netcd:\n- repository: keppel.$REGION.cloud.sap/ccloud-dockerhub-mirror/sapcc/etcd\n+ repository: keppel.global.cloud.sap/ccloud-dockerhub-mirror/sapcc/etcd\ntag: v3.3.14\netcdBackup:\n- repository: keppel.$REGION.cloud.sap/ccloud-dockerhub-mirror/sapcc/etcdbrctl\n+ repository: keppel.global.cloud.sap/ccloud-dockerhub-mirror/sapcc/etcdbrctl\ntag: 0.5.2\nflannel:\n- repository: keppel.$REGION.cloud.sap/ccloud-quay-mirror/coreos/flannel\n+ repository: keppel.global.cloud.sap/ccloud-quay-mirror/coreos/flannel\ntag: v0.12.0\nfluentd:\nrepository: keppel.global.cloud.sap/ccloud/kubernikus-fluentd\n@@ -534,13 +534,13 @@ imagesForVersion:\nrepository: keppel.global.cloud.sap/ccloud/dex\ntag: 38f4f8ea8d487470a1dd5b83d66b428d8b502f81\netcd:\n- repository: keppel.$REGION.cloud.sap/ccloud-dockerhub-mirror/sapcc/etcd\n+ repository: keppel.global.cloud.sap/ccloud-dockerhub-mirror/sapcc/etcd\ntag: v3.3.14\netcdBackup:\n- repository: keppel.$REGION.cloud.sap/ccloud-dockerhub-mirror/sapcc/etcdbrctl\n+ repository: keppel.global.cloud.sap/ccloud-dockerhub-mirror/sapcc/etcdbrctl\ntag: 0.5.2\nflannel:\n- repository: keppel.$REGION.cloud.sap/ccloud-quay-mirror/coreos/flannel\n+ repository: keppel.global.cloud.sap/ccloud-quay-mirror/coreos/flannel\ntag: v0.12.0\nfluentd:\nrepository: keppel.global.cloud.sap/ccloud/kubernikus-fluentd\n@@ -610,13 +610,13 @@ imagesForVersion:\nrepository: keppel.global.cloud.sap/ccloud/dex\ntag: 38f4f8ea8d487470a1dd5b83d66b428d8b502f81\netcd:\n- repository: keppel.$REGION.cloud.sap/ccloud-dockerhub-mirror/sapcc/etcd\n+ repository: keppel.global.cloud.sap/ccloud-dockerhub-mirror/sapcc/etcd\ntag: v3.3.14\netcdBackup:\n- repository: keppel.$REGION.cloud.sap/ccloud-dockerhub-mirror/sapcc/etcdbrctl\n+ repository: keppel.global.cloud.sap/ccloud-dockerhub-mirror/sapcc/etcdbrctl\ntag: 0.5.2\nflannel:\n- repository: keppel.$REGION.cloud.sap/ccloud-quay-mirror/coreos/flannel\n+ repository: keppel.global.cloud.sap/ccloud-quay-mirror/coreos/flannel\ntag: v0.12.0\nfluentd:\nrepository: keppel.global.cloud.sap/ccloud/kubernikus-fluentd\n@@ -686,13 +686,13 @@ imagesForVersion:\nrepository: keppel.global.cloud.sap/ccloud/dex\ntag: 38f4f8ea8d487470a1dd5b83d66b428d8b502f81\netcd:\n- repository: keppel.$REGION.cloud.sap/ccloud-dockerhub-mirror/sapcc/etcd\n+ repository: keppel.global.cloud.sap/ccloud-dockerhub-mirror/sapcc/etcd\ntag: v3.3.14\netcdBackup:\n- repository: keppel.$REGION.cloud.sap/ccloud-dockerhub-mirror/sapcc/etcdbrctl\n+ repository: keppel.global.cloud.sap/ccloud-dockerhub-mirror/sapcc/etcdbrctl\ntag: 0.5.2\nflannel:\n- repository: keppel.$REGION.cloud.sap/ccloud-quay-mirror/coreos/flannel\n+ repository: keppel.global.cloud.sap/ccloud-quay-mirror/coreos/flannel\ntag: v0.12.0\nfluentd:\nrepository: keppel.global.cloud.sap/ccloud/kubernikus-fluentd\n@@ -762,13 +762,13 @@ imagesForVersion:\nrepository: keppel.global.cloud.sap/ccloud/dex\ntag: 38f4f8ea8d487470a1dd5b83d66b428d8b502f81\netcd:\n- repository: keppel.$REGION.cloud.sap/ccloud-dockerhub-mirror/sapcc/etcd\n+ repository: keppel.global.cloud.sap/ccloud-dockerhub-mirror/sapcc/etcd\ntag: v3.3.14\netcdBackup:\n- repository: keppel.$REGION.cloud.sap/ccloud-dockerhub-mirror/sapcc/etcdbrctl\n+ repository: keppel.global.cloud.sap/ccloud-dockerhub-mirror/sapcc/etcdbrctl\ntag: 0.5.2\nflannel:\n- repository: keppel.$REGION.cloud.sap/ccloud-quay-mirror/coreos/flannel\n+ repository: keppel.global.cloud.sap/ccloud-quay-mirror/coreos/flannel\ntag: v0.12.0\nfluentd:\nrepository: keppel.global.cloud.sap/ccloud/kubernikus-fluentd\n@@ -838,13 +838,13 @@ imagesForVersion:\nrepository: keppel.global.cloud.sap/ccloud/dex\ntag: 38f4f8ea8d487470a1dd5b83d66b428d8b502f81\netcd:\n- repository: keppel.$REGION.cloud.sap/ccloud-dockerhub-mirror/sapcc/etcd\n+ repository: keppel.global.cloud.sap/ccloud-dockerhub-mirror/sapcc/etcd\ntag: v3.3.14\netcdBackup:\n- repository: keppel.$REGION.cloud.sap/ccloud-dockerhub-mirror/sapcc/etcdbrctl\n+ repository: keppel.global.cloud.sap/ccloud-dockerhub-mirror/sapcc/etcdbrctl\ntag: 0.5.2\nflannel:\n- repository: keppel.$REGION.cloud.sap/ccloud-quay-mirror/coreos/flannel\n+ repository: keppel.global.cloud.sap/ccloud-quay-mirror/coreos/flannel\ntag: v0.12.0\nfluentd:\nrepository: keppel.global.cloud.sap/ccloud/kubernikus-fluentd\n@@ -914,13 +914,13 @@ imagesForVersion:\nrepository: keppel.global.cloud.sap/ccloud/dex\ntag: 38f4f8ea8d487470a1dd5b83d66b428d8b502f81\netcd:\n- repository: keppel.$REGION.cloud.sap/ccloud-dockerhub-mirror/sapcc/etcd\n+ repository: keppel.global.cloud.sap/ccloud-dockerhub-mirror/sapcc/etcd\ntag: v3.3.14\netcdBackup:\n- repository: keppel.$REGION.cloud.sap/ccloud-dockerhub-mirror/sapcc/etcdbrctl\n+ repository: keppel.global.cloud.sap/ccloud-dockerhub-mirror/sapcc/etcdbrctl\ntag: 0.5.2\nflannel:\n- repository: keppel.$REGION.cloud.sap/ccloud-quay-mirror/coreos/flannel\n+ repository: keppel.global.cloud.sap/ccloud-quay-mirror/coreos/flannel\ntag: v0.12.0\nfluentd:\nrepository: keppel.global.cloud.sap/ccloud/kubernikus-fluentd\n@@ -990,13 +990,13 @@ imagesForVersion:\nrepository: keppel.global.cloud.sap/ccloud/dex\ntag: 38f4f8ea8d487470a1dd5b83d66b428d8b502f81\netcd:\n- repository: keppel.$REGION.cloud.sap/ccloud-dockerhub-mirror/sapcc/etcd\n+ repository: keppel.global.cloud.sap/ccloud-dockerhub-mirror/sapcc/etcd\ntag: v3.3.14\netcdBackup:\n- repository: keppel.$REGION.cloud.sap/ccloud-dockerhub-mirror/sapcc/etcdbrctl\n+ repository: keppel.global.cloud.sap/ccloud-dockerhub-mirror/sapcc/etcdbrctl\ntag: 0.5.2\nflannel:\n- repository: keppel.$REGION.cloud.sap/ccloud-quay-mirror/coreos/flannel\n+ repository: keppel.global.cloud.sap/ccloud-quay-mirror/coreos/flannel\ntag: v0.12.0\nfluentd:\nrepository: keppel.global.cloud.sap/ccloud/kubernikus-fluentd\n@@ -1248,81 +1248,81 @@ imagesForVersion:\ntag: v1.23.14\nsupported: false\nwormhole:\n- repository: keppel.$REGION.cloud.sap/ccloud/kubernikus\n+ repository: keppel.global.cloud.sap/ccloud/kubernikus\ntag: changeme\n1.23.15:\napiserver:\n- repository: keppel.$REGION.cloud.sap/ccloud/kube-apiserver\n+ repository: keppel.global.cloud.sap/ccloud/kube-apiserver\ntag: v1.23.15\ncinderCSIPlugin:\n- repository: keppel.$REGION.cloud.sap/ccloud/cinder-csi-plugin\n+ repository: keppel.global.cloud.sap/ccloud-dockerhub-mirror/k8scloudprovider/cinder-csi-plugin\ntag: v1.23.0\ncloudControllerManager:\n- repository: keppel.$REGION.cloud.sap/ccloud/openstack-cloud-controller-manager\n+ repository: keppel.global.cloud.sap/ccloud-dockerhub-mirror/k8scloudprovider/openstack-cloud-controller-manager\ntag: v1.23.4\ncontrollerManager:\n- repository: keppel.$REGION.cloud.sap/ccloud/kube-controller-manager\n+ repository: keppel.global.cloud.sap/ccloud-registry-k8s-io-mirror/kube-controller-manager\ntag: v1.23.15\ncoreDNS:\n- repository: keppel.$REGION.cloud.sap/ccloud-dockerhub-mirror/coredns/coredns\n+ repository: keppel.global.cloud.sap/ccloud-dockerhub-mirror/coredns/coredns\ntag: 1.9.1\ncsiAttacher:\n- repository: keppel.$REGION.cloud.sap/ccloud/csi-attacher\n+ repository: keppel.global.cloud.sap/ccloud-registry-k8s-io-mirror/sig-storage/csi-attacher\ntag: v3.4.0\ncsiLivenessProbe:\n- repository: keppel.$REGION.cloud.sap/ccloud/csi-livenessprobe\n+ repository: keppel.global.cloud.sap/ccloud-registry-k8s-io-mirror/sig-storage/livenessprobe\ntag: v2.6.0\ncsiNodeDriver:\n- repository: keppel.$REGION.cloud.sap/ccloud/csi-node-driver-registrar\n+ repository: keppel.global.cloud.sap/ccloud-registry-k8s-io-mirror/sig-storage/csi-node-driver-registrar\ntag: v2.5.0\ncsiProvisioner:\n- repository: keppel.$REGION.cloud.sap/ccloud/csi-provisioner\n+ repository: keppel.global.cloud.sap/ccloud-registry-k8s-io-mirror/sig-storage/csi-provisioner\ntag: v3.1.0\ncsiResizer:\n- repository: keppel.$REGION.cloud.sap/ccloud/csi-resizer\n+ repository: keppel.global.cloud.sap/ccloud-registry-k8s-io-mirror/sig-storage/csi-resizer\ntag: v1.4.0\ncsiSnapshotController:\n- repository: keppel.$REGION.cloud.sap/ccloud/csi-snapshot-controller\n+ repository: keppel.global.cloud.sap/ccloud-registry-k8s-io-mirror/sig-storage/snapshot-controller\ntag: v5.0.1\ncsiSnapshotter:\n- repository: keppel.$REGION.cloud.sap/ccloud/csi-snapshotter\n+ repository: keppel.global.cloud.sap/ccloud-registry-k8s-io-mirror/sig-storage/csi-snapshotter\ntag: v5.0.1\ndashboard:\n- repository: keppel.$REGION.cloud.sap/ccloud-dockerhub-mirror/kubernetesui/dashboard\n+ repository: keppel.global.cloud.sap/ccloud-dockerhub-mirror/kubernetesui/dashboard\ntag: v2.0.4\ndashboardProxy:\n- repository: keppel.$REGION.cloud.sap/ccloud/keycloak-gatekeeper\n+ repository: keppel.global.cloud.sap/ccloud/keycloak-gatekeeper\ntag: 6.0.1\ndefault: true\ndex:\n- repository: keppel.$REGION.cloud.sap/ccloud/dex\n+ repository: keppel.global.cloud.sap/ccloud/dex\ntag: 38f4f8ea8d487470a1dd5b83d66b428d8b502f81\netcd:\n- repository: keppel.$REGION.cloud.sap/ccloud/etcd\n+ repository: keppel.global.cloud.sap/ccloud/etcd\ntag: v3.4.13-bootstrap-3\netcdBackup:\n- repository: keppel.$REGION.cloud.sap/ccloud/etcdbrctl\n+ repository: keppel.global.cloud.sap/ccloud/etcdbrctl\ntag: v0.15.4\nflannel:\n- repository: keppel.$REGION.cloud.sap/ccloud-dockerhub-mirror/flannelcni/flannel\n+ repository: keppel.global.cloud.sap/ccloud-dockerhub-mirror/flannelcni/flannel\ntag: v0.17.0\nfluentd:\n- repository: keppel.$REGION.cloud.sap/ccloud/kubernikus-fluentd\n+ repository: keppel.global.cloud.sap/ccloud/kubernikus-fluentd\ntag: v1.14.6-1.1\nkubeProxy:\n- repository: keppel.$REGION.cloud.sap/ccloud/kube-proxy\n+ repository: keppel.global.cloud.sap/ccloud-registry-k8s-io-mirror/kube-proxy\ntag: v1.23.15\nkubelet:\n- repository: keppel.$REGION.cloud.sap/ccloud/kubelet\n+ repository: keppel.global.cloud.sap/ccloud/kubelet\ntag: v1.23.15\npause:\n- repository: keppel.$REGION.cloud.sap/ccloud-dockerhub-mirror/sapcc/pause-amd64\n+ repository: keppel.global.cloud.sap/ccloud-dockerhub-mirror/sapcc/pause-amd64\ntag: '3.1'\nrecycler:\nrepository: keppel.global.cloud.sap/ccloud-k8sgcr-mirror/debian-base\ntag: v2.0.0\nscheduler:\n- repository: keppel.$REGION.cloud.sap/ccloud/kube-scheduler\n+ repository: keppel.global.cloud.sap/ccloud-registry-k8s-io-mirror/kube-scheduler\ntag: v1.23.15\nsupported: true\nwormhole:\n@@ -1818,86 +1818,86 @@ imagesForVersion:\ntag: v1.24.8\nsupported: false\nwormhole:\n- repository: keppel.$REGION.cloud.sap/ccloud/kubernikus\n+ repository: keppel.global.cloud.sap/ccloud/kubernikus\ntag: changeme\n1.24.9:\napiserver:\n- repository: keppel.$REGION.cloud.sap/ccloud/kube-apiserver\n+ repository: keppel.global.cloud.sap/ccloud/kube-apiserver\ntag: v1.24.9\ncinderCSIPlugin:\n- repository: keppel.$REGION.cloud.sap/ccloud/cinder-csi-plugin\n+ repository: keppel.global.cloud.sap/ccloud-dockerhub-mirror/k8scloudprovider/cinder-csi-plugin\ntag: v1.24.5\ncloudControllerManager:\n- repository: keppel.$REGION.cloud.sap/ccloud/openstack-cloud-controller-manager\n+ repository: keppel.global.cloud.sap/ccloud-dockerhub-mirror/k8scloudprovider/openstack-cloud-controller-manager\ntag: v1.24.5\ncniPlugins:\n- repository: keppel.$REGION.cloud.sap/ccloud/cni-plugins\n+ repository: keppel.global.cloud.sap/ccloud/cni-plugins\ntag: v1.1.1\ncontrollerManager:\n- repository: keppel.$REGION.cloud.sap/ccloud/kube-controller-manager\n+ repository: keppel.global.cloud.sap/ccloud-registry-k8s-io-mirror/kube-controller-manager\ntag: v1.24.9\ncoreDNS:\n- repository: keppel.$REGION.cloud.sap/ccloud-dockerhub-mirror/coredns/coredns\n+ repository: keppel.global.cloud.sap/ccloud-dockerhub-mirror/coredns/coredns\ntag: 1.9.1\ncsiAttacher:\n- repository: keppel.$REGION.cloud.sap/ccloud/csi-attacher\n+ repository: keppel.global.cloud.sap/ccloud-registry-k8s-io-mirror/sig-storage/csi-attacher\ntag: v3.5.0\ncsiLivenessProbe:\n- repository: keppel.$REGION.cloud.sap/ccloud/csi-livenessprobe\n+ repository: keppel.global.cloud.sap/ccloud-registry-k8s-io-mirror/sig-storage/livenessprobe\ntag: v2.7.0\ncsiNodeDriver:\n- repository: keppel.$REGION.cloud.sap/ccloud/csi-node-driver-registrar\n+ repository: keppel.global.cloud.sap/ccloud-registry-k8s-io-mirror/sig-storage/csi-node-driver-registrar\ntag: v2.5.1\ncsiProvisioner:\n- repository: keppel.$REGION.cloud.sap/ccloud/csi-provisioner\n+ repository: keppel.global.cloud.sap/ccloud-registry-k8s-io-mirror/sig-storage/csi-provisioner\ntag: v3.2.1\ncsiResizer:\n- repository: keppel.$REGION.cloud.sap/ccloud/csi-resizer\n+ repository: keppel.global.cloud.sap/ccloud-registry-k8s-io-mirror/sig-storage/csi-resizer\ntag: v1.5.0\ncsiSnapshotController:\n- repository: keppel.$REGION.cloud.sap/ccloud/csi-snapshot-controller\n+ repository: keppel.global.cloud.sap/ccloud-registry-k8s-io-mirror/sig-storage/snapshot-controller\ntag: v5.0.1\ncsiSnapshotter:\n- repository: keppel.$REGION.cloud.sap/ccloud/csi-snapshotter\n+ repository: keppel.global.cloud.sap/ccloud-registry-k8s-io-mirror/sig-storage/csi-snapshotter\ntag: v5.0.1\ndashboard:\n- repository: keppel.$REGION.cloud.sap/ccloud-dockerhub-mirror/kubernetesui/dashboard\n+ repository: keppel.global.cloud.sap/ccloud-dockerhub-mirror/kubernetesui/dashboard\ntag: v2.0.4\ndashboardProxy:\n- repository: keppel.$REGION.cloud.sap/ccloud/keycloak-gatekeeper\n+ repository: keppel.global.cloud.sap/ccloud/keycloak-gatekeeper\ntag: 6.0.1\ndex:\n- repository: keppel.$REGION.cloud.sap/ccloud/dex\n+ repository: keppel.global.cloud.sap/ccloud/dex\ntag: 38f4f8ea8d487470a1dd5b83d66b428d8b502f81\netcd:\n- repository: keppel.$REGION.cloud.sap/ccloud/etcd\n+ repository: keppel.global.cloud.sap/ccloud/etcd\ntag: v3.4.13-bootstrap-3\netcdBackup:\n- repository: keppel.$REGION.cloud.sap/ccloud/etcdbrctl\n+ repository: keppel.global.cloud.sap/ccloud/etcdbrctl\ntag: v0.15.4\nflannel:\n- repository: keppel.$REGION.cloud.sap/ccloud-dockerhub-mirror/flannelcni/flannel\n+ repository: keppel.global.cloud.sap/ccloud-dockerhub-mirror/flannelcni/flannel\ntag: v0.19.1\nflannelCNIPlugin:\n- repository: keppel.$REGION.cloud.sap/ccloud-dockerhub-mirror/flannelcni/flannel-cni-plugin\n+ repository: keppel.global.cloud.sap/ccloud-dockerhub-mirror/flannelcni/flannel-cni-plugin\ntag: v1.1.0\nfluentd:\n- repository: keppel.$REGION.cloud.sap/ccloud/kubernikus-fluentd\n+ repository: keppel.global.cloud.sap/ccloud/kubernikus-fluentd\ntag: v1.14.6-1.1\nkubeProxy:\n- repository: keppel.$REGION.cloud.sap/ccloud/kube-proxy\n+ repository: keppel.global.cloud.sap/ccloud-registry-k8s-io-mirror/kube-proxy\ntag: v1.24.9\nkubelet:\n- repository: keppel.$REGION.cloud.sap/ccloud/kubelet\n+ repository: keppel.global.cloud.sap/ccloud/kubelet\ntag: v1.24.9\npause:\n- repository: keppel.$REGION.cloud.sap/ccloud-dockerhub-mirror/sapcc/pause-amd64\n+ repository: keppel.global.cloud.sap/ccloud-dockerhub-mirror/sapcc/pause-amd64\ntag: '3.1'\nrecycler:\nrepository: keppel.global.cloud.sap/ccloud-k8sgcr-mirror/debian-base\ntag: v2.0.0\nscheduler:\n- repository: keppel.$REGION.cloud.sap/ccloud/kube-scheduler\n+ repository: keppel.global.cloud.sap/ccloud-registry-k8s-io-mirror/kube-scheduler\ntag: v1.24.9\nsupported: true\nwormhole:\n@@ -2003,7 +2003,7 @@ imagesForVersion:\nrepository: keppel.global.cloud.sap/ccloud-registry-k8s-io-mirror/kube-controller-manager\ntag: v1.25.4\ncoreDNS:\n- repository: keppel.$REGION.cloud.sap/ccloud-dockerhub-mirror/coredns/coredns\n+ repository: keppel.global.cloud.sap/ccloud-dockerhub-mirror/coredns/coredns\ntag: 1.9.1\ncsiAttacher:\nrepository: keppel.global.cloud.sap/ccloud-registry-k8s-io-mirror/sig-storage/csi-attacher\n@@ -2067,86 +2067,86 @@ imagesForVersion:\ntag: v1.25.4\nsupported: false\nwormhole:\n- repository: keppel.$REGION.cloud.sap/ccloud/kubernikus\n+ repository: keppel.global.cloud.sap/ccloud/kubernikus\ntag: changeme\n1.25.5:\napiserver:\n- repository: keppel.$REGION.cloud.sap/ccloud/kube-apiserver\n+ repository: keppel.global.cloud.sap/ccloud/kube-apiserver\ntag: v1.25.5\ncinderCSIPlugin:\n- repository: keppel.$REGION.cloud.sap/ccloud/cinder-csi-plugin\n+ repository: keppel.global.cloud.sap/ccloud-dockerhub-mirror/k8scloudprovider/cinder-csi-plugin\ntag: v1.25.3\ncloudControllerManager:\n- repository: keppel.$REGION.cloud.sap/ccloud/openstack-cloud-controller-manager\n+ repository: keppel.global.cloud.sap/ccloud-dockerhub-mirror/k8scloudprovider/openstack-cloud-controller-manager\ntag: v1.25.3\ncniPlugins:\n- repository: keppel.$REGION.cloud.sap/ccloud/cni-plugins\n+ repository: keppel.global.cloud.sap/ccloud/cni-plugins\ntag: v1.1.1\ncontrollerManager:\n- repository: keppel.$REGION.cloud.sap/ccloud/kube-controller-manager\n+ repository: keppel.global.cloud.sap/ccloud-registry-k8s-io-mirror/kube-controller-manager\ntag: v1.25.5\ncoreDNS:\n- repository: keppel.$REGION.cloud.sap/ccloud-dockerhub-mirror/coredns/coredns\n+ repository: keppel.global.cloud.sap/ccloud-dockerhub-mirror/coredns/coredns\ntag: 1.9.1\ncsiAttacher:\n- repository: keppel.$REGION.cloud.sap/ccloud/csi-attacher\n+ repository: keppel.global.cloud.sap/ccloud-registry-k8s-io-mirror/sig-storage/csi-attacher\ntag: v3.5.0\ncsiLivenessProbe:\n- repository: keppel.$REGION.cloud.sap/ccloud/csi-livenessprobe\n+ repository: keppel.global.cloud.sap/ccloud-registry-k8s-io-mirror/sig-storage/livenessprobe\ntag: v2.7.0\ncsiNodeDriver:\n- repository: keppel.$REGION.cloud.sap/ccloud/csi-node-driver-registrar\n+ repository: keppel.global.cloud.sap/ccloud-registry-k8s-io-mirror/sig-storage/csi-node-driver-registrar\ntag: v2.5.1\ncsiProvisioner:\n- repository: keppel.$REGION.cloud.sap/ccloud/csi-provisioner\n+ repository: keppel.global.cloud.sap/ccloud-registry-k8s-io-mirror/sig-storage/csi-provisioner\ntag: v3.2.1\ncsiResizer:\n- repository: keppel.$REGION.cloud.sap/ccloud/csi-resizer\n+ repository: keppel.global.cloud.sap/ccloud-registry-k8s-io-mirror/sig-storage/csi-resizer\ntag: v1.5.0\ncsiSnapshotController:\n- repository: keppel.$REGION.cloud.sap/ccloud/csi-snapshot-controller\n+ repository: keppel.global.cloud.sap/ccloud-registry-k8s-io-mirror/sig-storage/snapshot-controller\ntag: v5.0.1\ncsiSnapshotter:\n- repository: keppel.$REGION.cloud.sap/ccloud/csi-snapshotter\n+ repository: keppel.global.cloud.sap/ccloud-registry-k8s-io-mirror/sig-storage/csi-snapshotter\ntag: v5.0.1\ndashboard:\n- repository: keppel.$REGION.cloud.sap/ccloud-dockerhub-mirror/kubernetesui/dashboard\n+ repository: keppel.global.cloud.sap/ccloud-dockerhub-mirror/kubernetesui/dashboard\ntag: v2.0.4\ndashboardProxy:\n- repository: keppel.$REGION.cloud.sap/ccloud/keycloak-gatekeeper\n+ repository: keppel.global.cloud.sap/ccloud/keycloak-gatekeeper\ntag: 6.0.1\ndex:\n- repository: keppel.$REGION.cloud.sap/ccloud/dex\n+ repository: keppel.global.cloud.sap/ccloud/dex\ntag: 38f4f8ea8d487470a1dd5b83d66b428d8b502f81\netcd:\n- repository: keppel.$REGION.cloud.sap/ccloud/etcd\n+ repository: keppel.global.cloud.sap/ccloud/etcd\ntag: v3.4.13-bootstrap-3\netcdBackup:\n- repository: keppel.$REGION.cloud.sap/ccloud/etcdbrctl\n+ repository: keppel.global.cloud.sap/ccloud/etcdbrctl\ntag: v0.15.4\nflannel:\n- repository: keppel.$REGION.cloud.sap/ccloud-dockerhub-mirror/flannelcni/flannel\n+ repository: keppel.global.cloud.sap/ccloud-dockerhub-mirror/flannelcni/flannel\ntag: v0.19.1\nflannelCNIPlugin:\n- repository: keppel.$REGION.cloud.sap/ccloud-dockerhub-mirror/flannelcni/flannel-cni-plugin\n+ repository: keppel.global.cloud.sap/ccloud-dockerhub-mirror/flannelcni/flannel-cni-plugin\ntag: v1.1.0\nfluentd:\n- repository: keppel.$REGION.cloud.sap/ccloud/kubernikus-fluentd\n+ repository: keppel.global.cloud.sap/ccloud/kubernikus-fluentd\ntag: v1.14.6-1.1\nkubeProxy:\n- repository: keppel.$REGION.cloud.sap/ccloud/kube-proxy\n+ repository: keppel.global.cloud.sap/ccloud-registry-k8s-io-mirror/kube-proxy\ntag: v1.25.5\nkubelet:\n- repository: keppel.$REGION.cloud.sap/ccloud/kubelet\n+ repository: keppel.global.cloud.sap/ccloud/kubelet\ntag: v1.25.5\npause:\n- repository: keppel.$REGION.cloud.sap/ccloud-dockerhub-mirror/sapcc/pause-amd64\n+ repository: keppel.global.cloud.sap/ccloud-dockerhub-mirror/sapcc/pause-amd64\ntag: '3.1'\nrecycler:\nrepository: keppel.global.cloud.sap/ccloud-k8sgcr-mirror/debian-base\ntag: v2.0.0\nscheduler:\n- repository: keppel.$REGION.cloud.sap/ccloud/kube-scheduler\n+ repository: keppel.global.cloud.sap/ccloud-registry-k8s-io-mirror/kube-scheduler\ntag: v1.25.5\nsupported: true\nwormhole:\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
More keppel.global (#735)
|
596,227 |
05.01.2023 18:51:34
| -3,600 |
dc3973d8985ce2905e64448f6077ca792ea2115e
|
Use --chart-directory flag to find seed chart
|
[
{
"change_type": "MODIFY",
"old_path": "pkg/controller/ground.go",
"new_path": "pkg/controller/ground.go",
"diff": "@@ -304,7 +304,7 @@ func (op *GroundControl) handler(key string) error {\nif err != nil {\nreturn err\n}\n- if err := seedReconciler.ReconcileSeeding(helmValues); err != nil {\n+ if err := seedReconciler.ReconcileSeeding(path.Join(op.Config.Helm.ChartDirectory, \"seed\"), helmValues); err != nil {\nmetrics.SeedReconciliationFailuresTotal.With(prometheus.Labels{\"kluster_name\": kluster.Spec.Name}).Inc()\nreturn fmt.Errorf(\"Seeding reconciliation failed: %w\", err)\n}\n@@ -355,7 +355,7 @@ func (op *GroundControl) handler(key string) error {\nmetrics.SeedReconciliationFailuresTotal.With(prometheus.Labels{\"kluster_name\": kluster.Spec.Name}).Inc()\nreturn err\n}\n- if err := seedReconciler.ReconcileSeeding(helmValues); err != nil {\n+ if err := seedReconciler.ReconcileSeeding(path.Join(op.Config.Helm.ChartDirectory, \"seed\"), helmValues); err != nil {\nmetrics.SeedReconciliationFailuresTotal.With(prometheus.Labels{\"kluster_name\": kluster.Spec.Name}).Inc()\nreturn fmt.Errorf(\"Seeding reconciliation failed: %w\", err)\n}\n"
},
{
"change_type": "MODIFY",
"old_path": "pkg/controller/ground/reconciler.go",
"new_path": "pkg/controller/ground/reconciler.go",
"diff": "@@ -87,7 +87,7 @@ func NewSeedReconciler(clients *config.Clients, kluster *v1.Kluster, logger log.\nreturn SeedReconciler{Clients: clients, Kluster: kluster, Logger: logger}\n}\n-func (sr *SeedReconciler) ReconcileSeeding(values map[string]interface{}) error {\n+func (sr *SeedReconciler) ReconcileSeeding(chartPath string, values map[string]interface{}) error {\nconfig, err := sr.Clients.Satellites.ConfigFor(sr.Kluster)\nif err != nil {\nreturn err\n@@ -105,7 +105,7 @@ func (sr *SeedReconciler) ReconcileSeeding(values map[string]interface{}) error\nif err != nil {\nreturn err\n}\n- planned, err := getPlannedObjects(&config, version, apiVersions, values)\n+ planned, err := getPlannedObjects(&config, version, apiVersions, chartPath, values)\nif err != nil {\nreturn err\n}\n@@ -148,9 +148,9 @@ func (sr *SeedReconciler) ReconcileSeeding(values map[string]interface{}) error\n}\n// Gets all resources as rendered by the seed chart\n-func getPlannedObjects(config *rest.Config, kubeVersion *version.Info, apiVersions chartutil.VersionSet, values map[string]interface{}) ([]unstructured.Unstructured, error) {\n+func getPlannedObjects(config *rest.Config, kubeVersion *version.Info, apiVersions chartutil.VersionSet, chartPath string, values map[string]interface{}) ([]unstructured.Unstructured, error) {\nplanned := make([]unstructured.Unstructured, 0)\n- seedChart, err := loader.Load(SeedChartPath)\n+ seedChart, err := loader.Load(chartPath)\nif err != nil {\nreturn planned, err\n}\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
Use --chart-directory flag to find seed chart (#740)
|
596,227 |
09.01.2023 13:21:48
| -3,600 |
0becc9c4b78b3956cc77c7bdebac1ccddd8f7f5b
|
Enable seed reconciliation for new clusters
|
[
{
"change_type": "MODIFY",
"old_path": "pkg/api/handlers/create_cluster.go",
"new_path": "pkg/api/handlers/create_cluster.go",
"diff": "@@ -83,6 +83,8 @@ func (d *createCluster) Handle(params operations.CreateClusterParams, principal\nreturn NewErrorResponse(&operations.CreateClusterDefault{}, 400, err.Error())\n}\n+ kluster.Labels[\"kubernikus.cloud.sap/seed-reconcile\"] = \"true\"\n+\nif kluster.ClusterCIDR() == \"\" && !kluster.Spec.NoCloud {\nreturn NewErrorResponse(&operations.CreateClusterDefault{}, 400, \"Specifying an empty ClusterCIDR is only allowed with noCloud: true\")\n}\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
Enable seed reconciliation for new clusters (#741)
|
596,227 |
09.01.2023 14:23:54
| -3,600 |
d502004ce24d976a675b982ab8dd3473d702cd00
|
Fix assignment to nil label map
|
[
{
"change_type": "MODIFY",
"old_path": "pkg/api/handlers/create_cluster.go",
"new_path": "pkg/api/handlers/create_cluster.go",
"diff": "@@ -83,6 +83,9 @@ func (d *createCluster) Handle(params operations.CreateClusterParams, principal\nreturn NewErrorResponse(&operations.CreateClusterDefault{}, 400, err.Error())\n}\n+ if kluster.Labels == nil {\n+ kluster.Labels = make(map[string]string)\n+ }\nkluster.Labels[\"kubernikus.cloud.sap/seed-reconcile\"] = \"true\"\nif kluster.ClusterCIDR() == \"\" && !kluster.Spec.NoCloud {\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
Fix assignment to nil label map (#742)
|
596,227 |
13.01.2023 10:38:56
| -3,600 |
ea86f5d1fac32f37d243ba306672bd1c0b7a3c16
|
Fix issues regarding seed reconciliation
Adds a missing clusterrolebindung
Run twice on cluster creation due to discovery client caching
Fix considered GVKs
|
[
{
"change_type": "MODIFY",
"old_path": "charts/seed/templates/csi.yaml",
"new_path": "charts/seed/templates/csi.yaml",
"diff": "@@ -292,6 +292,19 @@ subjects:\nname: csi-cinder-controller-sa\nnamespace: kube-system\n---\n+kind: ClusterRoleBinding\n+apiVersion: rbac.authorization.k8s.io/v1\n+metadata:\n+ name: snapshot-controller-role\n+subjects:\n+ - kind: ServiceAccount\n+ name: csi-cinder-controller-sa\n+ namespace: kube-system\n+roleRef:\n+ kind: ClusterRole\n+ name: snapshot-controller-runner\n+ apiGroup: rbac.authorization.k8s.io\n+---\napiVersion: storage.k8s.io/v1\nkind: CSIDriver\nmetadata:\n"
},
{
"change_type": "MODIFY",
"old_path": "pkg/controller/ground.go",
"new_path": "pkg/controller/ground.go",
"diff": "@@ -27,6 +27,7 @@ import (\n\"github.com/sapcc/kubernikus/pkg/api/models\"\nv1 \"github.com/sapcc/kubernikus/pkg/apis/kubernikus/v1\"\n+ \"github.com/sapcc/kubernikus/pkg/client/openstack/project\"\n\"github.com/sapcc/kubernikus/pkg/controller/config\"\n\"github.com/sapcc/kubernikus/pkg/controller/ground\"\n\"github.com/sapcc/kubernikus/pkg/controller/ground/bootstrap/ccm\"\n@@ -299,16 +300,17 @@ func (op *GroundControl) handler(key string) error {\nif err != nil {\nreturn err\n}\n- seedReconciler := ground.NewSeedReconciler(&op.Clients, kluster, op.Logger)\n- err = seedReconciler.EnrichHelmValuesForSeed(projectClient, helmValues)\n+ // we need to reconcile twice.\n+ // on the first run CRD will be added but these are not in the discovery client cache at that point in time, so creating CRD instances will silently fail.\n+ // we fail silently so reconciliation continues even when CRD's are not established.\n+ // we await CRD creation in the first run.\n+ // on the second run a new discovery client is created, so we know about all CRD's and respective instance can be created.\n+ for i := 0; i < 2; i++ {\n+ err = op.reconcileSeed(kluster, projectClient, helmValues)\nif err != nil {\nreturn err\n}\n- if err := seedReconciler.ReconcileSeeding(path.Join(op.Config.Helm.ChartDirectory, \"seed\"), helmValues); err != nil {\n- metrics.SeedReconciliationFailuresTotal.With(prometheus.Labels{\"kluster_name\": kluster.Spec.Name}).Inc()\n- return fmt.Errorf(\"Seeding reconciliation failed: %w\", err)\n}\n- op.Logger.Log(\"msg\", \"reconciled seeding successfully\", \"kluster\", kluster.GetName(), \"v\", 2)\nif err := op.updatePhase(kluster, models.KlusterPhaseRunning); err != nil {\nop.Logger.Log(\n\"msg\", \"failed to update status of kluster\",\n@@ -339,6 +341,8 @@ func (op *GroundControl) handler(key string) error {\nif err != nil {\nreturn err\n}\n+\n+ if kluster.Labels[SeedReconcileLabelKey] == \"true\" {\nhelmValues, err := helm_util.KlusterToHelmValues(kluster, klusterSecret, kluster.Spec.Version, &op.Config.Images, accessMode)\nif err != nil {\nreturn err\n@@ -347,17 +351,12 @@ func (op *GroundControl) handler(key string) error {\nif err != nil {\nreturn err\n}\n-\n- if kluster.Labels[SeedReconcileLabelKey] == \"true\" {\n- seedReconciler := ground.NewSeedReconciler(&op.Clients, kluster, op.Logger)\n- err = seedReconciler.EnrichHelmValuesForSeed(projectClient, helmValues)\n- if err != nil {\n- metrics.SeedReconciliationFailuresTotal.With(prometheus.Labels{\"kluster_name\": kluster.Spec.Name}).Inc()\n- return err\n- }\n- if err := seedReconciler.ReconcileSeeding(path.Join(op.Config.Helm.ChartDirectory, \"seed\"), helmValues); err != nil {\n- metrics.SeedReconciliationFailuresTotal.With(prometheus.Labels{\"kluster_name\": kluster.Spec.Name}).Inc()\n- return fmt.Errorf(\"Seeding reconciliation failed: %w\", err)\n+ if err := op.reconcileSeed(kluster, projectClient, helmValues); err != nil {\n+ op.Logger.Log(\n+ \"msg\", \"Failed seed reconciliation\",\n+ \"kluster\", kluster.GetName(),\n+ \"project\", kluster.Account(),\n+ \"err\", err)\n}\n}\n@@ -482,6 +481,20 @@ func (op *GroundControl) handler(key string) error {\nreturn nil\n}\n+func (op *GroundControl) reconcileSeed(kluster *v1.Kluster, projectClient project.ProjectClient, helmValues map[string]interface{}) error {\n+ seedReconciler := ground.NewSeedReconciler(&op.Clients, kluster, op.Logger)\n+ if err := seedReconciler.EnrichHelmValuesForSeed(projectClient, helmValues); err != nil {\n+ metrics.SeedReconciliationFailuresTotal.With(prometheus.Labels{\"kluster_name\": kluster.Spec.Name}).Inc()\n+ return fmt.Errorf(\"Enrichting seed values failed: %w\", err)\n+ }\n+ if err := seedReconciler.ReconcileSeeding(path.Join(op.Config.Helm.ChartDirectory, \"seed\"), helmValues); err != nil {\n+ metrics.SeedReconciliationFailuresTotal.With(prometheus.Labels{\"kluster_name\": kluster.Spec.Name}).Inc()\n+ return fmt.Errorf(\"Seeding reconciliation failed: %w\", err)\n+ }\n+ op.Logger.Log(\"msg\", \"reconciled seeding successfully\", \"kluster\", kluster.GetName(), \"v\", 2)\n+ return nil\n+}\n+\nfunc (op *GroundControl) klusterAdd(obj interface{}) {\nc := obj.(*v1.Kluster)\nkey, err := cache.MetaNamespaceKeyFunc(c)\n"
},
{
"change_type": "MODIFY",
"old_path": "pkg/controller/ground/gvk.go",
"new_path": "pkg/controller/ground/gvk.go",
"diff": "@@ -32,7 +32,7 @@ var managedGVKs = []schema.GroupVersionKind{\nKind: \"StorageClass\",\n},\n{\n- Group: \"storage.k8s.io/v1\",\n+ Group: \"snapshot.storage.k8s.io/v1\",\nVersion: \"v1\",\nKind: \"VolumeSnapshotClass\",\n},\n"
},
{
"change_type": "MODIFY",
"old_path": "pkg/controller/ground/reconciler.go",
"new_path": "pkg/controller/ground/reconciler.go",
"diff": "@@ -4,6 +4,7 @@ import (\n\"bytes\"\n\"context\"\n\"fmt\"\n+ \"time\"\njsonpatch \"github.com/evanphx/json-patch\"\n\"github.com/go-kit/kit/log\"\n@@ -12,11 +13,13 @@ import (\n\"helm.sh/helm/v3/pkg/chartutil\"\n\"helm.sh/helm/v3/pkg/engine\"\n\"helm.sh/helm/v3/pkg/releaseutil\"\n+ extensionsv1 \"k8s.io/apiextensions-apiserver/pkg/apis/apiextensions/v1\"\n\"k8s.io/apimachinery/pkg/api/errors\"\n\"k8s.io/apimachinery/pkg/api/meta\"\nmetav1 \"k8s.io/apimachinery/pkg/apis/meta/v1\"\n\"k8s.io/apimachinery/pkg/apis/meta/v1/unstructured\"\n\"k8s.io/apimachinery/pkg/types\"\n+ \"k8s.io/apimachinery/pkg/util/wait\"\n\"k8s.io/apimachinery/pkg/util/yaml\"\n\"k8s.io/apimachinery/pkg/version\"\n\"k8s.io/client-go/discovery\"\n@@ -337,8 +340,36 @@ func (sr *SeedReconciler) createPlanned(client dynamic.Interface, mapping *meta.\n\"kind\", fmt.Sprintf(\"%s\", obj.GetKind()),\n\"v\", 6)\n_, err := makeScopedClient(client, mapping, obj.GetNamespace()).Create(context.TODO(), obj, metav1.CreateOptions{})\n+ if err != nil {\nreturn err\n}\n+ if obj.GroupVersionKind().Kind == \"CustomResourceDefinition\" {\n+ sr.Logger.Log(\n+ \"msg\", \"Seed reconciliation: awaiting crd established\",\n+ \"name\", obj.GetName(),\n+ \"namespace\", obj.GetNamespace(),\n+ \"kind\", fmt.Sprintf(\"%s\", obj.GetKind()),\n+ \"v\", 6)\n+ return wait.Poll(500*time.Millisecond, 20*time.Second, func() (done bool, err error) {\n+ crd, err := makeScopedClient(client, mapping, obj.GetNamespace()).Get(context.TODO(), obj.GetName(), metav1.GetOptions{})\n+ if err != nil {\n+ return false, err\n+ }\n+ status := crd.Object[\"status\"].(map[string]interface{})\n+ conditions := status[\"conditions\"].([]interface{})\n+ for _, condition := range conditions {\n+ c := condition.(map[string]interface{})\n+ ty := c[\"type\"].(string)\n+ statusStr := c[\"status\"].(string)\n+ if ty == string(extensionsv1.Established) && statusStr == string(metav1.ConditionTrue) {\n+ return true, nil\n+ }\n+ }\n+ return false, nil\n+ })\n+ }\n+ return nil\n+}\nfunc (sr *SeedReconciler) patchDeployed(client dynamic.Interface, mapping *meta.RESTMapping, planned, deployed *unstructured.Unstructured) error {\n// skip if flagged\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
Fix issues regarding seed reconciliation (#744)
- Adds a missing clusterrolebindung
- Run twice on cluster creation due to discovery client caching
- Fix considered GVKs
|
596,227 |
19.01.2023 15:32:24
| -3,600 |
7c525e0ac97dcf2c522685b4cb52fcfba1218d6d
|
Allow recreation of resources during seed reconcile
|
[
{
"change_type": "MODIFY",
"old_path": "pkg/controller/ground/reconciler.go",
"new_path": "pkg/controller/ground/reconciler.go",
"diff": "@@ -39,6 +39,11 @@ const ManagedByLabelValue string = \"kubernikus\"\nconst SkipPatchKey string = \"kubernikus.cloud.sap/skip-manage\"\nconst SkipPatchValue string = \"true\"\n+var recreateKinds map[string]struct{} = map[string]struct{}{\n+ \"RoleBinding\": {},\n+ \"ClusterRoleBinding\": {},\n+}\n+\ntype objectDiff struct {\nplanned unstructured.Unstructured\ndeployed *unstructured.Unstructured\n@@ -223,9 +228,7 @@ func getManagedObjects(clients *config.Clients, mapper meta.RESTMapper, kluster\nif err != nil {\nreturn nil, err\n}\n- for _, oneManaged := range managedList.Items {\n- managed = append(managed, oneManaged)\n- }\n+ managed = append(managed, managedList.Items...)\n}\nreturn managed, nil\n}\n@@ -291,7 +294,7 @@ func (sr *SeedReconciler) deleteOrphanedObjects(client dynamic.Interface, mapper\n\"msg\", \"Seed reconciliation: deleting orphaned\",\n\"name\", oneOrphaned.GetName(),\n\"namespace\", oneOrphaned.GetNamespace(),\n- \"kind\", fmt.Sprintf(\"%s\", oneOrphaned.GetKind()),\n+ \"kind\", oneOrphaned.GetKind(),\n\"v\", 6)\nmapping, err := mapper.RESTMapping(oneOrphaned.GroupVersionKind().GroupKind(), oneOrphaned.GroupVersionKind().Version)\nif err != nil {\n@@ -337,7 +340,7 @@ func (sr *SeedReconciler) createPlanned(client dynamic.Interface, mapping *meta.\n\"msg\", \"Seed reconciliation: creating planned\",\n\"name\", obj.GetName(),\n\"namespace\", obj.GetNamespace(),\n- \"kind\", fmt.Sprintf(\"%s\", obj.GetKind()),\n+ \"kind\", obj.GetKind(),\n\"v\", 6)\n_, err := makeScopedClient(client, mapping, obj.GetNamespace()).Create(context.TODO(), obj, metav1.CreateOptions{})\nif err != nil {\n@@ -376,6 +379,8 @@ func (sr *SeedReconciler) patchDeployed(client dynamic.Interface, mapping *meta.\nif val, ok := deployed.GetLabels()[SkipPatchKey]; ok && val == SkipPatchValue {\nreturn nil\n}\n+ // make an unmodified deep copy of the planned object, which we could need re-creation\n+ plannedCopy := planned.DeepCopy()\n// copy over certain values to keep patches small\ndeployedMetadata := deployed.Object[\"metadata\"].(map[string]interface{})\nplannedMetadata := planned.Object[\"metadata\"].(map[string]interface{})\n@@ -392,7 +397,7 @@ func (sr *SeedReconciler) patchDeployed(client dynamic.Interface, mapping *meta.\n// Depending on the concrete resource there still patches that are not strictly\n// required fallthrough here. A prime example is the Container Spec of Deployments,\n// DaemonSets and so on, which has a bunch of optional fields, which aren't part\n- // of the planned maifest but of the deployed resources. That in turn creates some\n+ // of the planned manifest but of the deployed resources. That in turn creates some\n// larger patches.\noriginal, err := deployed.MarshalJSON()\n@@ -411,13 +416,54 @@ func (sr *SeedReconciler) patchDeployed(client dynamic.Interface, mapping *meta.\nreturn nil\n}\n+ // try to apply patch first\n+ err = sr.patchResource(client, mapping, deployed, patch)\n+ if err == nil {\n+ return nil\n+ } else if errors.IsInvalid(err) {\n+ // if the patch is invalid we can try to recreate certain resources\n+ if recreateAllowed(deployed.GetKind()) {\n+ return sr.recreateResource(client, mapping, plannedCopy)\n+ }\n+ }\n+ return err\n+}\n+\n+func recreateAllowed(kind string) bool {\n+ _, ok := recreateKinds[kind]\n+ return ok\n+}\n+\n+func (sr *SeedReconciler) patchResource(client dynamic.Interface, mapping *meta.RESTMapping, deployed *unstructured.Unstructured, patch []byte) error {\nsr.Logger.Log(\n\"msg\", \"Seed reconciliation: patching deployed\",\n\"name\", deployed.GetName(),\n\"namespace\", deployed.GetNamespace(),\n- \"kind\", fmt.Sprintf(\"%s\", deployed.GetKind()),\n+ \"kind\", deployed.GetKind(),\n\"patch\", string(patch),\n\"v\", 6)\n- _, err = makeScopedClient(client, mapping, deployed.GetNamespace()).Patch(context.TODO(), deployed.GetName(), types.MergePatchType, patch, metav1.PatchOptions{})\n+ _, err := makeScopedClient(client, mapping, deployed.GetNamespace()).Patch(context.TODO(), deployed.GetName(), types.MergePatchType, patch, metav1.PatchOptions{})\n+ return err\n+}\n+\n+func (sr *SeedReconciler) recreateResource(client dynamic.Interface, mapping *meta.RESTMapping, planned *unstructured.Unstructured) error {\n+ sr.Logger.Log(\n+ \"msg\", \"Seed reconciliation: recreating deployed\",\n+ \"name\", planned.GetName(),\n+ \"namespace\", planned.GetNamespace(),\n+ \"kind\", planned.GetKind(),\n+ \"v\", 6)\n+ // refuse to delete any resource called kubernikus:admin.\n+ // this could delete the clusterrolebinding we need to get\n+ // into the cluster and lock ourselves out\n+ if planned.GetName() == \"kubernikus:admin\" {\n+ return fmt.Errorf(\"refusing to recreate a resource with name kubernikus:admin\")\n+ }\n+ scoped := makeScopedClient(client, mapping, planned.GetNamespace())\n+ err := scoped.Delete(context.TODO(), planned.GetName(), metav1.DeleteOptions{})\n+ if err != nil {\n+ return nil\n+ }\n+ _, err = scoped.Create(context.TODO(), planned, metav1.CreateOptions{})\nreturn err\n}\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
Allow recreation of resources during seed reconcile (#749)
|
596,227 |
19.01.2023 15:35:50
| -3,600 |
1c09c5c697e3067e7f91e6563056b611a07ecd39
|
Fix replicate images CI step
* Deduplicate images before validating
* Serialize keppel validate
[ci skip]
|
[
{
"change_type": "MODIFY",
"old_path": "ci/pipeline.yaml.erb",
"new_path": "ci/pipeline.yaml.erb",
"diff": "@@ -282,9 +282,12 @@ jobs:\n- name: replicate-images\nplan:\n+ - in_parallel:\n- get: kubernikus.git\n+ - get: docs-builder.image\n+ passed: [docs]\ntrigger: true\n- passed: [build]\n+ passed: [docs]\n- task: replicate\ntimeout: 15m\nconfig:\n@@ -301,7 +304,7 @@ jobs:\nargs:\n- -cx\n- |\n- yq e '.imagesForVersion.*.*|.repository + \":\" + .tag' kubernikus.git/charts/images.yaml | grep -v changeme | sed -E 's/\\$REGION|global/eu-de-1/g' | xargs -n1 -P8 keppel validate\n+ yq e '.imagesForVersion.*.*|.repository + \":\" + .tag' kubernikus.git/charts/images.yaml | grep -v -E 'changeme|/ccloud/' | sed -E 's/\\$REGION|global/eu-de-1/g' | sort -u | xargs -n1 keppel validate\n- name: master\nserial: true\n@@ -310,7 +313,7 @@ jobs:\n- get: secrets.git\n- get: kubernikus.builds\nresource: kubernikus.git\n- passed: [replicate-images]\n+ passed: [replicate-images, build]\ntrigger: true\n- in_parallel:\n- task: kubernikus\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
Fix replicate images CI step (#747)
* Deduplicate images before validating
* Serialize keppel validate
[ci skip]
|
596,226 |
03.02.2023 11:40:07
| -3,600 |
f98f2dde496cea121165ee4057be9b3b0b48b7bf
|
Fixes in seed reconciliation, logging
|
[
{
"change_type": "MODIFY",
"old_path": "charts/seed/templates/coredns.yaml",
"new_path": "charts/seed/templates/coredns.yaml",
"diff": "@@ -101,11 +101,7 @@ spec:\nprotocol: TCP\nport: 9153\n---\n-{{- if semverCompare \">= 1.16\" .Capabilities.KubeVersion.Version -}}\napiVersion: apps/v1\n-{{- else if semverCompare \">= 1.13\" .Capabilities.KubeVersion.Version -}}\n-apiVersion: extensions/v1beta1\n-{{- end }}\nkind: Deployment\nmetadata:\nname: coredns\n"
},
{
"change_type": "MODIFY",
"old_path": "charts/seed/templates/csi.yaml",
"new_path": "charts/seed/templates/csi.yaml",
"diff": "@@ -378,7 +378,6 @@ metadata:\nname: csi-cinder-snapclass\ndriver: cinder.csi.openstack.org\ndeletionPolicy: Delete\n-{{- end }}\n---\nkind: DaemonSet\napiVersion: apps/v1\n@@ -497,3 +496,4 @@ spec:\n- name: secret-cinderplugin\nsecret:\nsecretName: cloud-config\n+{{- end }}\n"
},
{
"change_type": "MODIFY",
"old_path": "charts/seed/templates/kubedns.yaml",
"new_path": "charts/seed/templates/kubedns.yaml",
"diff": "@@ -18,7 +18,7 @@ metadata:\nlabels:\naddonmanager.kubernetes.io/mode: EnsureExists\n---\n-apiVersion: extensions/v1beta1\n+apiVersion: apps/v1\nkind: Deployment\nmetadata:\nname: kube-dns\n"
},
{
"change_type": "MODIFY",
"old_path": "pkg/controller/ground/reconciler.go",
"new_path": "pkg/controller/ground/reconciler.go",
"diff": "@@ -74,7 +74,7 @@ func (sr *SeedReconciler) EnrichHelmValuesForSeed(client project.ProjectClient,\nreturn err\n}\n// required to adapat old kube-dns deployments\n- _, err = k8sClient.ExtensionsV1beta1().Deployments(\"kube-system\").Get(context.TODO(), \"kube-dns\", metav1.GetOptions{})\n+ _, err = k8sClient.AppsV1().Deployments(\"kube-system\").Get(context.TODO(), \"kube-dns\", metav1.GetOptions{})\nvar isKubeDns bool\nif err == nil {\nisKubeDns = true\n@@ -120,6 +120,8 @@ func (sr *SeedReconciler) ReconcileSeeding(chartPath string, values map[string]i\nsr.Logger.Log(\n\"msg\", \"Seed reconciliation: planned objects\",\n\"count\", len(planned),\n+ \"kluster\", sr.Kluster.GetName(),\n+ \"project\", sr.Kluster.Account(),\n\"v\", 6)\ngroupRessources, err := restmapper.GetAPIGroupResources(discover)\n@@ -150,7 +152,8 @@ func (sr *SeedReconciler) ReconcileSeeding(chartPath string, values map[string]i\n}\nsr.Logger.Log(\n\"msg\", \"Seed reconciliation: successful\",\n- \"kluster\", sr.Kluster.Name,\n+ \"kluster\", sr.Kluster.GetName(),\n+ \"project\", sr.Kluster.Account(),\n\"v\", 5)\nreturn nil\n}\n@@ -295,6 +298,8 @@ func (sr *SeedReconciler) deleteOrphanedObjects(client dynamic.Interface, mapper\n\"name\", oneOrphaned.GetName(),\n\"namespace\", oneOrphaned.GetNamespace(),\n\"kind\", oneOrphaned.GetKind(),\n+ \"kluster\", sr.Kluster.GetName(),\n+ \"project\", sr.Kluster.Account(),\n\"v\", 6)\nmapping, err := mapper.RESTMapping(oneOrphaned.GroupVersionKind().GroupKind(), oneOrphaned.GroupVersionKind().Version)\nif err != nil {\n@@ -341,6 +346,8 @@ func (sr *SeedReconciler) createPlanned(client dynamic.Interface, mapping *meta.\n\"name\", obj.GetName(),\n\"namespace\", obj.GetNamespace(),\n\"kind\", obj.GetKind(),\n+ \"kluster\", sr.Kluster.GetName(),\n+ \"project\", sr.Kluster.Account(),\n\"v\", 6)\n_, err := makeScopedClient(client, mapping, obj.GetNamespace()).Create(context.TODO(), obj, metav1.CreateOptions{})\nif err != nil {\n@@ -352,6 +359,8 @@ func (sr *SeedReconciler) createPlanned(client dynamic.Interface, mapping *meta.\n\"name\", obj.GetName(),\n\"namespace\", obj.GetNamespace(),\n\"kind\", fmt.Sprintf(\"%s\", obj.GetKind()),\n+ \"kluster\", sr.Kluster.GetName(),\n+ \"project\", sr.Kluster.Account(),\n\"v\", 6)\nreturn wait.Poll(500*time.Millisecond, 20*time.Second, func() (done bool, err error) {\ncrd, err := makeScopedClient(client, mapping, obj.GetNamespace()).Get(context.TODO(), obj.GetName(), metav1.GetOptions{})\n@@ -441,6 +450,8 @@ func (sr *SeedReconciler) patchResource(client dynamic.Interface, mapping *meta.\n\"namespace\", deployed.GetNamespace(),\n\"kind\", deployed.GetKind(),\n\"patch\", string(patch),\n+ \"kluster\", sr.Kluster.GetName(),\n+ \"project\", sr.Kluster.Account(),\n\"v\", 6)\n_, err := makeScopedClient(client, mapping, deployed.GetNamespace()).Patch(context.TODO(), deployed.GetName(), types.MergePatchType, patch, metav1.PatchOptions{})\nreturn err\n@@ -452,6 +463,8 @@ func (sr *SeedReconciler) recreateResource(client dynamic.Interface, mapping *me\n\"name\", planned.GetName(),\n\"namespace\", planned.GetNamespace(),\n\"kind\", planned.GetKind(),\n+ \"kluster\", sr.Kluster.GetName(),\n+ \"project\", sr.Kluster.Account(),\n\"v\", 6)\n// refuse to delete any resource called kubernikus:admin.\n// this could delete the clusterrolebinding we need to get\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
Fixes in seed reconciliation, logging (#754)
|
596,226 |
10.02.2023 15:12:55
| -3,600 |
fd1d63c81a664f5763c2126c8f5f14a5fc8b5706
|
{skip ci] Add flatcar update reset scripts
|
[
{
"change_type": "ADD",
"old_path": null,
"new_path": "scripts/fix-flatcar-update-reset.sh",
"diff": "+#!/bin/bash\n+set -eo pipefail\n+\n+if [ \"$2\" == \"\" ]; then\n+ echo usage $0 PARENT_CONTEXT CLUSTER_FQDN\n+ exit\n+fi\n+unset KUBECONTEXT\n+unset KUBENAMESPACE\n+\n+PARENT_CONTEXT=$1\n+CLUSTER=$2\n+CLUSTER_NAME=${2%-*}\n+CLUSTER_CONFIG=$TMPDIR$CLUSTER\n+\n+KUBECONTEXT=$PARENT_CONTEXT ./kubeconfig.sh $2 >$CLUSTER_CONFIG\n+echo Check status with:\n+echo env KUBECONFIG=$CLUSTER_CONFIG kubectl --context=$CLUSTER -n default get pods\n+./flatcar-update-reset.sh $CLUSTER $CLUSTER_CONFIG\n+\n+rm $CLUSTER_CONFIG\n"
},
{
"change_type": "ADD",
"old_path": null,
"new_path": "scripts/flatcar-update-reset.sh",
"diff": "+#!/bin/bash\n+set -eo pipefail\n+if [ \"$1\" == \"\" ]; then\n+ echo \"usage: $0 CONTEXT [KUBECONFIG]\"\n+ exit 1\n+fi\n+\n+CONTEXT=$1\n+\n+if [ \"$2\" != \"\" ]; then\n+ echo using kubeconfig $2\n+ export KUBECONFIG=$2\n+fi\n+\n+kubectl apply --context $CONTEXT -n default -f - <<EOF\n+apiVersion: apps/v1\n+kind: DaemonSet\n+metadata:\n+ name: flatcar-update-reset\n+spec:\n+ selector:\n+ matchLabels:\n+ app: flatcar-update-reset\n+ template:\n+ metadata:\n+ labels:\n+ app: flatcar-update-reset\n+ spec:\n+ tolerations:\n+ - operator: Exists\n+ hostPID: true\n+ initContainers:\n+ - name: init\n+ image: keppel.global.cloud.sap/ccloud-dockerhub-mirror/library/alpine:latest\n+ securityContext:\n+ privileged: true\n+ command:\n+ - sh\n+ - -c\n+ args:\n+ - |-\n+ set -xe\n+ chroot /host update_engine_client -reset_status\n+ chroot /host update_engine_client -check_for_update\n+ volumeMounts:\n+ - name: host\n+ mountPath: \"/host\"\n+ containers:\n+ - name: pause\n+ image: keppel.global.cloud.sap/ccloud-dockerhub-mirror/sapcc/pause-amd64:3.1\n+ volumes:\n+ - name: host\n+ hostPath:\n+ path: \"/\"\n+EOF\n+\n+timeout -v 300 kubectl rollout status --context $CONTEXT -n default daemonset/flatcar-update-reset\n+kubectl delete --context $CONTEXT -n default daemonset/flatcar-update-reset\n"
}
] |
Go
|
Apache License 2.0
|
sapcc/kubernikus
|
{skip ci] Add flatcar update reset scripts
|
8,490 |
02.01.2017 14:03:28
| -3,600 |
f5d765c52385dd7a040ecccf5cea4440d1c432eb
|
Add a adaptive SC-List decoder (ASCL, works only with a CRC).
|
[
{
"change_type": "MODIFY",
"old_path": "src/Launcher/BFER/Polar/Launcher_BFER_polar.cpp",
"new_path": "src/Launcher/BFER/Polar/Launcher_BFER_polar.cpp",
"diff": "@@ -80,7 +80,7 @@ void Launcher_BFER_polar<B,R,Q>\n\"disable the systematic encoding.\"};\n// ------------------------------------------------------------------------------------------------------- decoder\n- this->opt_args[{\"dec-type\", \"D\"}].push_back(\"SC, SCL, SCAN\");\n+ this->opt_args[{\"dec-type\", \"D\"}].push_back(\"SC, SCL, ASCL, SCAN\");\nthis->opt_args[{\"dec-ite\", \"i\"}] =\n{\"positive_int\",\n\"maximal number of iterations in the SCAN decoder.\"};\n@@ -197,6 +197,9 @@ std::vector<std::pair<std::string,std::string>> Launcher_BFER_polar<B,R,Q>\nif (this->params.decoder.type == \"SCL\")\np.push_back(std::make_pair(\"Num. of lists (L)\", std::to_string(this->params.decoder.L)));\n+ if (this->params.decoder.type == \"ASCL\")\n+ p.push_back(std::make_pair(\"Max num. of lists (L)\", std::to_string(this->params.decoder.L)));\n+\nreturn p;\n}\n"
},
{
"change_type": "ADD",
"old_path": null,
"new_path": "src/Module/Decoder/Polar/ASCL/Decoder_polar_ASCL_fast_CA_sys.hpp",
"diff": "+#ifndef DECODER_POLAR_ASCL_FAST_SYS_CA\n+#define DECODER_POLAR_ASCL_FAST_SYS_CA\n+\n+#include \"../SC/Decoder_polar_SC_fast_sys.hpp\"\n+#include \"../SCL/CRC/Decoder_polar_SCL_fast_CA_sys.hpp\"\n+\n+#include \"Module/CRC/CRC.hpp\"\n+\n+template <typename B, typename R, class API_polar>\n+class Decoder_polar_ASCL_fast_CA_sys : public Decoder_polar_SCL_fast_CA_sys<B,R,API_polar>\n+{\n+private:\n+ Decoder_polar_SC_fast_sys<B,R,API_polar> sc_decoder;\n+ const int L_max;\n+\n+public:\n+ Decoder_polar_ASCL_fast_CA_sys(const int& K, const int& N, const int& max_L, const mipp::vector<B>& frozen_bits,\n+ CRC<B>& crc, const int n_frames = 1,\n+ const std::string name = \"Decoder_polar_ASCL_fast_CA_sys\");\n+ virtual ~Decoder_polar_ASCL_fast_CA_sys(){};\n+\n+ void load (const mipp::vector<R>& Y_N);\n+ void hard_decode( );\n+ void store (mipp::vector<B>& V_N ) const;\n+ void unpack (mipp::vector<B>& V_N ) const;\n+ void store_fast (mipp::vector<B>& V ) const;\n+};\n+\n+#include \"Decoder_polar_ASCL_fast_CA_sys.hxx\"\n+\n+#endif /* DECODER_POLAR_ASCL_FAST_SYS_CA_CA */\n"
},
{
"change_type": "ADD",
"old_path": null,
"new_path": "src/Module/Decoder/Polar/ASCL/Decoder_polar_ASCL_fast_CA_sys.hxx",
"diff": "+#include \"Decoder_polar_ASCL_fast_CA_sys.hpp\"\n+#include \"Tools/Algo/Bit_packer.hpp\"\n+\n+template <typename B, typename R, class API_polar>\n+Decoder_polar_ASCL_fast_CA_sys<B,R,API_polar>\n+::Decoder_polar_ASCL_fast_CA_sys(const int& K, const int& N, const int& L_max, const mipp::vector<B>& frozen_bits,\n+ CRC<B>& crc, const int n_frames, const std::string name)\n+: Decoder_polar_SCL_fast_CA_sys<B,R,API_polar>(K, N, L_max, frozen_bits, crc, n_frames, name),\n+ sc_decoder (K, N , frozen_bits, n_frames, name),\n+ L_max(L_max)\n+{\n+ assert(L_max > 0);\n+}\n+\n+template <typename B, typename R, class API_polar>\n+void Decoder_polar_ASCL_fast_CA_sys<B,R,API_polar>\n+::load(const mipp::vector<R>& Y_N)\n+{\n+ sc_decoder.load(Y_N);\n+}\n+\n+template <typename B, typename R, class API_polar>\n+void Decoder_polar_ASCL_fast_CA_sys<B,R,API_polar>\n+::hard_decode()\n+{\n+ this->L = 1;\n+ sc_decoder.hard_decode();\n+\n+ // check the CRC\n+ auto crc_decode_result = this->crc_check(sc_decoder.s);\n+\n+ // delete the path if the CRC result is negative\n+ if (!crc_decode_result && L_max > 1)\n+ {\n+ std::copy(sc_decoder.l.begin(), sc_decoder.l.begin() + this->N, this->Y_N.begin());\n+ do\n+ {\n+ int first_node_id = 0, off_l = 0, off_s = 0;\n+\n+ this->L <<= 1;\n+ this->init_buffers();\n+ this->recursive_decode(off_l, off_s, this->m, first_node_id);\n+ }\n+ while (!this->select_best_path() && this->L < L_max);\n+ }\n+}\n+\n+template <typename B, typename R, class API_polar>\n+void Decoder_polar_ASCL_fast_CA_sys<B,R,API_polar>\n+::store(mipp::vector<B>& V_K) const\n+{\n+ if (this->L == 1) sc_decoder. store(V_K);\n+ else Decoder_polar_SCL_fast_CA_sys<B,R,API_polar>::store(V_K);\n+}\n+\n+template <typename B, typename R, class API_polar>\n+void Decoder_polar_ASCL_fast_CA_sys<B,R,API_polar>\n+::store_fast(mipp::vector<B>& V) const\n+{\n+ if (this->L == 1) sc_decoder. store_fast(V);\n+ else Decoder_polar_SCL_fast_CA_sys<B,R,API_polar>::store_fast(V);\n+}\n+\n+template <typename B, typename R, class API_polar>\n+void Decoder_polar_ASCL_fast_CA_sys<B,R,API_polar>\n+::unpack(mipp::vector<B>& V_N) const\n+{\n+ if (this->L == 1) sc_decoder. unpack(V_N);\n+ else Decoder_polar_SCL_fast_CA_sys<B,R,API_polar>::unpack(V_N);\n+}\n"
},
{
"change_type": "MODIFY",
"old_path": "src/Module/Decoder/Polar/SC/Decoder_polar_SC_fast_sys.hpp",
"new_path": "src/Module/Decoder/Polar/SC/Decoder_polar_SC_fast_sys.hpp",
"diff": "#include \"../decoder_polar_functions.h\"\n#include \"../../Decoder.hpp\"\n+template <typename B, typename R, class API_polar>\n+class Decoder_polar_ASCL_fast_CA_sys;\n+\ntemplate <typename B, typename R, class API_polar>\nclass Decoder_polar_SC_fast_sys : public Decoder<B,R>\n{\n+ friend Decoder_polar_ASCL_fast_CA_sys<B,R,API_polar>;\n+\nprotected:\nconst int m; // graph depth\nmipp::vector<R> l; // lambda, LR or LLR\n@@ -28,11 +33,11 @@ public:\nvirtual ~Decoder_polar_SC_fast_sys();\nprotected:\n- void load (const mipp::vector<R>& Y_N);\n+ virtual void load (const mipp::vector<R>& Y_N);\nvirtual void hard_decode( );\n- void store ( mipp::vector<B>& V_K) const;\n- void store_fast ( mipp::vector<B>& V_N) const;\n- void unpack ( mipp::vector<B>& V_N) const;\n+ virtual void store ( mipp::vector<B>& V_K) const;\n+ virtual void store_fast ( mipp::vector<B>& V_N) const;\n+ virtual void unpack ( mipp::vector<B>& V_N) const;\nvirtual void recursive_decode(const int off_l, const int off_s, const int reverse_depth, int &node_id);\n};\n"
},
{
"change_type": "MODIFY",
"old_path": "src/Module/Decoder/Polar/SCL/CRC/Decoder_polar_SCL_fast_CA_sys.hpp",
"new_path": "src/Module/Decoder/Polar/SCL/CRC/Decoder_polar_SCL_fast_CA_sys.hpp",
"diff": "template <typename B, typename R, class API_polar>\nclass Decoder_polar_SCL_fast_CA_sys : public Decoder_polar_SCL_fast_sys<B,R,API_polar>\n{\n-private:\n+protected:\nCRC<B>& crc;\nmipp::vector<B> U_test;\n@@ -17,7 +17,8 @@ public:\nvirtual ~Decoder_polar_SCL_fast_CA_sys(){};\nprotected:\n- virtual void select_best_path();\n+ bool crc_check (mipp::vector<B> &s);\n+ virtual int select_best_path( );\n};\n#include \"Decoder_polar_SCL_fast_CA_sys.hxx\"\n"
},
{
"change_type": "MODIFY",
"old_path": "src/Module/Decoder/Polar/SCL/CRC/Decoder_polar_SCL_fast_CA_sys.hxx",
"new_path": "src/Module/Decoder/Polar/SCL/CRC/Decoder_polar_SCL_fast_CA_sys.hxx",
"diff": "@@ -11,23 +11,17 @@ Decoder_polar_SCL_fast_CA_sys<B,R,API_polar>\n}\ntemplate <typename B, typename R, class API_polar>\n-void Decoder_polar_SCL_fast_CA_sys<B,R,API_polar>\n-::select_best_path()\n-{\n- auto j = 0;\n- auto n_active_paths_cpy = this->n_active_paths;\n- for (auto i = 0; i < n_active_paths_cpy; i++)\n+bool Decoder_polar_SCL_fast_CA_sys<B,R,API_polar>\n+::crc_check(mipp::vector<B> &s)\n{\n- const auto path = this->paths[j];\n-\n// // extract the info bits from the codeword\n// auto k = 0;\n// for (auto leaf = 0; leaf < this->N; leaf++)\n// if (!this->frozen_bits[leaf])\n-// U_test[k++] = this->s[path][leaf];\n+// U_test[k++] = s[leaf];\n// // check the CRC\n-// bool decode_result = crc.check(U_test, this->get_simd_inter_frame_level());\n+// return crc.check(U_test, this->get_simd_inter_frame_level());\n// extract the info bits (packed) from the codeword\nauto bytes = (unsigned char*)U_test.data();\n@@ -38,21 +32,37 @@ void Decoder_polar_SCL_fast_CA_sys<B,R,API_polar>\nauto b = 0;\nwhile (b < 8 && (leaf < this->N))\n{\n- byte |= !this->frozen_bits[leaf] ? ((unsigned char)(this->s[path][leaf] != 0)) << b++ : 0;\n+ byte |= !this->frozen_bits[leaf] ? ((unsigned char)(s[leaf] != 0)) << b++ : 0;\nleaf++;\n}\nbytes[B_pos++] = byte;\n}\n// check the CRC\n- bool decode_result = crc.check_packed(U_test, this->get_simd_inter_frame_level());\n+ return crc.check_packed(U_test, this->get_simd_inter_frame_level());\n+}\n+\n+template <typename B, typename R, class API_polar>\n+int Decoder_polar_SCL_fast_CA_sys<B,R,API_polar>\n+::select_best_path()\n+{\n+ auto n_valid_paths = 0;\n+ auto n_active_paths_cpy = this->n_active_paths;\n+ for (auto i = 0; i < n_active_paths_cpy; i++)\n+ {\n+ const auto path = this->paths[n_valid_paths];\n+\n+ // check the CRC\n+ auto decode_result = crc_check(this->s[path]);\n// delete the path if the CRC result is negative\nif (!decode_result)\n- this->delete_path(j);\n+ this->delete_path(n_valid_paths);\nelse\n- j++;\n+ n_valid_paths++;\n}\nthis->Decoder_polar_SCL_fast_sys<B,R,API_polar>::select_best_path();\n+\n+ return n_valid_paths;\n}\n"
},
{
"change_type": "MODIFY",
"old_path": "src/Module/Decoder/Polar/SCL/Decoder_polar_SCL_fast_sys.hpp",
"new_path": "src/Module/Decoder/Polar/SCL/Decoder_polar_SCL_fast_sys.hpp",
"diff": "@@ -13,7 +13,7 @@ class Decoder_polar_SCL_fast_sys : public Decoder<B,R>\n{\nprotected:\nconst int m; // graph depth\n- const int L; // maximum paths number\n+ int L; // maximum paths number\nconst mipp::vector<B>& frozen_bits;\nconst Pattern_polar_parser<B> polar_patterns;\n@@ -42,15 +42,16 @@ public:\nconst int n_frames = 1, const std::string name = \"Decoder_polar_SCL_fast_sys\");\nvirtual ~Decoder_polar_SCL_fast_sys();\n- void load (const mipp::vector<R>& Y_N);\n- void hard_decode( );\n+ virtual void load (const mipp::vector<R>& Y_N);\n+ virtual void hard_decode( );\nvirtual void store (mipp::vector<B>& V_N ) const;\n- void unpack (mipp::vector<B>& V_N ) const;\n+ virtual void unpack (mipp::vector<B>& V_N ) const;\nvirtual void store_fast (mipp::vector<B>& V ) const;\n-private:\n+protected:\ninline void recursive_decode(const int off_l, const int off_s, const int rev_depth, int &node_id );\n+private:\ninline void update_paths_r0 (const int rev_depth, const int off_l, const int off_s, const int n_elmts);\ninline void update_paths_r1 (const int rev_depth, const int off_l, const int off_s, const int n_elmts);\ninline void update_paths_rep(const int rev_depth, const int off_l, const int off_s, const int n_elmts);\n@@ -64,8 +65,9 @@ private:\ninline int up_ref_array_idx(const int path, const int r_d ); // return the array\nprotected:\n+ inline void init_buffers ( );\ninline void delete_path (int path_id);\n- virtual inline void select_best_path( );\n+ virtual inline int select_best_path( );\n};\n#include \"Decoder_polar_SCL_fast_sys.hxx\"\n"
},
{
"change_type": "MODIFY",
"old_path": "src/Module/Decoder/Polar/SCL/Decoder_polar_SCL_fast_sys.hxx",
"new_path": "src/Module/Decoder/Polar/SCL/Decoder_polar_SCL_fast_sys.hxx",
"diff": "#include \"Tools/Perf/MIPP/mipp.h\"\n#include \"Tools/Perf/Transpose/transpose_selector.h\"\n#include \"Tools/Display/bash_tools.h\"\n+#include \"Tools/Math/utils.h\"\n#include \"Decoder_polar_SCL_fast_sys.hpp\"\n@@ -45,7 +46,7 @@ Decoder_polar_SCL_fast_sys<B,R,API_polar>\nmetrics (L),\nY_N ( N + mipp::nElReg<R>() ),\nl (L, mipp::vector<R>(N + mipp::nElReg<R>())),\n- s (L, mipp::vector<B>(N + mipp::nElReg<B>(), 0)),\n+ s (L, mipp::vector<B>(N + mipp::nElReg<B>())),\nmetrics_vec (3, std::vector<float>()),\nmetrics_idx (3, std::vector<int >()),\ndup_count (L, 0),\n@@ -54,12 +55,14 @@ Decoder_polar_SCL_fast_sys<B,R,API_polar>\nis_even (L),\nbest_path (0),\nn_active_paths(1),\n- n_array_ref (L, std::vector<int>(m, 0)),\n- path_2_array (L, std::vector<int>(m, 0))\n+ n_array_ref (L, std::vector<int>(m)),\n+ path_2_array (L, std::vector<int>(m))\n{\nstatic_assert(API_polar::get_n_frames() == 1, \"The inter-frame API_polar is not supported.\");\nstatic_assert(sizeof(B) == sizeof(R), \"Sizes of the bits and reals have to be identical.\");\n+ assert(is_power_of_2(L));\n+\nmetrics_vec[0].resize(L * 2);\nmetrics_idx[0].resize(L * 2);\n@@ -69,14 +72,10 @@ Decoder_polar_SCL_fast_sys<B,R,API_polar>\nmetrics_vec[2].resize((L <= 2 ? 2 : 8) * L);\nmetrics_idx[2].resize((L <= 2 ? 2 : 8) * L);\n- for (auto j = 0 ; j < 3 ; j++)\n- for (size_t i = 0 ; i < metrics_idx[j].size() ; i++)\n- metrics_idx[j][i] = i;\n-\nfor (auto i = 0; i <= m; i++)\n{\nllr_indexes.push_back(std::vector<int>(std::exp2(i)));\n- for (size_t j = 0 ; j < llr_indexes[i].size() ; j++)\n+ for (auto j = 0 ; j < (int)llr_indexes[i].size() ; j++)\nllr_indexes[i][j] = j;\n}\n}\n@@ -90,23 +89,30 @@ Decoder_polar_SCL_fast_sys<B,R,API_polar>\ntemplate <typename B, typename R, class API_polar>\nvoid Decoder_polar_SCL_fast_sys<B,R,API_polar>\n-::load(const mipp::vector<R>& Y_N)\n+::init_buffers()\n{\n- std::fill(metrics.begin(), metrics.end(), std::numeric_limits<float>::min());\n+ std::fill(metrics.begin(), metrics.begin() + L, std::numeric_limits<float>::min());\nfor (auto i = 0; i < L; i++) paths[i] = i;\nn_active_paths = 1;\n- for (auto i = 0; i < 3; i++)\n- std::fill(metrics_vec[i].begin(), metrics_vec[i].end(), std::numeric_limits<float>::max());\n-\n- std::copy(Y_N.begin(), Y_N.end(), this->Y_N.begin());\n-\n// at the beginning, path 0 points to array 0\nstd::fill(n_array_ref [0].begin(), n_array_ref [0].end(), 1);\nstd::fill(path_2_array[0].begin(), path_2_array[0].end(), 0);\nfor (auto i = 1; i < L; i++)\nstd::fill(n_array_ref[i].begin(), n_array_ref[i].end(), 0);\n+\n+ for (auto i = 0 ; i < 2 * L ; i++) metrics_idx[0][i] = i;\n+ for (auto i = 0 ; i < 4 * L ; i++) metrics_idx[1][i] = i;\n+ for (auto i = 0 ; i < (L <= 2 ? 2 : 8) * L ; i++) metrics_idx[2][i] = i;\n+}\n+\n+template <typename B, typename R, class API_polar>\n+void Decoder_polar_SCL_fast_sys<B,R,API_polar>\n+::load(const mipp::vector<R>& Y_N)\n+{\n+ std::copy(Y_N.begin(), Y_N.begin() + this->N, this->Y_N.begin());\n+ init_buffers();\n}\ntemplate <typename B, typename R, class API_polar>\n@@ -383,21 +389,15 @@ void Decoder_polar_SCL_fast_sys<B,R,API_polar>\nfor (auto j = 0 ; j < 4 ; j++)\nmetrics_vec[1][4 * paths[i] +j] = std::numeric_limits<float>::max();\n- std::partial_sort(metrics_idx[1].begin(), metrics_idx[1].begin() + L, metrics_idx[1].end(),\n+ std::partial_sort(metrics_idx[1].begin(), metrics_idx[1].begin() + L, metrics_idx[1].begin() + L * 4,\n[this](int x, int y) {\nreturn metrics_vec[1][x] < metrics_vec[1][y];\n});\n- // TODO: disable this code\n// L first of the lists are the L best paths\nfor (auto i = 0; i < L; i++)\ndup_count[metrics_idx[1][i] / 4]++;\n-// // TODO: enable this code\n-// // L first of the lists are the L best paths\n-// for (auto i = 0; i < n_active_paths; i++)\n-// dup_count[metrics_idx[1][paths[i]] / 4]++;\n-\n// erase paths\nauto k = 0;\nauto n_active_paths_cpy = n_active_paths;\n@@ -421,7 +421,7 @@ void Decoder_polar_SCL_fast_sys<B,R,API_polar>\ndup_count[path] = 0;\n}\n}\n- std::fill(dup_count.begin(), dup_count.end(), 0); // TODO: remove this fill\n+ std::fill(dup_count.begin(), dup_count.begin() + L, 0);\n}\n}\n@@ -457,7 +457,7 @@ void Decoder_polar_SCL_fast_sys<B,R,API_polar>\nelse // n_active_paths == L\n{\n// sort hypothetic metrics\n- std::sort(metrics_idx[0].begin(), metrics_idx[0].end(),\n+ std::sort(metrics_idx[0].begin(), metrics_idx[0].begin() + L * 2,\n[this](int x, int y) {\nreturn metrics_vec[0][x] < metrics_vec[0][y];\n});\n@@ -550,7 +550,7 @@ void Decoder_polar_SCL_fast_sys<B,R,API_polar>\n}\n}\n- std::partial_sort(metrics_idx[2].begin(), metrics_idx[2].begin() + L, metrics_idx[2].end(),\n+ std::partial_sort(metrics_idx[2].begin(), metrics_idx[2].begin() + L, metrics_idx[2].begin() + (L <= 2 ? 2 : 8) * L,\n[this](int x, int y){\nreturn metrics_vec[2][x] < metrics_vec[2][y];\n});\n@@ -581,7 +581,7 @@ void Decoder_polar_SCL_fast_sys<B,R,API_polar>\ndup_count[path] = 0;\n}\n}\n- std::fill(dup_count.begin(), dup_count.end(), 0); // TODO: remove this fill\n+ std::fill(dup_count.begin(), dup_count.begin() + L, 0);\n}\ntemplate <typename B, typename R, class API_polar>\n@@ -697,7 +697,7 @@ void Decoder_polar_SCL_fast_sys<B,R,API_polar>\n}\ntemplate <typename B, typename R, class API_polar>\n-void Decoder_polar_SCL_fast_sys<B,R,API_polar>\n+int Decoder_polar_SCL_fast_sys<B,R,API_polar>\n::select_best_path()\n{\nbest_path = -1;\n@@ -707,6 +707,8 @@ void Decoder_polar_SCL_fast_sys<B,R,API_polar>\nif (best_path == -1)\nbest_path = 0;\n+\n+ return n_active_paths;\n}\ntemplate <typename B, typename R, class API_polar>\n"
},
{
"change_type": "MODIFY",
"old_path": "src/Module/Decoder/Polar/SCL/Decoder_polar_SCL_naive.hxx",
"new_path": "src/Module/Decoder/Polar/SCL/Decoder_polar_SCL_naive.hxx",
"diff": "#include <map>\n#include \"Tools/Display/bash_tools.h\"\n+#include \"Tools/Math/utils.h\"\n#include \"Decoder_polar_SCL_naive.hpp\"\n@@ -19,6 +20,8 @@ Decoder_polar_SCL_naive<B,R,F,G>\nfrozen_bits(frozen_bits),\nL(L)\n{\n+ assert(is_power_of_2(L));\n+\nthis->active_paths.insert(0);\nfor (auto i = 0; i < L; i++)\n{\n"
},
{
"change_type": "MODIFY",
"old_path": "src/Tools/Factory/Polar/Factory_decoder_polar.cpp",
"new_path": "src/Tools/Factory/Polar/Factory_decoder_polar.cpp",
"diff": "#include \"Module/Decoder/Polar/SCL/CRC/Decoder_polar_SCL_naive_CA.hpp\"\n#include \"Module/Decoder/Polar/SCL/CRC/Decoder_polar_SCL_naive_CA_sys.hpp\"\n#include \"Module/Decoder/Polar/SCL/CRC/Decoder_polar_SCL_fast_CA_sys.hpp\"\n+#include \"Module/Decoder/Polar/ASCL/Decoder_polar_ASCL_fast_CA_sys.hpp\"\n// #define API_POLAR_DYNAMIC 1\n@@ -197,6 +198,25 @@ Decoder<B,R>* Factory_decoder_polar<B,R>\ndecoder = new Decoder_polar_SCL_fast_CA_sys<B, R, API_polar>(params.code.K, params.code.N_code, params.decoder.L, frozen_bits, *crc, params.simulation.inter_frame_level);\n}\n}\n+\n+ if (params.decoder.type == \"ASCL\")\n+ {\n+ if (params.decoder.simd_strategy == \"INTRA\")\n+ {\n+ using API_polar = API_polar_dynamic_intra\n+ <B, R, f_LLR <R>, g_LLR <B,R>, g0_LLR <R>, h_LLR <B,R>, xo_STD <B>,\n+ f_LLR_i<R>, g_LLR_i<B,R>, g0_LLR_i<R>, h_LLR_i<B,R>, xo_STD_i<B>>;\n+ if (!params.crc.poly.empty())\n+ decoder = new Decoder_polar_ASCL_fast_CA_sys<B, R, API_polar>(params.code.K, params.code.N_code, params.decoder.L, frozen_bits, *crc, params.simulation.inter_frame_level);\n+ }\n+ else if (params.decoder.simd_strategy.empty())\n+ {\n+ using API_polar = API_polar_dynamic_seq\n+ <B, R, f_LLR<R>, g_LLR<B,R>, g0_LLR<R>, h_LLR<B,R>, xo_STD<B>>;\n+ if (!params.crc.poly.empty())\n+ decoder = new Decoder_polar_ASCL_fast_CA_sys<B, R, API_polar>(params.code.K, params.code.N_code, params.decoder.L, frozen_bits, *crc, params.simulation.inter_frame_level);\n+ }\n+ }\n}\n}\n"
},
{
"change_type": "MODIFY",
"old_path": "src/Tools/Math/utils.h",
"new_path": "src/Tools/Math/utils.h",
"diff": "@@ -54,4 +54,10 @@ inline void saturate(mipp::vector<T> &array, const T min, const T max)\ntemplate <typename B, typename R>\nB sgn(R val) { return (B)((R(0) < val) - (val < R(0))); }\n+template <typename T>\n+constexpr bool is_power_of_2(T x)\n+{\n+ return (x > 0) && !(x & (x - 1));\n+}\n+\n#endif /* MATH_UTILS_H */\n"
},
{
"change_type": "MODIFY",
"old_path": "src/Tools/Perf/Reorderer/Reorderer.hxx",
"new_path": "src/Tools/Perf/Reorderer/Reorderer.hxx",
"diff": "#include <cmath>\n#include <cassert>\n+#include \"Tools/Math/utils.h\"\n+\n#include \"Reorderer.hpp\"\n#ifdef _MSC_VER\n#pragma warning( disable : 4717 )\n#endif\n-constexpr bool is_power_of_2(unsigned x)\n-{\n- return (x > 0) && !(x & (x - 1));\n-}\n-\ntemplate <typename T>\nvoid Reorderer<T>\n::apply(const std::vector<const T*> in_data, T* out_data, const int data_length)\n"
}
] |
C++
|
MIT License
|
aff3ct/aff3ct
|
Add a adaptive SC-List decoder (ASCL, works only with a CRC).
|
8,490 |
03.01.2017 10:14:51
| -3,600 |
e1f3674d5df81b5eb256d7247dd2b1fb47216bea
|
Add a pseudo adaptive methode in the ASCL decoder (like in the McGill paper).
|
[
{
"change_type": "MODIFY",
"old_path": "src/Module/Decoder/Polar/ASCL/Decoder_polar_ASCL_fast_CA_sys.hpp",
"new_path": "src/Module/Decoder/Polar/ASCL/Decoder_polar_ASCL_fast_CA_sys.hpp",
"diff": "@@ -12,6 +12,7 @@ class Decoder_polar_ASCL_fast_CA_sys : public Decoder_polar_SCL_fast_CA_sys<B,R,\nprivate:\nDecoder_polar_SC_fast_sys<B,R,API_polar> sc_decoder;\nconst int L_max;\n+ const bool is_full_adaptive;\npublic:\nDecoder_polar_ASCL_fast_CA_sys(const int& K, const int& N, const int& max_L, const mipp::vector<B>& frozen_bits,\n"
},
{
"change_type": "MODIFY",
"old_path": "src/Module/Decoder/Polar/ASCL/Decoder_polar_ASCL_fast_CA_sys.hxx",
"new_path": "src/Module/Decoder/Polar/ASCL/Decoder_polar_ASCL_fast_CA_sys.hxx",
"diff": "@@ -7,7 +7,7 @@ Decoder_polar_ASCL_fast_CA_sys<B,R,API_polar>\nCRC<B>& crc, const int n_frames, const std::string name)\n: Decoder_polar_SCL_fast_CA_sys<B,R,API_polar>(K, N, L_max, frozen_bits, crc, n_frames, name),\nsc_decoder (K, N , frozen_bits, n_frames, name),\n- L_max(L_max)\n+ L_max(L_max), is_full_adaptive(true)\n{\nassert(L_max > 0);\n}\n@@ -31,6 +31,8 @@ void Decoder_polar_ASCL_fast_CA_sys<B,R,API_polar>\n// delete the path if the CRC result is negative\nif (!crc_decode_result && L_max > 1)\n+ {\n+ if (is_full_adaptive)\n{\nstd::copy(sc_decoder.l.begin(), sc_decoder.l.begin() + this->N, this->Y_N.begin());\ndo\n@@ -42,6 +44,14 @@ void Decoder_polar_ASCL_fast_CA_sys<B,R,API_polar>\nthis->recursive_decode(off_l, off_s, this->m, first_node_id);\n}\nwhile (!this->select_best_path() && this->L < L_max);\n+\n+ }\n+ else // pseudo adaptive mode\n+ {\n+ this->L = this->L_max;\n+ Decoder_polar_SCL_fast_CA_sys<B,R,API_polar>::load(sc_decoder.l);\n+ Decoder_polar_SCL_fast_CA_sys<B,R,API_polar>::hard_decode();\n+ }\n}\n}\n"
}
] |
C++
|
MIT License
|
aff3ct/aff3ct
|
Add a pseudo adaptive methode in the ASCL decoder (like in the McGill paper).
|
8,490 |
03.01.2017 17:09:06
| -3,600 |
05d4889e3d511b19d4f96e0b01a772f762918653
|
Improve the AList reader.
|
[
{
"change_type": "MODIFY",
"old_path": "conf/dec/LDPC/10GBPS-ETHERNET_1723_2048.alist",
"new_path": "conf/dec/LDPC/10GBPS-ETHERNET_1723_2048.alist",
"diff": "+# LDPC code used in IEEE 802.3an (10Gbps Ethernet)\n2048 384\n6 32\n6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6 6\n"
},
{
"change_type": "MODIFY",
"old_path": "src/Tools/Code/LDPC/AList_reader/AList_reader.cpp",
"new_path": "src/Tools/Code/LDPC/AList_reader/AList_reader.cpp",
"diff": "-#include <fstream>\n+#include <string>\n+#include <sstream>\n+#include <vector>\n-#include \"AList_reader.hpp\"\n#include \"Tools/Display/bash_tools.h\"\n-AList_reader::\n-AList_reader(std::string filename)\n+#include \"AList_reader.hpp\"\n+\n+std::vector<std::string> split(const std::string &s)\n+{\n+ std::string buf; // have a buffer string\n+ std::stringstream ss(s); // insert the string into a stream\n+ std::vector<std::string> tokens; // create vector to hold our words\n+\n+ while (ss >> buf)\n+ tokens.push_back(buf);\n+\n+ return tokens;\n+}\n+\n+AList_reader\n+::AList_reader(std::string filename)\n: n_VN(0), n_CN(0), VN_max_degree(0), CN_max_degree(0), n_branches(0), VN_to_CN(), CN_to_VN()\n{\nstd::ifstream file(filename.c_str(), std::ios::in);\nif (file.is_open())\n+ {\n+ auto success = this->read_format1(file);\n+ file.close();\n+\n+ if (!success)\n+ {\n+ file.open(filename.c_str(), std::ios::in);\n+ success = this->read_format2(file);\n+ file.close();\n+\n+ if (!success)\n+ {\n+ std::cerr << bold_red(\"(EE) The \\\"\")\n+ << bold_red(filename)\n+ << bold_red(\"\\\" file is not supported by \")\n+ << bold_red(\"the AList reader, exiting.\")\n+ << std::endl;\n+ std::exit(-1);\n+ }\n+ }\n+ }\n+ else\n+ {\n+ std::cerr << bold_red(\"(EE) Can't open \\\"\") << bold_red(filename) << bold_red(\"\\\" file, exiting.\") << std::endl;\n+ std::exit(-1);\n+ }\n+}\n+\n+bool AList_reader\n+::read_format1(std::ifstream &file)\n{\nfile >> this->n_VN;\nfile >> this->n_CN;\nfile >> this->VN_max_degree;\nfile >> this->CN_max_degree;\n- assert(this->n_VN > 0 &&\n+ if (this->n_VN > 0 &&\nthis->n_CN > 0 &&\nthis->VN_max_degree > 0 &&\n- this->CN_max_degree > 0);\n-\n+ this->CN_max_degree > 0)\n+ {\nthis->VN_to_CN.resize(this->n_VN);\nfor (auto i = 0; i < (int)this->n_VN; i++)\n{\nint n_connections = 0;\nfile >> n_connections;\n- assert(n_connections > 0);\n-\n+ if (n_connections > 0)\n+ {\nthis->VN_to_CN[i].resize(n_connections);\nthis->n_branches += n_connections;\n}\n+ else\n+ return false;\n+ }\nthis->CN_to_VN.resize(this->n_CN);\nfor (auto i = 0; i < (int)this->n_CN; i++)\n@@ -39,9 +87,10 @@ AList_reader(std::string filename)\nint n_connections = 0;\nfile >> n_connections;\n- assert(n_connections > 0);\n-\n+ if (n_connections > 0)\nthis->CN_to_VN[i].resize(n_connections);\n+ else\n+ return false;\n}\nfor (auto i = 0; i < (int)this->n_VN; i++)\n@@ -51,12 +100,15 @@ AList_reader(std::string filename)\nint C_node_id = 0;\nfile >> C_node_id;\n- assert((C_node_id > 0 && j < (int)this->VN_to_CN[i].size()) ||\n- (C_node_id <= 0 && j >= (int)this->VN_to_CN[i].size()));\n-\n+ if ((C_node_id > 0 && j < (int)this->VN_to_CN[i].size()) ||\n+ (C_node_id <= 0 && j >= (int)this->VN_to_CN[i].size()))\n+ {\nif (C_node_id)\nthis->VN_to_CN[i][j] = C_node_id -1;\n}\n+ else\n+ return false;\n+ }\n}\nfor (auto i = 0; i < (int)this->n_CN; i++)\n@@ -66,26 +118,130 @@ AList_reader(std::string filename)\nint V_node_id = 0;\nfile >> V_node_id;\n- assert((V_node_id > 0 && j < (int)this->CN_to_VN[i].size()) ||\n- (V_node_id <= 0 && j >= (int)this->CN_to_VN[i].size()));\n-\n+ if ((V_node_id > 0 && j < (int)this->CN_to_VN[i].size()) ||\n+ (V_node_id <= 0 && j >= (int)this->CN_to_VN[i].size()))\n+ {\nif (V_node_id)\nthis->CN_to_VN[i][j] = V_node_id -1;\n}\n+ else\n+ return false;\n+ }\n+ }\n+ }\n+ else\n+ return false;\n+\n+ return true;\n}\n- file.close();\n+bool AList_reader\n+::read_format2(std::ifstream &file)\n+{\n+ std::string line;\n+\n+ while (std::getline(file, line)) if (line[0] != '#') break;\n+ if (file.eof() || file.fail() || file.bad()) return false;\n+ auto values = split(line);\n+ if (values.size() < 2) return false;\n+\n+ this->n_VN = std::stoi(values[0]);\n+ this->n_CN = std::stoi(values[1]);\n+\n+ if (this->n_VN <= 0 || this->n_CN <= 0) return false;\n+\n+ while (std::getline(file, line)) if (line[0] != '#') break;\n+ if (file.eof() || file.fail() || file.bad()) return false;\n+ values = split(line);\n+ if (values.size() < 2) return false;\n+\n+ this->VN_max_degree = std::stoi(values[0]);\n+ this->CN_max_degree = std::stoi(values[1]);\n+\n+ if (this->VN_max_degree <= 0 || this->CN_max_degree <= 0) return false;\n+\n+ while (std::getline(file, line)) if (line[0] != '#') break;\n+ if (file.eof() || file.fail() || file.bad()) return false;\n+ values = split(line);\n+ if (values.size() < this->n_VN) return false;\n+\n+ this->VN_to_CN.resize(this->n_VN);\n+ for (auto i = 0; i < (int)this->n_VN; i++)\n+ {\n+ auto n_connections = std::stoi(values[i]);\n+ if (n_connections > 0)\n+ {\n+ this->VN_to_CN[i].resize(n_connections);\n+ this->n_branches += n_connections;\n}\nelse\n+ return false;\n+ }\n+\n+ while (std::getline(file, line)) if (line[0] != '#') break;\n+ if (file.eof() || file.fail() || file.bad()) return false;\n+ values = split(line);\n+ if (values.size() < this->n_CN) return false;\n+\n+ this->CN_to_VN.resize(this->n_CN);\n+ for (auto i = 0; i < (int)this->n_CN; i++)\n{\n- std::cerr << bold_red(\"(EE) Can't open \\\"\") << bold_red(filename) << bold_red(\"\\\" file, exiting.\")\n- << std::endl;\n- std::exit(-1);\n+ auto n_connections = std::stoi(values[i]);\n+ if (n_connections > 0)\n+ this->CN_to_VN[i].resize(n_connections);\n+ else\n+ return false;\n+ }\n+\n+ for (auto i = 0; i < (int)this->n_VN; i++)\n+ {\n+ while (std::getline(file, line)) if (line[0] != '#') break;\n+ if (file.eof() || file.fail() || file.bad()) return false;\n+ values = split(line);\n+\n+ if (values.size() < this->VN_to_CN[i].size()) return false;\n+\n+ for (auto j = 0; j < (int)this->VN_max_degree; j++)\n+ {\n+ auto C_node_id = (j < (int)values.size()) ? std::stoi(values[j]) : 0;\n+ if ((C_node_id > 0 && j < (int)this->VN_to_CN[i].size()) ||\n+ (C_node_id <= 0 && j >= (int)this->VN_to_CN[i].size()))\n+ {\n+ if (C_node_id)\n+ this->VN_to_CN[i][j] = C_node_id -1;\n}\n+ else\n+ return false;\n+ }\n+ }\n+\n+ for (auto i = 0; i < (int)this->n_CN; i++)\n+ {\n+ while (std::getline(file, line)) if (line[0] != '#') break;\n+ if (file.eof() || file.fail() || file.bad()) return false;\n+ values = split(line);\n+\n+ if (values.size() < this->CN_to_VN[i].size()) return false;\n+\n+ for (auto j = 0; j < (int)this->CN_max_degree; j++)\n+ {\n+ auto V_node_id = (j < (int)values.size()) ? std::stoi(values[j]) : 0;\n+ if ((V_node_id > 0 && j < (int)this->CN_to_VN[i].size()) ||\n+ (V_node_id <= 0 && j >= (int)this->CN_to_VN[i].size()))\n+ {\n+ if (V_node_id)\n+ this->CN_to_VN[i][j] = V_node_id -1;\n+ }\n+ else\n+ return false;\n+ }\n+ }\n+\n+ return true;\n}\n-AList_reader::\n-~AList_reader()\n+AList_reader\n+::~AList_reader()\n{\n}\n"
},
{
"change_type": "MODIFY",
"old_path": "src/Tools/Code/LDPC/AList_reader/AList_reader.hpp",
"new_path": "src/Tools/Code/LDPC/AList_reader/AList_reader.hpp",
"diff": "#ifndef ALIST_READER_HPP_\n#define ALIST_READER_HPP_\n+#include <fstream>\n#include <vector>\n#include \"Tools/Perf/MIPP/mipp.h\"\n@@ -34,6 +35,10 @@ public:\nmipp::vector<unsigned int > get_linear_VN_to_CN () const;\nmipp::vector<unsigned int > get_linear_CN_to_VN () const;\nmipp::vector<unsigned int > get_branches_transpose() const;\n+\n+private:\n+ bool read_format1(std::ifstream &file);\n+ bool read_format2(std::ifstream &file);\n};\n#endif /* ALIST_READER_HPP_ */\n"
}
] |
C++
|
MIT License
|
aff3ct/aff3ct
|
Improve the AList reader.
|
8,488 |
03.01.2017 11:19:46
| 18,000 |
1aa22530f334bfbffa162eafb7e543f70bec9459
|
Fix Gaussian Approximation implementation. Code has been given by Peter Trifonov.
|
[
{
"change_type": "MODIFY",
"old_path": "src/Tools/Code/Polar/Frozenbits_generator/Frozenbits_generator_GA.cpp",
"new_path": "src/Tools/Code/Polar/Frozenbits_generator/Frozenbits_generator_GA.cpp",
"diff": "#include \"Frozenbits_generator_GA.hpp\"\n-#ifndef M_PI\n-#define M_PI 3.1415926535897932384626433832795\n-#endif\n-\ntemplate <typename B>\nFrozenbits_generator_GA<B>\n::Frozenbits_generator_GA(const int K, const int N, const float sigma)\n@@ -35,7 +31,7 @@ void Frozenbits_generator_GA<B>\nthis->best_channels[i] = i;\nfor (auto i = 0; i < std::exp2(m); i++)\n- z[i] = 4.0 / std::pow((double)this->sigma, 2.0);\n+ z[i] = 2.0 / std::pow((double)this->sigma, 2.0);\nfor (auto l = 1; l <= m; l++)\n{\n@@ -45,8 +41,11 @@ void Frozenbits_generator_GA<B>\nfor (auto t = 0; t < (int)std::exp2(l - 1); t++)\n{\ndouble T = z[t * o1];\n- z[t * o1 ] = (T < 140) ? phi_inv(1.0 - std::pow(1.0 - phi(T), 2.0)) :\n- T < 2810 ? phi_inv((2.0 - phi(T)) * phi(T)) : (T - 2.77);\n+\n+ z[t * o1 ] = phi_inv(1.0 - std::pow(1.0 - phi(T), 2.0));\n+ if (z[t * o1] == HUGE_VAL)\n+ z[t * o1] = T + M_LN2 / (alpha * gamma);\n+\nz[t * o1 + o2] = 2.0 * T ;\n}\n}\n@@ -59,45 +58,19 @@ double Frozenbits_generator_GA<B>\n::phi(double t)\n{\nif (t < phi_pivot)\n- return std::exp(alpha * std::pow(t, gamma) + beta);\n+ return std::exp(0.0564 * t * t - 0.48560 * t);\nelse // if(t >= phi_pivot)\n- return std::sqrt(M_PI / t) * std::exp(- t / 4.0) * (1.0 - 10.0 / (7.0 * t));\n+ return std::exp(alpha * std::pow(t, gamma) + beta);\n}\ntemplate <typename B>\ndouble Frozenbits_generator_GA<B>\n::phi_inv(double t)\n{\n- double x1, x2, phi_m;\n- if (t >= phi_inv_pivot)\n- {\n- return std::pow(a * std::log(t) + b, c);\n- }\n- else\n- {\n- x1 = 10.0;\n- x2 = bisection_max;\n- while(true)\n- {\n- phi_m = phi((x2 + x1) / 2.0);\n-\n- if (phi_m < t)\n- {\n- if ((t - phi_m) / t < epsilon)\n- return (x2 + x1) / 2.0;\n+ if (t > phi_inv_pivot)\n+ return 4.304964539 * (1 - sqrt(1 + 0.9567131408*std::log(t)));\nelse\n- x2 = (x2 + x1) / 2.0;\n- }\n- else\n- {\n- if ((phi_m - t) / t < epsilon)\n- return (x2 + x1) / 2.0;\n- else\n- x1 = (x2 + x1) / 2.0;\n- }\n- }\n- return INFINITY;\n- }\n+ return std::pow(a * std::log(t) + b, c);\n}\n// ==================================================================================== explicit template instantiation\n"
},
{
"change_type": "MODIFY",
"old_path": "src/Tools/Code/Polar/Frozenbits_generator/Frozenbits_generator_GA.hpp",
"new_path": "src/Tools/Code/Polar/Frozenbits_generator/Frozenbits_generator_GA.hpp",
"diff": "@@ -22,8 +22,8 @@ private:\nconst double b = -beta / alpha;\nconst double c = 1.0 / gamma;\n- const double phi_pivot = 10.0;\n- const double phi_inv_pivot = 0.0394;\n+ const double phi_pivot = 0.867861;\n+ const double phi_inv_pivot = 0.6845772418;\n// const double bisection_max = 1000000000;\nconst double bisection_max = std::numeric_limits<double>::max();\n"
}
] |
C++
|
MIT License
|
aff3ct/aff3ct
|
Fix Gaussian Approximation implementation. Code has been given by Peter Trifonov.
|
8,488 |
03.01.2017 11:22:51
| 18,000 |
5dd4350b66e8440494309f0909cc52dd9c75511f
|
Remove warning message for inaccurate GA after fix.
|
[
{
"change_type": "MODIFY",
"old_path": "src/Tools/Code/Polar/Frozenbits_generator/Frozenbits_generator_GA.cpp",
"new_path": "src/Tools/Code/Polar/Frozenbits_generator/Frozenbits_generator_GA.cpp",
"diff": "@@ -13,8 +13,6 @@ Frozenbits_generator_GA<B>\n::Frozenbits_generator_GA(const int K, const int N, const float sigma)\n: Frozenbits_generator<B>(K, N, sigma), m((int)std::log2(N)), z((int)std::exp2(m), 0)\n{\n- std::cerr << bold_yellow(\"(WW) The GA (Gaussian Approximation) frozen bits generator is inaccurate.\")\n- << std::endl;\n}\ntemplate <typename B>\n"
}
] |
C++
|
MIT License
|
aff3ct/aff3ct
|
Remove warning message for inaccurate GA after fix.
|
8,490 |
04.01.2017 15:13:10
| -3,600 |
7ce7854ce399787407aeebc7c07a03e32fbc1230
|
Improve the AList reader once again.
|
[
{
"change_type": "MODIFY",
"old_path": "src/Tools/Code/LDPC/AList_reader/AList_reader.cpp",
"new_path": "src/Tools/Code/LDPC/AList_reader/AList_reader.cpp",
"diff": "#include <string>\n#include <sstream>\n#include <vector>\n+#include <algorithm>\n#include \"Tools/Display/bash_tools.h\"\n@@ -18,6 +19,16 @@ std::vector<std::string> split(const std::string &s)\nreturn tokens;\n}\n+bool getline(std::ifstream &file, std::string &line)\n+{\n+ while (std::getline(file, line))\n+ if (line[0] != '#' && !std::all_of(line.begin(),line.end(),isspace))\n+ break;\n+ if (file.eof() || file.fail() || file.bad()) return false;\n+\n+ return true;\n+}\n+\nAList_reader\n::AList_reader(std::string filename)\n: n_VN(0), n_CN(0), VN_max_degree(0), CN_max_degree(0), n_branches(0), VN_to_CN(), CN_to_VN()\n@@ -35,6 +46,12 @@ AList_reader\nsuccess = this->read_format2(file);\nfile.close();\n+ if (!success)\n+ {\n+ file.open(filename.c_str(), std::ios::in);\n+ success = this->read_format3(file);\n+ file.close();\n+\nif (!success)\n{\nstd::cerr << bold_red(\"(EE) The \\\"\")\n@@ -46,6 +63,7 @@ AList_reader\n}\n}\n}\n+ }\nelse\n{\nstd::cerr << bold_red(\"(EE) Can't open \\\"\") << bold_red(filename) << bold_red(\"\\\" file, exiting.\") << std::endl;\n@@ -53,6 +71,7 @@ AList_reader\n}\n}\n+// perfect AList format\nbool AList_reader\n::read_format1(std::ifstream &file)\n{\n@@ -135,13 +154,13 @@ bool AList_reader\nreturn true;\n}\n+// format with comments or/and no padding with 0\nbool AList_reader\n::read_format2(std::ifstream &file)\n{\nstd::string line;\n- while (std::getline(file, line)) if (line[0] != '#') break;\n- if (file.eof() || file.fail() || file.bad()) return false;\n+ if (!getline(file, line)) return false;\nauto values = split(line);\nif (values.size() < 2) return false;\n@@ -150,8 +169,7 @@ bool AList_reader\nif (this->n_VN <= 0 || this->n_CN <= 0) return false;\n- while (std::getline(file, line)) if (line[0] != '#') break;\n- if (file.eof() || file.fail() || file.bad()) return false;\n+ if (!getline(file, line)) return false;\nvalues = split(line);\nif (values.size() < 2) return false;\n@@ -160,8 +178,7 @@ bool AList_reader\nif (this->VN_max_degree <= 0 || this->CN_max_degree <= 0) return false;\n- while (std::getline(file, line)) if (line[0] != '#') break;\n- if (file.eof() || file.fail() || file.bad()) return false;\n+ if (!getline(file, line)) return false;\nvalues = split(line);\nif (values.size() < this->n_VN) return false;\n@@ -178,8 +195,7 @@ bool AList_reader\nreturn false;\n}\n- while (std::getline(file, line)) if (line[0] != '#') break;\n- if (file.eof() || file.fail() || file.bad()) return false;\n+ if (!getline(file, line)) return false;\nvalues = split(line);\nif (values.size() < this->n_CN) return false;\n@@ -195,8 +211,7 @@ bool AList_reader\nfor (auto i = 0; i < (int)this->n_VN; i++)\n{\n- while (std::getline(file, line)) if (line[0] != '#') break;\n- if (file.eof() || file.fail() || file.bad()) return false;\n+ if (!getline(file, line)) return false;\nvalues = split(line);\nif (values.size() < this->VN_to_CN[i].size()) return false;\n@@ -217,8 +232,7 @@ bool AList_reader\nfor (auto i = 0; i < (int)this->n_CN; i++)\n{\n- while (std::getline(file, line)) if (line[0] != '#') break;\n- if (file.eof() || file.fail() || file.bad()) return false;\n+ if (!getline(file, line)) return false;\nvalues = split(line);\nif (values.size() < this->CN_to_VN[i].size()) return false;\n@@ -240,6 +254,65 @@ bool AList_reader\nreturn true;\n}\n+// format ANR NAND\n+bool AList_reader\n+::read_format3(std::ifstream &file)\n+{\n+ std::string line;\n+\n+ if (!getline(file, line)) return false;\n+ auto values = split(line);\n+ if (values.size() < 2) return false;\n+\n+ this->n_CN = std::stoi(values[0]);\n+ this->n_VN = std::stoi(values[1]);\n+\n+ if (this->n_VN <= 0 || this->n_CN <= 0) return false;\n+\n+ if (!getline(file, line)) return false;\n+ values = split(line);\n+ if (values.size() < this->n_CN) return false;\n+\n+ this->CN_max_degree = 0;\n+ this->CN_to_VN.resize(this->n_CN);\n+ for (auto i = 0; i < (int)this->n_CN; i++)\n+ {\n+ auto n_connections = std::stoi(values[i]);\n+ this->CN_max_degree = std::max(this->CN_max_degree, (unsigned)n_connections);\n+ if (n_connections > 0)\n+ {\n+ this->CN_to_VN[i].resize(n_connections);\n+ this->n_branches += n_connections;\n+ }\n+ else\n+ return false;\n+ }\n+\n+ this->VN_to_CN.resize(this->n_VN);\n+ for (auto i = 0; i < (int)this->n_CN; i++)\n+ {\n+ if (!getline(file, line)) return false;\n+ values = split(line);\n+\n+ if (values.size() < this->CN_to_VN[i].size()) return false;\n+\n+ for (auto j = 0; j < (int)this->CN_to_VN[i].size(); j++)\n+ {\n+ auto V_node_id = std::stoi(values[j]);\n+ this->CN_to_VN[i][j] = V_node_id;\n+\n+ this->VN_to_CN[V_node_id].push_back(i);\n+ }\n+ }\n+\n+ // find the VN max degree\n+ this->VN_max_degree = 0;\n+ for (auto i = 0; i < (int)this->n_VN; i++)\n+ this->VN_max_degree = std::max(this->VN_max_degree, (unsigned)this->VN_to_CN[i].size());\n+\n+ return true;\n+}\n+\nAList_reader\n::~AList_reader()\n{\n"
},
{
"change_type": "MODIFY",
"old_path": "src/Tools/Code/LDPC/AList_reader/AList_reader.hpp",
"new_path": "src/Tools/Code/LDPC/AList_reader/AList_reader.hpp",
"diff": "@@ -39,6 +39,7 @@ public:\nprivate:\nbool read_format1(std::ifstream &file);\nbool read_format2(std::ifstream &file);\n+ bool read_format3(std::ifstream &file);\n};\n#endif /* ALIST_READER_HPP_ */\n"
}
] |
C++
|
MIT License
|
aff3ct/aff3ct
|
Improve the AList reader once again.
|
8,490 |
05.01.2017 18:10:57
| -3,600 |
cb997ea518786b4bfa215082c052a59a99bae057
|
Fix a bug in the number of check nodes computation.
|
[
{
"change_type": "MODIFY",
"old_path": "src/Module/Decoder/LDPC/BP/Flooding/Decoder_LDPC_BP_flooding.cpp",
"new_path": "src/Module/Decoder/LDPC/BP/Flooding/Decoder_LDPC_BP_flooding.cpp",
"diff": "@@ -17,7 +17,7 @@ Decoder_LDPC_BP_flooding<B,R>\n: Decoder_SISO<B,R> (K, N, n_frames, 1, name ),\nn_ite (n_ite ),\nn_V_nodes (N ), // same as N but more explicit\n- n_C_nodes (N - K ),\n+ n_C_nodes ((int)alist_data.get_n_CN() ),\nn_branches ((int)alist_data.get_n_branches() ),\nenable_syndrome (enable_syndrome ),\ninit_flag (false ),\n"
},
{
"change_type": "MODIFY",
"old_path": "src/Module/Decoder/LDPC/BP/Layered/Decoder_LDPC_BP_layered.cpp",
"new_path": "src/Module/Decoder/LDPC/BP/Layered/Decoder_LDPC_BP_layered.cpp",
"diff": "@@ -15,7 +15,7 @@ Decoder_LDPC_BP_layered<B,R>\nconst std::string name)\n: Decoder_SISO<B,R>(K, N, n_frames, 1, name ),\nn_ite (n_ite ),\n- n_C_nodes (N - K ),\n+ n_C_nodes ((int)alist_data.get_n_CN() ),\nenable_syndrome (enable_syndrome ),\ninit_flag (false ),\nCN_to_VN (alist_data.get_CN_to_VN() ),\n"
}
] |
C++
|
MIT License
|
aff3ct/aff3ct
|
Fix a bug in the number of check nodes computation.
|
8,490 |
06.01.2017 17:13:24
| -3,600 |
d069fc176810bb422366a859501acabb19551603
|
Fix a wrong param in the completion script.
|
[
{
"change_type": "MODIFY",
"old_path": "aff3ct_completion.sh",
"new_path": "aff3ct_completion.sh",
"diff": "@@ -47,7 +47,7 @@ _aff3ct() {\n--sim-inter-lvl --cde-info-bits -K --cde-size -N --src-type \\\n--src-path --enc-type --enc-path --mod-type --mod-bps --mod-ups \\\n--mod-const-path --dmod-max --dmod-no-sig2 --chn-type \\\n- --chn-path --chn-blk-fad --qnt-type --qnt-int --qnt-bits \\\n+ --chn-path --chn-blk-fad --qnt-type --qnt-dec --qnt-bits \\\n--qnt-range --dec-type --dec-implem --term-no --term-freq \\\n--sim-seed --sim-mpi-comm\"\nfi\n@@ -198,7 +198,7 @@ _aff3ct() {\n--sim-snr-min | -m | --snr-min-max | -M | --sim-snr-min | -m | \\\n--snr-min-max | -M | --sim-snr-step | -s | --sim-stop-time | \\\n--sim-threads | -t | --sim-inter-lvl | --cde-info-bits | -K | \\\n- --cde-size | -N | --mod-bps | --mod-ups | --qnt-int | --qnt-bits | \\\n+ --cde-size | -N | --mod-bps | --mod-ups | --qnt-dec | --qnt-bits | \\\n--qnt-range | --qnt-type | --dec-type | --dec-implem | --sim-benchs | \\\n-b | --sim-debug-limit | --mnt-max-fe | -e | --term-type | \\\n--sim-siga-min | -a | --sim-siga-max | -A | --sim-siga-step | \\\n"
}
] |
C++
|
MIT License
|
aff3ct/aff3ct
|
Fix a wrong param in the completion script.
|
8,490 |
09.01.2017 10:30:44
| -3,600 |
8507f747507e26fb4241d136ebfa0718a6af9d5f
|
Fix the LDPC SPA 10GE curve.
|
[
{
"change_type": "MODIFY",
"old_path": "tests/data/BFER/ldpc/BP_FLOODING/SPA/ETH_SPA_1723_2048.txt",
"new_path": "tests/data/BFER/ldpc/BP_FLOODING/SPA/ETH_SPA_1723_2048.txt",
"diff": "@@ -14,6 +14,8 @@ Trace:\n# ** SNR step (s) = 0.200000 dB\n# ** Type of bits = int (32 bits)\n# ** Type of reals = float (32 bits)\n+# ** Inter frame level = 1\n+# ** Seed = 0\n# ** Multi-threading (t) = 56 thread(s)\n# * Code ------------------------------------------\n# ** Type = LDPC\n@@ -38,6 +40,7 @@ Trace:\n# ** Type (D) = BP_FLOODING\n# ** Implementation = SPA\n# ** Num. of iterations (i) = 100\n+# ** Syndrome detection = on\n# * Monitor ---------------------------------------\n# ** Frame error count (e) = 100\n#\n@@ -50,26 +53,27 @@ Trace:\n# SNR | FRA | BE | FE | BER | FER || CTHR | ITHR | LATENCY || SIM_CTHR | ET/RT\n# (dB) | | | | | || (Mb/s) | (Mb/s) | (us) || (Mb/s) | (hhmmss)\n# ------|----------|----------|----------|----------|----------||----------|----------|----------||----------|----------\n- 0.00 | 155 | 26045 | 155 | 9.77e-02 | 1.00e+00 || 0.01 | 0.01 | 1.37e+05 || 0.68 | 00h00'00\n- 0.20 | 155 | 24609 | 155 | 9.23e-02 | 1.00e+00 || 0.02 | 0.01 | 1.36e+05 || 0.70 | 00h00'00\n- 0.40 | 155 | 23262 | 155 | 8.73e-02 | 1.00e+00 || 0.01 | 0.01 | 1.37e+05 || 0.70 | 00h00'00\n- 0.60 | 155 | 22052 | 155 | 8.27e-02 | 1.00e+00 || 0.02 | 0.01 | 1.35e+05 || 0.71 | 00h00'00\n- 0.80 | 155 | 20767 | 155 | 7.79e-02 | 1.00e+00 || 0.02 | 0.01 | 1.35e+05 || 0.70 | 00h00'00\n- 1.00 | 155 | 19469 | 155 | 7.30e-02 | 1.00e+00 || 0.02 | 0.01 | 1.35e+05 || 0.70 | 00h00'00\n- 1.20 | 155 | 18150 | 155 | 6.81e-02 | 1.00e+00 || 0.02 | 0.01 | 1.36e+05 || 0.70 | 00h00'00\n- 1.40 | 155 | 16942 | 155 | 6.35e-02 | 1.00e+00 || 0.02 | 0.01 | 1.34e+05 || 0.70 | 00h00'00\n- 1.60 | 155 | 15656 | 155 | 5.87e-02 | 1.00e+00 || 0.02 | 0.01 | 1.36e+05 || 0.70 | 00h00'00\n- 1.80 | 155 | 14557 | 155 | 5.46e-02 | 1.00e+00 || 0.02 | 0.01 | 1.35e+05 || 0.70 | 00h00'00\n- 2.00 | 155 | 13427 | 155 | 5.04e-02 | 1.00e+00 || 0.02 | 0.01 | 1.35e+05 || 0.70 | 00h00'00\n- 2.20 | 155 | 12232 | 155 | 4.59e-02 | 1.00e+00 || 0.02 | 0.01 | 1.35e+05 || 0.70 | 00h00'00\n- 2.40 | 155 | 11034 | 155 | 4.14e-02 | 1.00e+00 || 0.01 | 0.01 | 1.37e+05 || 0.70 | 00h00'00\n- 2.60 | 156 | 9832 | 155 | 3.66e-02 | 9.94e-01 || 0.02 | 0.01 | 1.36e+05 || 0.70 | 00h00'00\n- 2.80 | 160 | 8608 | 153 | 3.13e-02 | 9.56e-01 || 0.02 | 0.01 | 1.33e+05 || 0.71 | 00h00'00\n- 3.00 | 184 | 7483 | 150 | 2.36e-02 | 8.15e-01 || 0.02 | 0.01 | 1.19e+05 || 0.80 | 00h00'00\n- 3.20 | 267 | 6447 | 148 | 1.40e-02 | 5.54e-01 || 0.02 | 0.02 | 88928.36 || 1.08 | 00h00'00\n- 3.40 | 647 | 5660 | 144 | 5.09e-03 | 2.23e-01 || 0.04 | 0.04 | 45980.70 || 2.11 | 00h00'00\n- 3.60 | 2273 | 4708 | 121 | 1.20e-03 | 5.32e-02 || 0.10 | 0.08 | 21047.45 || 4.82 | 00h00'00\n- 3.80 | 15079 | 4064 | 103 | 1.57e-04 | 6.83e-03 || 0.18 | 0.15 | 11587.76 || 9.51 | 00h00'03\n- 4.00 | 129191 | 3850 | 101 | 1.73e-05 | 7.82e-04 || 0.24 | 0.20 | 8693.56 || 12.92 | 00h00'20\n- 4.20 | 2064210 | 3730 | 100 | 1.05e-06 | 4.84e-05 || 0.28 | 0.23 | 7387.95 || 15.24 | 00h04'37\n- 4.40 | 60240756 | 3359 | 100 | 3.24e-08 | 1.66e-06 || 0.31 | 0.26 | 6558.97 || 17.14 | 01h59'59\n\\ No newline at end of file\n+ 0.00 | 155 | 26111 | 155 | 9.78e-02 | 1.00e+00 || 0.01 | 0.01 | 1.40e+05 || 0.55 | 00h00'00\n+ 0.20 | 155 | 24690 | 155 | 9.24e-02 | 1.00e+00 || 0.02 | 0.01 | 1.36e+05 || 0.71 | 00h00'00\n+ 0.40 | 155 | 23292 | 155 | 8.72e-02 | 1.00e+00 || 0.01 | 0.01 | 1.41e+05 || 0.68 | 00h00'00\n+ 0.60 | 155 | 22035 | 155 | 8.25e-02 | 1.00e+00 || 0.01 | 0.01 | 1.38e+05 || 0.70 | 00h00'00\n+ 0.80 | 155 | 20815 | 155 | 7.79e-02 | 1.00e+00 || 0.01 | 0.01 | 1.37e+05 || 0.70 | 00h00'00\n+ 1.00 | 155 | 19429 | 155 | 7.28e-02 | 1.00e+00 || 0.02 | 0.01 | 1.35e+05 || 0.71 | 00h00'00\n+ 1.20 | 155 | 18154 | 155 | 6.80e-02 | 1.00e+00 || 0.02 | 0.01 | 1.36e+05 || 0.71 | 00h00'00\n+ 1.40 | 155 | 16872 | 155 | 6.32e-02 | 1.00e+00 || 0.01 | 0.01 | 1.39e+05 || 0.66 | 00h00'00\n+ 1.60 | 155 | 15672 | 155 | 5.87e-02 | 1.00e+00 || 0.02 | 0.01 | 1.36e+05 || 0.71 | 00h00'00\n+ 1.80 | 155 | 14492 | 155 | 5.43e-02 | 1.00e+00 || 0.02 | 0.01 | 1.35e+05 || 0.69 | 00h00'00\n+ 2.00 | 155 | 13357 | 155 | 5.00e-02 | 1.00e+00 || 0.01 | 0.01 | 1.37e+05 || 0.69 | 00h00'00\n+ 2.20 | 155 | 12186 | 155 | 4.56e-02 | 1.00e+00 || 0.01 | 0.01 | 1.37e+05 || 0.68 | 00h00'00\n+ 2.40 | 155 | 10921 | 155 | 4.09e-02 | 1.00e+00 || 0.01 | 0.01 | 1.37e+05 || 0.69 | 00h00'00\n+ 2.60 | 158 | 9668 | 155 | 3.55e-02 | 9.81e-01 || 0.02 | 0.01 | 1.35e+05 || 0.69 | 00h00'00\n+ 2.80 | 176 | 8596 | 154 | 2.83e-02 | 8.75e-01 || 0.02 | 0.01 | 1.23e+05 || 0.75 | 00h00'00\n+ 3.00 | 254 | 7690 | 151 | 1.76e-02 | 5.94e-01 || 0.02 | 0.02 | 95080.34 || 1.00 | 00h00'00\n+ 3.20 | 576 | 6709 | 142 | 6.76e-03 | 2.47e-01 || 0.04 | 0.03 | 50027.80 || 1.87 | 00h00'00\n+ 3.40 | 1798 | 5417 | 120 | 1.75e-03 | 6.67e-02 || 0.09 | 0.08 | 22476.26 || 4.30 | 00h00'00\n+ 3.60 | 10712 | 4575 | 107 | 2.48e-04 | 9.99e-03 || 0.17 | 0.15 | 11758.46 || 9.17 | 00h00'02\n+ 3.80 | 112048 | 4385 | 102 | 2.27e-05 | 9.10e-04 || 0.25 | 0.21 | 8308.16 || 13.52 | 00h00'16\n+ 4.00 | 1371034 | 4238 | 100 | 1.79e-06 | 7.29e-05 || 0.30 | 0.25 | 6927.05 || 16.28 | 00h02'52\n+ 4.20 | 37471528 | 4033 | 100 | 6.25e-08 | 2.67e-06 || 0.33 | 0.28 | 6126.49 || 18.37 | 01h09'36\n+ 4.40 | 1.12e+09 | 3016 | 100 | 1.56e-09 | 8.93e-08 || 0.37 | 0.31 | 5580.79 || 20.13 | 31h39'07\n+# End of the simulation.\n\\ No newline at end of file\n"
}
] |
C++
|
MIT License
|
aff3ct/aff3ct
|
Fix the LDPC SPA 10GE curve.
|
8,490 |
09.01.2017 13:54:07
| -3,600 |
78c34a94aaf827351e02bedd82012b4393ac70be
|
Improve a little bit the alist reader.
|
[
{
"change_type": "MODIFY",
"old_path": "src/Tools/Code/LDPC/AList_reader/AList_reader.cpp",
"new_path": "src/Tools/Code/LDPC/AList_reader/AList_reader.cpp",
"diff": "@@ -31,7 +31,7 @@ bool getline(std::ifstream &file, std::string &line)\nAList_reader\n::AList_reader(std::string filename)\n-: n_VN(0), n_CN(0), VN_max_degree(0), CN_max_degree(0), n_branches(0), VN_to_CN(), CN_to_VN()\n+: n_VN(0), n_CN(0), VN_max_degree(0), CN_max_degree(0), n_branches(0), VN_to_CN(), CN_to_VN(), branches_transpose()\n{\nstd::ifstream file(filename.c_str(), std::ios::in);\n@@ -69,6 +69,8 @@ AList_reader\nstd::cerr << bold_red(\"(EE) Can't open \\\"\") << bold_red(filename) << bold_red(\"\\\" file, exiting.\") << std::endl;\nstd::exit(-1);\n}\n+\n+ this->compute_branches_transpose();\n}\n// perfect AList format\n@@ -408,11 +410,10 @@ mipp::vector<unsigned int> AList_reader\nreturn linear_C_to_V;\n}\n-mipp::vector<unsigned int> AList_reader\n-::get_branches_transpose() const\n+void AList_reader\n+::compute_branches_transpose()\n{\n- mipp::vector<unsigned int> transpose(this->n_branches);\n-\n+ branches_transpose.resize(this->get_n_branches());\nmipp::vector<unsigned char> connections(VN_to_CN.size(), 0);\nauto k = 0;\n@@ -430,10 +431,14 @@ mipp::vector<unsigned int> AList_reader\nassert(connections[id_V] <= (int)VN_to_CN[id_V].size());\n- transpose[k] = branch_id;\n+ branches_transpose[k] = branch_id;\nk++;\n}\n}\n+}\n- return transpose;\n+const mipp::vector<unsigned int>& AList_reader\n+::get_branches_transpose() const\n+{\n+ return branches_transpose;\n}\n"
},
{
"change_type": "MODIFY",
"old_path": "src/Tools/Code/LDPC/AList_reader/AList_reader.hpp",
"new_path": "src/Tools/Code/LDPC/AList_reader/AList_reader.hpp",
"diff": "@@ -16,6 +16,7 @@ private:\nstd::vector<std::vector<unsigned int>> VN_to_CN;\nstd::vector<std::vector<unsigned int>> CN_to_VN;\n+ mipp::vector<unsigned int> branches_transpose;\npublic:\nAList_reader(std::string filename);\n@@ -34,12 +35,13 @@ public:\nmipp::vector<unsigned char> get_n_CN_per_VN () const;\nmipp::vector<unsigned int > get_linear_VN_to_CN () const;\nmipp::vector<unsigned int > get_linear_CN_to_VN () const;\n- mipp::vector<unsigned int > get_branches_transpose() const;\n+ const mipp::vector<unsigned int >& get_branches_transpose() const;\nprivate:\nbool read_format1(std::ifstream &file);\nbool read_format2(std::ifstream &file);\nbool read_format3(std::ifstream &file);\n+ void compute_branches_transpose();\n};\n#endif /* ALIST_READER_HPP_ */\n"
}
] |
C++
|
MIT License
|
aff3ct/aff3ct
|
Improve a little bit the alist reader.
|
8,490 |
10.01.2017 10:51:08
| -3,600 |
3e52e9ff1ad1a7d4649e4037936bee375c8d3694
|
Fix a bug in MIPP (in SSE2 mode).
|
[
{
"change_type": "MODIFY",
"old_path": "src/Tools/Perf/MIPP/mipp_impl_SSE.hxx",
"new_path": "src/Tools/Perf/MIPP/mipp_impl_SSE.hxx",
"diff": "#endif\n// ---------------------------------------------------------------------------------------------------------- shuff\n-#ifdef __SSE3__\n+#ifdef __SSSE3__\ntemplate <>\ninline reg shuff<float>(const reg v, const reg cm) {\nreturn _mm_castsi128_ps(_mm_shuffle_epi8(_mm_castps_si128(v), _mm_castps_si128(cm)));\n#endif\n// --------------------------------------------------------------------------------------------------------- shuff2\n-#ifdef __SSE3__\n+#ifdef __SSSE3__\ntemplate <>\ninline reg shuff2<float>(const reg v, const reg cm) {\nreturn mipp::shuff<float>(v, cm);\n"
}
] |
C++
|
MIT License
|
aff3ct/aff3ct
|
Fix a bug in MIPP (in SSE2 mode).
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.