-
Notifications
You must be signed in to change notification settings - Fork 70
/
main.go
270 lines (239 loc) · 9.78 KB
/
main.go
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
246
247
248
249
250
251
252
253
254
255
256
257
258
259
260
261
262
263
264
265
266
267
268
269
270
/*
Copyright 2021 The VolSync authors.
This program is free software: you can redistribute it and/or modify
it under the terms of the GNU Affero General Public License as published
by the Free Software Foundation, either version 3 of the License, or
(at your option) any later version.
This program is distributed in the hope that it will be useful,
but WITHOUT ANY WARRANTY; without even the implied warranty of
MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
GNU Affero General Public License for more details.
You should have received a copy of the GNU Affero General Public License
along with this program. If not, see <https://www.gnu.org/licenses/>.
*/
package main
import (
"bufio"
"context"
"flag"
"fmt"
"os"
"runtime"
"time"
// Import all Kubernetes client auth plugins (e.g. Azure, GCP, OIDC, etc.)
// to ensure that exec-entrypoint and run can make use of them.
_ "k8s.io/client-go/plugin/pkg/client/auth"
"k8s.io/client-go/rest"
_ "embed"
snapv1 "github.com/kubernetes-csi/external-snapshotter/client/v8/apis/volumesnapshot/v1"
volumepopulatorv1beta1 "github.com/kubernetes-csi/volume-data-source-validator/client/apis/volumepopulator/v1beta1"
ocpsecurityv1 "github.com/openshift/api/security/v1"
"github.com/spf13/pflag"
"github.com/spf13/viper"
"go.uber.org/zap/zapcore"
kruntime "k8s.io/apimachinery/pkg/runtime"
utilruntime "k8s.io/apimachinery/pkg/util/runtime"
clientgoscheme "k8s.io/client-go/kubernetes/scheme"
ctrl "sigs.k8s.io/controller-runtime"
"sigs.k8s.io/controller-runtime/pkg/client"
"sigs.k8s.io/controller-runtime/pkg/healthz"
"sigs.k8s.io/controller-runtime/pkg/log/zap"
"sigs.k8s.io/controller-runtime/pkg/manager"
metricsserver "sigs.k8s.io/controller-runtime/pkg/metrics/server"
volsyncv1alpha1 "github.com/backube/volsync/api/v1alpha1"
"github.com/backube/volsync/controllers"
"github.com/backube/volsync/controllers/mover"
"github.com/backube/volsync/controllers/platform"
"github.com/backube/volsync/controllers/utils"
//+kubebuilder:scaffold:imports
)
var (
scheme = kruntime.NewScheme()
setupLog = ctrl.Log.WithName("setup")
volsyncVersion = "0.0.0"
//go:embed config/openshift/mover_scc.yaml
volsyncMoverSCCYamlRaw []byte
// See each mover_<movertype>_register.go where they add themselves to
// enabledMovers
enabledMovers = []func() error{}
)
func init() {
utilruntime.Must(clientgoscheme.AddToScheme(scheme))
utilruntime.Must(snapv1.AddToScheme(scheme))
utilruntime.Must(volsyncv1alpha1.AddToScheme(scheme))
utilruntime.Must(ocpsecurityv1.AddToScheme(scheme))
utilruntime.Must(volumepopulatorv1beta1.AddToScheme(scheme))
//+kubebuilder:scaffold:scheme
}
// registerMovers Registers the data movers to be used by the VolSync operator.
func registerMovers() error {
// logger isn't initialized yet, write to stdout
bufout := bufio.NewWriter(os.Stdout)
defer bufout.Flush()
for _, moverRegisterFunc := range enabledMovers {
if err := moverRegisterFunc(); err != nil {
return err
}
}
fmt.Fprintf(bufout, "Registered Movers: %v\n", mover.GetEnabledMoverList())
return nil
}
// printInfo Prints version information about the operating system and movers to STDOUT.
func printInfo() {
setupLog.Info(fmt.Sprintf("Go Version: %s", runtime.Version()))
setupLog.Info(fmt.Sprintf("Go OS/Arch: %s/%s", runtime.GOOS, runtime.GOARCH))
setupLog.Info(fmt.Sprintf("Operator Version: %s", volsyncVersion))
for _, b := range mover.Catalog {
setupLog.Info(b.VersionInfo())
}
}
// configureChecks Configures the manager with a healthz and readyz check.
func configureChecks(mgr manager.Manager) error {
if err := mgr.AddHealthzCheck("healthz", healthz.Ping); err != nil {
return fmt.Errorf("unable to setup health check: %w", err)
}
if err := mgr.AddReadyzCheck("readyz", healthz.Ping); err != nil {
return fmt.Errorf("unable to setup ready check: %w", err)
}
return nil
}
// addCommandFlags Configures flags to be bound to the VolSync command.
func addCommandFlags(probeAddr *string, metricsAddr *string, enableLeaderElection *bool) {
flag.StringVar(metricsAddr, "metrics-bind-address", ":8080", "The address the metric endpoint binds to.")
flag.StringVar(probeAddr, "health-probe-bind-address", ":8081", "The address the probe endpoint binds to.")
flag.BoolVar(enableLeaderElection, "leader-elect", false,
"Enable leader election for controller manager. "+
"Enabling this will ensure there is only one active controller manager.")
flag.StringVar(&utils.SCCName, "scc-name",
utils.DefaultSCCName, "The name of the volsync security context constraint")
opts := zap.Options{
Development: true,
TimeEncoder: zapcore.ISO8601TimeEncoder,
}
opts.BindFlags(flag.CommandLine)
// Import flags into pflag so they can be bound by viper
pflag.CommandLine.AddGoFlagSet(flag.CommandLine)
pflag.Parse()
if err := viper.BindPFlags(pflag.CommandLine); err != nil {
setupLog.Error(err, "Unable to bind command line flags")
os.Exit(1)
}
ctrl.SetLogger(zap.New(zap.UseFlagOptions(&opts)))
}
// Prereq CRs we want to always be present in certain environments but do not want to reconcile often (just at startup)
func ensureCRs(cfg *rest.Config) {
setupClient, err := client.New(cfg, client.Options{Scheme: scheme})
if err != nil {
setupLog.Error(err, "error creating client")
os.Exit(1)
}
// Privileged mover SCC required in OpenShift envs
setupLog.Info("Privileged Mover SCC", "scc-name", utils.SCCName)
err = platform.EnsureVolSyncMoverSCCIfOpenShift(context.Background(), setupClient, setupLog,
utils.SCCName, volsyncMoverSCCYamlRaw)
if err != nil {
setupLog.Error(err, "unable to reconcile volsync mover scc", "scc-name", utils.SCCName)
os.Exit(1)
}
// VolumePopulator CR should be registered if the VolumePopulator CRD is present
err = controllers.EnsureVolSyncVolumePopulatorCRIfCRDPresent(context.Background(), setupClient, setupLog)
if err != nil {
setupLog.Error(err, "unable to reconcile VolumePopulator CR")
os.Exit(1)
}
}
func initPodLogsClient(cfg *rest.Config) {
_, err := utils.InitPodLogsClient(cfg)
if err != nil {
setupLog.Error(err, "unable to create client-go clientset for pod logs")
os.Exit(1)
}
setupLog.Info("Mover Status Log", "log max bytes", utils.GetMoverLogMaxBytes(),
"tail lines", utils.GetMoverLogTailLines(), "debug", utils.IsMoverLogDebug())
}
// nolint: funlen
func main() {
err := registerMovers()
if err != nil {
fmt.Printf("error registering data movers: %+v", err)
os.Exit(1)
}
var probeAddr, metricsAddr string
var enableLeaderElection bool
addCommandFlags(&probeAddr, &metricsAddr, &enableLeaderElection)
printInfo()
leaseDuration := 137 * time.Second
renewDeadline := 107 * time.Second
retryPeriod := 26 * time.Second
cfg := ctrl.GetConfigOrDie()
mgr, err := ctrl.NewManager(cfg, ctrl.Options{
Scheme: scheme,
Metrics: metricsserver.Options{BindAddress: metricsAddr},
HealthProbeBindAddress: probeAddr,
LeaderElection: enableLeaderElection,
LeaderElectionID: "b95b3104.backube",
// LeaderElectionReleaseOnCancel defines if the leader should step down voluntarily
// when the Manager ends. This requires the binary to immediately end when the
// Manager is stopped, otherwise, this setting is unsafe. Setting this significantly
// speeds up voluntary leader transitions as the new leader don't have to wait
// LeaseDuration time first.
LeaderElectionReleaseOnCancel: true,
LeaseDuration: &leaseDuration,
RenewDeadline: &renewDeadline,
RetryPeriod: &retryPeriod,
})
if err != nil {
setupLog.Error(err, "unable to start manager")
os.Exit(1)
}
// Before starting controllers - create or patch volsync mover SCC and VolumePopulator CR if necessary
ensureCRs(cfg)
initPodLogsClient(cfg)
// Index fields that are required for the ReplicationSource controller
if err := controllers.IndexFieldsForReplicationSource(context.Background(), mgr.GetFieldIndexer()); err != nil {
setupLog.Error(err, "unable to index fields for controller", "controller", "ReplicationSource")
os.Exit(1)
}
if err = (&controllers.ReplicationSourceReconciler{
Client: mgr.GetClient(),
Log: ctrl.Log.WithName("controllers").WithName("ReplicationSource"),
Scheme: mgr.GetScheme(),
EventRecorder: mgr.GetEventRecorderFor("volsync-controller"),
}).SetupWithManager(mgr); err != nil {
setupLog.Error(err, "unable to create controller", "controller", "ReplicationSource")
os.Exit(1)
}
if err = (&controllers.ReplicationDestinationReconciler{
Client: mgr.GetClient(),
Log: ctrl.Log.WithName("controllers").WithName("ReplicationDestination"),
Scheme: mgr.GetScheme(),
EventRecorder: mgr.GetEventRecorderFor("volsync-controller"),
}).SetupWithManager(mgr); err != nil {
setupLog.Error(err, "unable to create controller", "controller", "ReplicationDestination")
os.Exit(1)
}
// Index fields that are required for the VolumePopulator controller
if err := controllers.IndexFieldsForVolumePopulator(context.Background(), mgr.GetFieldIndexer()); err != nil {
setupLog.Error(err, "unable to index fields for controller", "controller", "VolumePopulator")
os.Exit(1)
}
if err = (&controllers.VolumePopulatorReconciler{
Client: mgr.GetClient(),
Log: ctrl.Log.WithName("controllers").WithName("VolumePopulator"),
Scheme: mgr.GetScheme(),
EventRecorder: mgr.GetEventRecorderFor("volsync-controller"),
}).SetupWithManager(mgr); err != nil {
setupLog.Error(err, "unable to create controller", "controller", "VolumePopulator")
os.Exit(1)
}
//+kubebuilder:scaffold:builder
if err := configureChecks(mgr); err != nil {
setupLog.Error(err, "unable to setup checks")
os.Exit(1)
}
setupLog.Info("starting manager")
if err := mgr.Start(ctrl.SetupSignalHandler()); err != nil {
setupLog.Error(err, "problem running manager")
os.Exit(1)
}
}