diff --git a/README.md b/README.md index 6fd5f355f1..f1603dc72d 100644 --- a/README.md +++ b/README.md @@ -18,7 +18,7 @@ It allows to easily manage Kubernetes clusters that are natively integrated with * Architecured to be operated as a managed service * Masters are managed centrally - * Nodes are dececentralized in customer's projects + * Nodes are decentralized in customer's projects * 100% Vanilla Kubernetes * 100% Compatible Openstack API * Air-Gapped Masters and Nodes diff --git a/docs/development/changing_docs.md b/docs/development/changing_docs.md index 5d51c743e4..b08aa3adab 100644 --- a/docs/development/changing_docs.md +++ b/docs/development/changing_docs.md @@ -6,7 +6,7 @@ title: Changing Docs The self-hosted documentation served by the apiserver is generated using Hugo. -It constists of 2 parts: +It consists of 2 parts: * Hugo Theme in `contrib/kubernikus-docs-builder/data` * Markdown docs in `docs` diff --git a/docs/development/helm_dev.md b/docs/development/helm_dev.md index f654b5ce02..ce32cb9cd0 100644 --- a/docs/development/helm_dev.md +++ b/docs/development/helm_dev.md @@ -17,7 +17,7 @@ You will need to provide all of the options which are: ``` --auth-project-domain monsoon3``` the domain of the project ``` --auth-username techuser1``` the user with which to interact -### Chaging values.yaml +### Changing values.yaml After the fact some values have to be added to the ```values.yaml```: ``` api: @@ -37,7 +37,7 @@ The api part represents the host of the ingress definition and has to have some These openstack things need also be added: ```region``` the region in which this is running -```lbSubnetID``` the private _subnet_ to which the loadbalancer should be added +```lbSubnetID``` the private _subnet_ to which the load balancer should be added ```routerID``` the id of your router. ### The user diff --git a/docs/features.md b/docs/features.md index a1d533bc44..9455bafb10 100644 --- a/docs/features.md +++ b/docs/features.md @@ -7,7 +7,7 @@ type: index * Architecured to be operated as a managed service * Masters are managed centrally - * Nodes are dececentralized in customer's projects + * Nodes are decentralized in customer's projects ## Compliant diff --git a/docs/guide/best_practices.md b/docs/guide/best_practices.md index d9bf4bafa9..f0fcd3597a 100644 --- a/docs/guide/best_practices.md +++ b/docs/guide/best_practices.md @@ -8,4 +8,4 @@ title: Best Practices ## Automating Authentication Refresh -## Integrationn for CI Systems +## Integration for CI Systems diff --git a/pkg/api/auth/authorizer_test.go b/pkg/api/auth/authorizer_test.go index a99a1795c9..3cd14ea36a 100644 --- a/pkg/api/auth/authorizer_test.go +++ b/pkg/api/auth/authorizer_test.go @@ -36,10 +36,10 @@ func TestAuthorizer(t *testing.T) { assert.NoError(t, err) req := httptest.NewRequest("GET", "/api/v1/clusters", nil) - assert.NoError(t, authorizer.Authorize(req, &admin), "admin can list clustes") + assert.NoError(t, authorizer.Authorize(req, &admin), "admin can list clusters") assert.NoError(t, authorizer.Authorize(req, &user), "user can list clusters") req = httptest.NewRequest("POST", "/api/v1/clusters", nil) - assert.NoError(t, authorizer.Authorize(req, &admin), "admin can create clustes") - assert.Error(t, authorizer.Authorize(req, &user), "user can not create clustes") + assert.NoError(t, authorizer.Authorize(req, &admin), "admin can create clusters") + assert.Error(t, authorizer.Authorize(req, &user), "user can not create clusters") } diff --git a/pkg/api/rest/api_test.go b/pkg/api/rest/api_test.go index 77ada3da0f..a860460cc1 100644 --- a/pkg/api/rest/api_test.go +++ b/pkg/api/rest/api_test.go @@ -208,7 +208,7 @@ func TestClusterUpdate(t *testing.T) { }, } jsonPayload, err := updateObject.MarshalBinary() - assert.NoError(t, err, "marsheling update payload failed") + assert.NoError(t, err, "marshaling update payload failed") req := createRequest("PUT", "/api/v1/clusters/nase", string(jsonPayload)) code, _, body := result(handler, req) if !assert.Equal(t, 200, code) { diff --git a/pkg/apis/kubernikus/factory.go b/pkg/apis/kubernikus/factory.go index 0d0a73829e..95849afc29 100644 --- a/pkg/apis/kubernikus/factory.go +++ b/pkg/apis/kubernikus/factory.go @@ -84,7 +84,7 @@ func (klusterFactory) KlusterFor(spec models.KlusterSpec) (*v1.Kluster, error) { } if k.Spec.Version != util.DEFAULT_KUBERNETES_VERSION { - return nil, fmt.Errorf("Unabled to create cluster. Unsupported Kubernetes version.") + return nil, fmt.Errorf("Unable to create cluster. Unsupported Kubernetes version.") } if k.ObjectMeta.Name == "" { diff --git a/pkg/client/helm/client.go b/pkg/client/helm/client.go index 1a595900d1..58ea38e658 100644 --- a/pkg/client/helm/client.go +++ b/pkg/client/helm/client.go @@ -39,7 +39,7 @@ func NewClient(kubeClient kubernetes.Interface, kubeConfig *rest.Config, logger //Lets see how this goes: We close the tunnel as soon as the client is GC'ed. runtime.SetFinalizer(client, func(_ *helm.Client) { logger.Log( - "msg", "trearing down tunnel to tiller", + "msg", "tearing down tunnel to tiller", "host", tillerHost, "v", 2) tunnel.Close() diff --git a/pkg/client/openstack/client.go b/pkg/client/openstack/client.go index f35a52f39a..1a0f6ddf42 100644 --- a/pkg/client/openstack/client.go +++ b/pkg/client/openstack/client.go @@ -497,7 +497,7 @@ func (c *client) getDomainID(client *gophercloud.ServiceClient, domainName strin c.domainNameToID.Store(domainName, domains[0].ID) return false, nil default: - return false, errors.New("More then one domain found") + return false, errors.New("More than one domain found") } }) id, _ := c.domainNameToID.Load(domainName) diff --git a/pkg/cmd/kubernikus/certificates/sign.go b/pkg/cmd/kubernikus/certificates/sign.go index 3d1bd2d9d2..4903cbaf8e 100644 --- a/pkg/cmd/kubernikus/certificates/sign.go +++ b/pkg/cmd/kubernikus/certificates/sign.go @@ -57,7 +57,7 @@ func (o *SignOptions) BindFlags(flags *pflag.FlagSet) { flags.StringVar(&o.KubeConfig, "kubeconfig", o.KubeConfig, "Path to the kubeconfig file to use to talk to the Kubernetes apiserver. If unset, try the environment variable KUBECONFIG, as well as in-cluster configuration") flags.StringVar(&o.Namespace, "namespace", o.Namespace, "Namespace where the kluster is located") flags.StringVar(&o.CN, "cn", o.CN, "Common name in the certificate") - flags.StringVar(&o.Organization, "organizaion", o.Organization, "Common name in the certificate") + flags.StringVar(&o.Organization, "organization", o.Organization, "Common name in the certificate") flags.StringVar(&o.ApiURL, "api-url", o.ApiURL, "URL for the apiserver") } diff --git a/pkg/cmd/kubernikusctl/auth/refresh.go b/pkg/cmd/kubernikusctl/auth/refresh.go index 24da5c7eec..425bb39e32 100644 --- a/pkg/cmd/kubernikusctl/auth/refresh.go +++ b/pkg/cmd/kubernikusctl/auth/refresh.go @@ -104,7 +104,7 @@ func (o *RefreshOptions) Run(c *cobra.Command) error { } if ok, err := o.isCertificateValid(); err != nil { - return errors.Wrap(err, "Verification of certifcates failed.") + return errors.Wrap(err, "Verification of certificates failed.") } else { if ok && !o.force { glog.V(2).Infof("Certificates are good. Doing nothing.") @@ -200,7 +200,7 @@ func (o *RefreshOptions) loadKubeconfig() (err error) { } func (o *RefreshOptions) isKubernikusContext() (bool, error) { - caCert, err := o.getCACertifciate() + caCert, err := o.getCACertificate() if err != nil { return false, err } @@ -235,7 +235,7 @@ func (o *RefreshOptions) autoDetectKubernikusClientMetadata() (string, string, e parts := strings.Split(cert.Subject.CommonName, "@") if len(parts) != 2 { - return "", "", errors.Errorf("Couldln't extract username/domain from client certificate %v", parts) + return "", "", errors.Errorf("Couldn't extract username/domain from client certificate %v", parts) } return parts[0], parts[1], nil @@ -317,7 +317,7 @@ func (o *RefreshOptions) getRawCACertificate() ([]byte, error) { return certData, nil } -func (o *RefreshOptions) getCACertifciate() (*x509.Certificate, error) { +func (o *RefreshOptions) getCACertificate() (*x509.Certificate, error) { data, err := o.getRawCACertificate() if err != nil { return nil, err diff --git a/pkg/cmd/kubernikusctl/get/common.go b/pkg/cmd/kubernikusctl/get/common.go index 4f324641cf..29925d587a 100644 --- a/pkg/cmd/kubernikusctl/get/common.go +++ b/pkg/cmd/kubernikusctl/get/common.go @@ -39,7 +39,7 @@ func (o *GetOptions) SetupKubernikusClient() error { if o._url == "" { fmt.Println("Auto-Detecting Kubernikus Host ...") if o.url, err = o.Openstack.DefaultKubernikusURL(); err != nil { - glog.V(2).Infof("Error detecting kubernikust host: %+v", err) + glog.V(2).Infof("Error detecting kubernikus host: %+v", err) return errors.Errorf("You need to provide --url. Auto-Detection failed.") } } else { diff --git a/pkg/cmd/kubernikusctl/get/nodepool.go b/pkg/cmd/kubernikusctl/get/nodepool.go index c51fcff289..56274d4fbb 100644 --- a/pkg/cmd/kubernikusctl/get/nodepool.go +++ b/pkg/cmd/kubernikusctl/get/nodepool.go @@ -63,7 +63,7 @@ func validateNodePoolCommandArgs(args []string) error { return errors.Errorf("Surplus arguments to nodepool, %v", args) } if len(args) < 1 { - return errors.Errorf("No clustername given, %v", args) + return errors.Errorf("No cluster name given, %v", args) } return nil } diff --git a/pkg/cmd/kubernikusctl/kubernikusctl.go b/pkg/cmd/kubernikusctl/kubernikusctl.go index 6e323c2f32..c7a56496e3 100644 --- a/pkg/cmd/kubernikusctl/kubernikusctl.go +++ b/pkg/cmd/kubernikusctl/kubernikusctl.go @@ -10,7 +10,7 @@ func NewCommand(name string) *cobra.Command { c := &cobra.Command{ Use: name, Short: "Kubernikus Kubectl Plugin", - Long: "Plugin that extends kubectl with Kubernikus convinience features", + Long: "Plugin that extends kubectl with Kubernikus convenience features", } c.AddCommand( diff --git a/pkg/controller/base/controller.go b/pkg/controller/base/controller.go index 59ba05e10f..8afcfc3c54 100644 --- a/pkg/controller/base/controller.go +++ b/pkg/controller/base/controller.go @@ -21,7 +21,7 @@ const ( KLUSTER_RECHECK_INTERVAL = 5 * time.Minute ) -var ErrUnkownKluster = errors.New("unkown kluster") +var ErrUnkownKluster = errors.New("unknown kluster") type Controller interface { Run(int, <-chan struct{}, *sync.WaitGroup) diff --git a/pkg/controller/events/event.go b/pkg/controller/events/event.go index 8e96b12945..abdad70897 100644 --- a/pkg/controller/events/event.go +++ b/pkg/controller/events/event.go @@ -1,8 +1,8 @@ package events const ( - FailedCreateNode = "FailedCreateNode" - FailedDeleteNode = "FailedDeleteNode" - SuccessfullCreateNode = "SuccessfullCreateNode" - SuccessfullDeleteNode = "SuccessfullDeleteNode" + FailedCreateNode = "FailedCreateNode" + FailedDeleteNode = "FailedDeleteNode" + SuccessfulCreateNode = "SuccessfulCreateNode" + SuccessfulDeleteNode = "SuccessfulDeleteNode" ) diff --git a/pkg/controller/ground.go b/pkg/controller/ground.go index a6b8b94946..e43c3f6f48 100644 --- a/pkg/controller/ground.go +++ b/pkg/controller/ground.go @@ -101,7 +101,7 @@ func (op *GroundControl) Run(threadiness int, stopCh <-chan struct{}, wg *sync.W select { case <-ticker.C: op.Logger.Log( - "msg", "I now would do reconciliation if its was implemented", + "msg", "I now would do reconciliation if it was implemented", "kluster_recheck_interval", KLUSTER_RECHECK_INTERVAL, "v", 2) //op.queue.Add(true) @@ -310,7 +310,7 @@ func (op *GroundControl) klusterUpdate(cur, old interface{}) { func (op *GroundControl) updatePhase(kluster *v1.Kluster, phase models.KlusterPhase, message string) error { - //Never modify the cache, at leasts thats what I've been told + //Never modify the cache, at least that's what I've been told kluster, err := op.Clients.Kubernikus.Kubernikus().Klusters(kluster.Namespace).Get(kluster.Name, metav1.GetOptions{}) if err != nil { return err @@ -539,7 +539,7 @@ func (op *GroundControl) discoverOpenstackInfo(kluster *v1.Kluster) error { "project", kluster.Account()) copy.Spec.Openstack.RouterID = selectedRouter.ID } else { - return fmt.Errorf("Found %d routers in project. Autoconfiguration not possible.", numRouters) + return fmt.Errorf("Found %d routers in project. Auto-configuration not possible.", numRouters) } } diff --git a/pkg/controller/launch/eventing.go b/pkg/controller/launch/eventing.go index 62ebde47f4..9120b705e6 100644 --- a/pkg/controller/launch/eventing.go +++ b/pkg/controller/launch/eventing.go @@ -26,7 +26,7 @@ func (epm *EventingPoolManager) CreateNode() (id string, err error) { id, err = epm.PoolManager.CreateNode() if err == nil { - epm.Recorder.Eventf(epm.Kluster, api_v1.EventTypeNormal, events.SuccessfullCreateNode, "Successfully created node %v", id) + epm.Recorder.Eventf(epm.Kluster, api_v1.EventTypeNormal, events.SuccessfulCreateNode, "Successfully created node %v", id) } else { epm.Recorder.Eventf(epm.Kluster, api_v1.EventTypeWarning, events.FailedCreateNode, "Failed to created node: %v", err) } @@ -38,7 +38,7 @@ func (epm *EventingPoolManager) DeleteNode(id string) (err error) { err = epm.PoolManager.DeleteNode(id) if err == nil { - epm.Recorder.Eventf(epm.Kluster, api_v1.EventTypeNormal, events.SuccessfullDeleteNode, "Successfully deleted node %v", id) + epm.Recorder.Eventf(epm.Kluster, api_v1.EventTypeNormal, events.SuccessfulDeleteNode, "Successfully deleted node %v", id) } else { epm.Recorder.Eventf(epm.Kluster, api_v1.EventTypeWarning, events.FailedDeleteNode, "Failed to delete node: %v", err) } diff --git a/pkg/wormhole/server/tunnel.go b/pkg/wormhole/server/tunnel.go index 2846bb646e..3218ce7cc5 100644 --- a/pkg/wormhole/server/tunnel.go +++ b/pkg/wormhole/server/tunnel.go @@ -29,7 +29,7 @@ func NewTunnel(options *Options) (*Tunnel, error) { } caPool, err := loadCAFile(options.ClientCA) if err != nil { - return nil, fmt.Errorf("Faile to load ca file %s: %s", options.ClientCA, err) + return nil, fmt.Errorf("Failed to load ca file %s: %s", options.ClientCA, err) } tlsConfig.ClientAuth = tls.RequireAndVerifyClientCert tlsConfig.ClientCAs = caPool diff --git a/test/e2e/README.md b/test/e2e/README.md index 051fed239c..e1429fc533 100644 --- a/test/e2e/README.md +++ b/test/e2e/README.md @@ -48,7 +48,7 @@ Triggering a specific phase ca be done as seen below: ``` make tests-e2e ``` -Availabe phases are: +Available phases are: ``` create - create a new cluster api - run API tests