Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

typofix - https://github.com/vlajos/misspell_fixer #12441

Merged
merged 1 commit into from
Aug 10, 2015
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Jump to
Jump to file
Failed to load files.
Diff view
Diff view
4 changes: 2 additions & 2 deletions api/swagger-spec/v1.json
Original file line number Diff line number Diff line change
Expand Up @@ -11429,7 +11429,7 @@
},
"lastTimestamp": {
"type": "string",
"description": "the time at which the most recent occurance of this event was recorded"
"description": "the time at which the most recent occurrence of this event was recorded"
},
"count": {
"type": "integer",
Expand Down Expand Up @@ -11985,7 +11985,7 @@
},
"hostPath": {
"$ref": "v1.HostPathVolumeSource",
"description": "a HostPath provisioned by a developer or tester; for develment use only; see http://releases.k8s.io/HEAD/docs/user-guide/volumes.md#hostpath"
"description": "a HostPath provisioned by a developer or tester; for development use only; see http://releases.k8s.io/HEAD/docs/user-guide/volumes.md#hostpath"
},
"glusterfs": {
"$ref": "v1.GlusterfsVolumeSource",
Expand Down
2 changes: 1 addition & 1 deletion cluster/addons/dns/README.md
Original file line number Diff line number Diff line change
Expand Up @@ -164,7 +164,7 @@ If you see that, DNS is working correctly.

## How does it work?
SkyDNS depends on etcd for what to serve, but it doesn't really need all of
what etcd offers (at least not in the way we use it). For simplicty, we run
what etcd offers (at least not in the way we use it). For simplicity, we run
etcd and SkyDNS together in a pod, and we do not try to link etcd instances
across replicas. A helper container called [kube2sky](kube2sky/) also runs in
the pod and acts a bridge between Kubernetes and SkyDNS. It finds the
Expand Down
2 changes: 1 addition & 1 deletion cluster/aws/config-default.sh
Original file line number Diff line number Diff line change
Expand Up @@ -49,7 +49,7 @@ SERVICE_CLUSTER_IP_RANGE="10.0.0.0/16" # formerly PORTAL_NET
CLUSTER_IP_RANGE="${CLUSTER_IP_RANGE:-10.244.0.0/16}"
MASTER_IP_RANGE="${MASTER_IP_RANGE:-10.246.0.0/24}"
# If set to Elastic IP, master instance will be associated with this IP.
# If set to auto, a new Elastic IP will be aquired
# If set to auto, a new Elastic IP will be acquired
# Otherwise amazon-given public ip will be used (it'll change with reboot).
MASTER_RESERVED_IP="${MASTER_RESERVED_IP:-}"

Expand Down
2 changes: 1 addition & 1 deletion cluster/aws/config-test.sh
Original file line number Diff line number Diff line change
Expand Up @@ -45,7 +45,7 @@ SERVICE_CLUSTER_IP_RANGE="10.0.0.0/16" # formerly PORTAL_NET
CLUSTER_IP_RANGE="${CLUSTER_IP_RANGE:-10.245.0.0/16}"
MASTER_IP_RANGE="${MASTER_IP_RANGE:-10.246.0.0/24}"
# If set to Elastic IP, master instance will be associated with this IP.
# If set to auto, a new Elastic IP will be aquired
# If set to auto, a new Elastic IP will be acquired
# Otherwise amazon-given public ip will be used (it'll change with reboot).
MASTER_RESERVED_IP="${MASTER_RESERVED_IP:-}"

Expand Down
2 changes: 1 addition & 1 deletion cluster/aws/util.sh
Original file line number Diff line number Diff line change
Expand Up @@ -876,7 +876,7 @@ function kube-up {

# This is a race between instance start and volume attachment. There appears to be no way to start an AWS instance with a volume attached.
# To work around this, we wait for volume to be ready in setup-master-pd.sh
echo "Attaching peristent data volume (${MASTER_DISK_ID}) to master"
echo "Attaching persistent data volume (${MASTER_DISK_ID}) to master"
$AWS_CMD attach-volume --volume-id ${MASTER_DISK_ID} --device /dev/sdb --instance-id ${master_id}

sleep 10
Expand Down
2 changes: 1 addition & 1 deletion cluster/common.sh
Original file line number Diff line number Diff line change
Expand Up @@ -172,7 +172,7 @@ function get-kubeconfig-bearertoken() {
}

# Sets KUBE_VERSION variable to the version passed in as an argument, or if argument is
# latest_stable, latest_release, or latest_ci fetches and sets the correponding version number
# latest_stable, latest_release, or latest_ci fetches and sets the corresponding version number
#
# Args:
# $1 version string from command line
Expand Down
2 changes: 1 addition & 1 deletion cluster/gce/util.sh
Original file line number Diff line number Diff line change
Expand Up @@ -425,7 +425,7 @@ function create-node-template {
}

# Robustly try to add metadata on an instance.
# $1: The name of the instace.
# $1: The name of the instance.
# $2...$n: The metadata key=value pairs to add.
function add-instance-metadata {
local -r instance=$1
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -39,7 +39,7 @@ def test_update_rc_files(self, pmock):
pmock.return_value.write_lines.assert_called_with(['line1', 'line2',
'test1', 'test2'])

def test_update_rc_files_with_nonexistant_path(self):
def test_update_rc_files_with_nonexistent_path(self):
"""
Test an unhappy path if the bashrc/users do not exist.
"""
Expand Down
2 changes: 1 addition & 1 deletion cluster/juju/charms/trusty/kubernetes/hooks/start
Original file line number Diff line number Diff line change
Expand Up @@ -3,7 +3,7 @@
set -ex

# Start is guaranteed to be called once when after the unit is installed
# *AND* once everytime a machine is rebooted.
# *AND* once every time a machine is rebooted.

if [ ! -f $CHARM_DIR/.unit-state ]
then
Expand Down
2 changes: 1 addition & 1 deletion cluster/juju/return-node-ips.py
Original file line number Diff line number Diff line change
Expand Up @@ -16,7 +16,7 @@

import json
import sys
# This script helps parse out the private IP addreses from the
# This script helps parse out the private IP addresses from the
# `juju run` command's JSON object, see cluster/juju/util.sh

if len(sys.argv) > 1:
Expand Down
2 changes: 1 addition & 1 deletion cluster/juju/util.sh
Original file line number Diff line number Diff line change
Expand Up @@ -135,7 +135,7 @@ function sleep-status() {

# sleep because we cannot get the status back of where the minions are in the deploy phase
# thanks to a generic "started" state and our service not actually coming online until the
# minions have recieved the binary from the master distribution hub during relations
# minions have received the binary from the master distribution hub during relations
echo "Sleeping an additional minute to allow the cluster to settle"
sleep 60
}
2 changes: 1 addition & 1 deletion cluster/libvirt-coreos/util.sh
Original file line number Diff line number Diff line change
Expand Up @@ -296,7 +296,7 @@ function kube-push-internal {
local)
kube-push-local;;
*)
echo "The only known push methods are \"release\" to use the relase tarball or \"local\" to use the binaries built by make. KUBE_PUSH is set \"$KUBE_PUSH\"" >&2
echo "The only known push methods are \"release\" to use the release tarball or \"local\" to use the binaries built by make. KUBE_PUSH is set \"$KUBE_PUSH\"" >&2
return 1;;
esac
}
Expand Down
2 changes: 1 addition & 1 deletion cluster/mesos/docker/util.sh
Original file line number Diff line number Diff line change
Expand Up @@ -218,7 +218,7 @@ function validate-cluster {
echo "Validating ${KUBERNETES_PROVIDER} cluster" 1>&2

# Do not validate cluster size. There will be zero k8s minions until a pod is created.
# TODO(karlkfi): use componentstatuses or equivelent when it supports non-localhost core components
# TODO(karlkfi): use componentstatuses or equivalent when it supports non-localhost core components

# Validate immediate cluster reachability and responsiveness
echo "KubeDNS: $(cluster::mesos::docker::addon_status 'kube-dns')"
Expand Down
2 changes: 1 addition & 1 deletion cluster/saltbase/salt/kube-addons/kube-addon-update.sh
Original file line number Diff line number Diff line change
Expand Up @@ -25,7 +25,7 @@
# LIMITATIONS
# 1. controllers are not updated unless their name is changed
# 3. Services will not be updated unless their name is changed,
# but for services we acually want updates without name change.
# but for services we actually want updates without name change.
# 4. Json files are not handled at all. Currently addons must be
# in yaml files
# 5. exit code is probably not always correct (I haven't checked
Expand Down
2 changes: 1 addition & 1 deletion cluster/saltbase/salt/kube-master-addons/init.sls
Original file line number Diff line number Diff line change
Expand Up @@ -6,7 +6,7 @@
- mode: 755

# Used to restart kube-master-addons service each time salt is run
# Actually, it doens't work (the service is not restarted),
# Actually, it does not work (the service is not restarted),
# but master-addon service always terminates after it does it job,
# so it is (usually) not running and it will be started when
# salt is run.
Expand Down
2 changes: 1 addition & 1 deletion cluster/ubuntu/util.sh
Original file line number Diff line number Diff line change
Expand Up @@ -35,7 +35,7 @@ function test-build-release {
"${KUBE_ROOT}/build/release.sh"
}

# From user input set the necessary k8s and etcd configuration infomation
# From user input set the necessary k8s and etcd configuration information
function setClusterInfo() {
# Initialize MINION_IPS in setClusterInfo function
# MINION_IPS is defined as a global variable, and is concatenated with other nodeIP
Expand Down
4 changes: 2 additions & 2 deletions cmd/kube-apiserver/app/server.go
Original file line number Diff line number Diff line change
Expand Up @@ -479,7 +479,7 @@ func (s *APIServer) Run(_ []string) error {
}
// err == systemd.SdNotifyNoSocket when not running on a systemd system
if err := systemd.SdNotify("READY=1\n"); err != nil && err != systemd.SdNotifyNoSocket {
glog.Errorf("Unable to send systemd daemon sucessful start message: %v\n", err)
glog.Errorf("Unable to send systemd daemon successful start message: %v\n", err)
}
if err := secureServer.ListenAndServeTLS(s.TLSCertFile, s.TLSPrivateKeyFile); err != nil {
glog.Errorf("Unable to listen for secure (%v); will try again.", err)
Expand All @@ -497,7 +497,7 @@ func (s *APIServer) Run(_ []string) error {
if secureLocation == "" {
// err == systemd.SdNotifyNoSocket when not running on a systemd system
if err := systemd.SdNotify("READY=1\n"); err != nil && err != systemd.SdNotifyNoSocket {
glog.Errorf("Unable to send systemd daemon sucessful start message: %v\n", err)
glog.Errorf("Unable to send systemd daemon successful start message: %v\n", err)
}
}
glog.Infof("Serving insecurely on %s", insecureLocation)
Expand Down
2 changes: 1 addition & 1 deletion cmd/kube-controller-manager/app/controllermanager.go
Original file line number Diff line number Diff line change
Expand Up @@ -125,7 +125,7 @@ func (s *CMServer) AddFlags(fs *pflag.FlagSet) {
fs.IntVar(&s.DeletingPodsBurst, "deleting-pods-burst", 10, "Number of nodes on which pods are bursty deleted in case of node failure. For more details look into RateLimiter.")
fs.IntVar(&s.RegisterRetryCount, "register-retry-count", s.RegisterRetryCount, ""+
"The number of retries for initial node registration. Retry interval equals node-sync-period.")
fs.MarkDeprecated("register-retry-count", "This flag is currenty no-op and will be deleted.")
fs.MarkDeprecated("register-retry-count", "This flag is currently no-op and will be deleted.")
fs.DurationVar(&s.NodeMonitorGracePeriod, "node-monitor-grace-period", 40*time.Second,
"Amount of time which we allow running Node to be unresponsive before marking it unhealty. "+
"Must be N times more than kubelet's nodeStatusUpdateFrequency, "+
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -25,7 +25,7 @@
# LIMITATIONS
# 1. controllers are not updated unless their name is changed
# 3. Services will not be updated unless their name is changed,
# but for services we acually want updates without name change.
# but for services we actually want updates without name change.
# 4. Json files are not handled at all. Currently addons must be
# in yaml files
# 5. exit code is probably not always correct (I haven't checked
Expand Down
2 changes: 1 addition & 1 deletion contrib/ansible/roles/kubernetes/defaults/main.yml
Original file line number Diff line number Diff line change
Expand Up @@ -25,7 +25,7 @@ kube_token_dir: "{{ kube_config_dir }}/tokens"
kube_manifest_dir: "{{ kube_config_dir }}/manifests"

# This is the group that the cert creation scripts chgrp the
# cert files to. Not really changable...
# cert files to. Not really changeable...
kube_cert_group: kube-cert

# Internal DNS domain name.
Expand Down
2 changes: 1 addition & 1 deletion contrib/diurnal/time.go
Original file line number Diff line number Diff line change
Expand Up @@ -182,7 +182,7 @@ func parseTimeISO8601(s string) (time.Time, error) {
theTime.offMinute = v
s = s[2:]
default:
return time.Time{}, errors.New("an unknown error occured")
return time.Time{}, errors.New("an unknown error occurred")
}
state++
}
Expand Down
2 changes: 1 addition & 1 deletion contrib/for-demos/proxy-to-service/Dockerfile
Original file line number Diff line number Diff line change
@@ -1,4 +1,4 @@
# Alpine linux would be great for this, but it's DNS does not use seach paths.
# Alpine linux would be great for this, but it's DNS does not use search paths.
FROM progrium/busybox
MAINTAINER Tim Hockin "thockin@google.com"

Expand Down
2 changes: 1 addition & 1 deletion contrib/mesos/docs/ha.md
Original file line number Diff line number Diff line change
Expand Up @@ -34,7 +34,7 @@ In this case, if there are problems launching a replacement scheduler process th
##### Command Line Arguments

- `--ha` is required to enable scheduler HA and multi-scheduler leader election.
- `--km_path` or else (`--executor_path` and `--proxy_path`) should reference non-local-file URI's and must be identicial across schedulers.
- `--km_path` or else (`--executor_path` and `--proxy_path`) should reference non-local-file URI's and must be identical across schedulers.

If you have HDFS installed on your slaves then you can specify HDFS URI locations for the binaries:

Expand Down
4 changes: 2 additions & 2 deletions contrib/mesos/pkg/executor/executor_test.go
Original file line number Diff line number Diff line change
Expand Up @@ -263,7 +263,7 @@ func TestExecutorRegister(t *testing.T) {
}

// TestExecutorDisconnect ensures that the executor thinks that it is not
// connected after a call to Disconnected has occured.
// connected after a call to Disconnected has occurred.
func TestExecutorDisconnect(t *testing.T) {
mockDriver := &MockExecutorDriver{}
executor := NewTestKubernetesExecutor()
Expand Down Expand Up @@ -346,7 +346,7 @@ func TestExecutorLaunchAndKillTask(t *testing.T) {
select {
case <-updates:
case <-time.After(time.Second):
t.Fatalf("Executor should send an intial update on Registration")
t.Fatalf("Executor should send an initial update on Registration")
}

pod := NewTestPod(1)
Expand Down
2 changes: 1 addition & 1 deletion contrib/mesos/pkg/queue/interface.go
Original file line number Diff line number Diff line change
Expand Up @@ -58,7 +58,7 @@ type FIFO interface {
// Pop waits until an item is ready and returns it. If multiple items are
// ready, they are returned in the order in which they were added/updated.
// The item is removed from the queue (and the store) before it is returned,
// so if you don't succesfully process it, you need to add it back with Add().
// so if you don't successfully process it, you need to add it back with Add().
Pop() interface{}

// Await attempts to Pop within the given interval; upon success the non-nil
Expand Down
2 changes: 1 addition & 1 deletion contrib/mesos/pkg/scheduler/plugin.go
Original file line number Diff line number Diff line change
Expand Up @@ -350,7 +350,7 @@ func (k *kubeScheduler) doSchedule(task *podtask.T, err error) (string, error) {
return "", fmt.Errorf("task.offer assignment must be idempotent, task %+v: offer %+v", task, offer)
}

// write resource limits into the pod spec which is transfered to the executor. From here
// write resource limits into the pod spec which is transferred to the executor. From here
// on we can expect that the pod spec of a task has proper limits for CPU and memory.
// TODO(sttts): For a later separation of the kubelet and the executor also patch the pod on the apiserver
if unlimitedCPU := mresource.LimitPodCPU(&task.Pod, k.defaultContainerCPULimit); unlimitedCPU {
Expand Down
2 changes: 1 addition & 1 deletion contrib/prometheus/README.md
Original file line number Diff line number Diff line change
Expand Up @@ -52,7 +52,7 @@ This is a v1 api based, containerized prometheus ReplicationController, which sc

1. Use kubectl to handle auth & proxy the kubernetes API locally, emulating the old KUBERNETES_RO service.

1. The list of services to be monitored is passed as a command line aguments in
1. The list of services to be monitored is passed as a command line arguments in
the yaml file.

1. The startup scripts assumes that each service T will have
Expand Down
2 changes: 1 addition & 1 deletion contrib/service-loadbalancer/README.md
Original file line number Diff line number Diff line change
Expand Up @@ -191,7 +191,7 @@ $ mysql -u root -ppassword --host 104.197.63.17 --port 3306 -e 'show databases;'
### Troubleshooting:
- If you can curl or netcat the endpoint from the pod (with kubectl exec) and not from the node, you have not specified hostport and containerport.
- If you can hit the ips from the node but not from your machine outside the cluster, you have not opened firewall rules for the right network.
- If you can't hit the ips from within the container, either haproxy or the service_loadbalacer script is not runing.
- If you can't hit the ips from within the container, either haproxy or the service_loadbalacer script is not running.
1. Use ps in the pod
2. sudo restart haproxy in the pod
3. cat /etc/haproxy/haproxy.cfg in the pod
Expand Down
2 changes: 1 addition & 1 deletion contrib/service-loadbalancer/service_loadbalancer_test.go
Original file line number Diff line number Diff line change
Expand Up @@ -141,7 +141,7 @@ func TestGetServices(t *testing.T) {
{Port: 20, TargetPort: util.NewIntOrStringFromInt(ports[1])},
}

// 2 services targetting the same endpoints, one of which is declared as a tcp service.
// 2 services targeting the same endpoints, one of which is declared as a tcp service.
svc1 := getService(servicePorts)
svc2 := getService(servicePorts)
endpoints := []*api.Endpoints{
Expand Down
2 changes: 1 addition & 1 deletion docs/admin/cluster-large.md
Original file line number Diff line number Diff line change
Expand Up @@ -88,7 +88,7 @@ To avoid running into cluster addon resource issues, when creating a cluster wit
* [Kibana](http://releases.k8s.io/HEAD/cluster/addons/fluentd-elasticsearch/kibana-controller.yaml)
* Scale number of replicas for the following addons, if used, along with the size of cluster (there are multiple replicas of each so increasing replicas should help handle increased load, but, since load per replica also increases slightly, also consider increasing CPU/memory limits):
* [elasticsearch](http://releases.k8s.io/HEAD/cluster/addons/fluentd-elasticsearch/es-controller.yaml)
* Increase memory and CPU limits sligthly for each of the following addons, if used, along with the size of cluster (there is one replica per node but CPU/memory usage increases slightly along with cluster load/size as well):
* Increase memory and CPU limits slightly for each of the following addons, if used, along with the size of cluster (there is one replica per node but CPU/memory usage increases slightly along with cluster load/size as well):
* [FluentD with ElasticSearch Plugin](http://releases.k8s.io/HEAD/cluster/saltbase/salt/fluentd-es/fluentd-es.yaml)
* [FluentD with GCP Plugin](http://releases.k8s.io/HEAD/cluster/saltbase/salt/fluentd-gcp/fluentd-gcp.yaml)

Expand Down
4 changes: 2 additions & 2 deletions docs/admin/cluster-management.md
Original file line number Diff line number Diff line change
Expand Up @@ -35,7 +35,7 @@ Documentation for other releases can be found at

This document describes several topics related to the lifecycle of a cluster: creating a new cluster,
upgrading your cluster's
master and worker nodes, performing node maintainence (e.g. kernel upgrades), and upgrading the Kubernetes API version of a
master and worker nodes, performing node maintenance (e.g. kernel upgrades), and upgrading the Kubernetes API version of a
running cluster.

## Creating and configuring a Cluster
Expand Down Expand Up @@ -132,7 +132,7 @@ For pods with a replication controller, the pod will eventually be replaced by a

For pods with no replication controller, you need to bring up a new copy of the pod, and assuming it is not part of a service, redirect clients to it.

Perform maintainence work on the node.
Perform maintenance work on the node.

Make the node schedulable again:

Expand Down
2 changes: 1 addition & 1 deletion docs/admin/etcd.md
Original file line number Diff line number Diff line change
Expand Up @@ -41,7 +41,7 @@ objects.

Access Control: give *only* kube-apiserver read/write access to etcd. You do not
want apiserver's etcd exposed to every node in your cluster (or worse, to the
internet at large), because access to etcd is equivilent to root in your
internet at large), because access to etcd is equivalent to root in your
cluster.

Data Reliability: for reasonable safety, either etcd needs to be run as a
Expand Down
4 changes: 2 additions & 2 deletions docs/api-reference/definitions.html
Original file line number Diff line number Diff line change
Expand Up @@ -4224,7 +4224,7 @@ <h3 id="_v1_persistentvolumespec">v1.PersistentVolumeSpec</h3>
</tr>
<tr>
<td class="tableblock halign-left valign-top"><p class="tableblock">hostPath</p></td>
<td class="tableblock halign-left valign-top"><p class="tableblock">a HostPath provisioned by a developer or tester; for develment use only; see <a href="http://releases.k8s.io/HEAD/docs/user-guide/volumes.md#hostpath">http://releases.k8s.io/HEAD/docs/user-guide/volumes.md#hostpath</a></p></td>
<td class="tableblock halign-left valign-top"><p class="tableblock">a HostPath provisioned by a developer or tester; for development use only; see <a href="http://releases.k8s.io/HEAD/docs/user-guide/volumes.md#hostpath">http://releases.k8s.io/HEAD/docs/user-guide/volumes.md#hostpath</a></p></td>
<td class="tableblock halign-left valign-top"><p class="tableblock">false</p></td>
<td class="tableblock halign-left valign-top"><p class="tableblock"><a href="#_v1_hostpathvolumesource">v1.HostPathVolumeSource</a></p></td>
<td class="tableblock halign-left valign-top"></td>
Expand Down Expand Up @@ -5099,7 +5099,7 @@ <h3 id="_v1_event">v1.Event</h3>
</tr>
<tr>
<td class="tableblock halign-left valign-top"><p class="tableblock">lastTimestamp</p></td>
<td class="tableblock halign-left valign-top"><p class="tableblock">the time at which the most recent occurance of this event was recorded</p></td>
<td class="tableblock halign-left valign-top"><p class="tableblock">the time at which the most recent occurrence of this event was recorded</p></td>
<td class="tableblock halign-left valign-top"><p class="tableblock">false</p></td>
<td class="tableblock halign-left valign-top"><p class="tableblock">string</p></td>
<td class="tableblock halign-left valign-top"></td>
Expand Down
2 changes: 1 addition & 1 deletion docs/devel/development.md
Original file line number Diff line number Diff line change
Expand Up @@ -87,7 +87,7 @@ Note: If you have write access to the main repository at github.com/GoogleCloudP
git remote set-url --push upstream no_push
```

### Commiting changes to your fork
### Committing changes to your fork

```sh
git commit
Expand Down
2 changes: 1 addition & 1 deletion docs/devel/making-release-notes.md
Original file line number Diff line number Diff line change
Expand Up @@ -40,7 +40,7 @@ This documents the process for making release notes for a release.
Find the most-recent PR that was merged with the previous .0 release. Remember this as $LASTPR.
_TODO_: Figure out a way to record this somewhere to save the next release engineer time.

Find the most-recent PR that was merged with the current .0 release. Remeber this as $CURRENTPR.
Find the most-recent PR that was merged with the current .0 release. Remember this as $CURRENTPR.

### 2) Run the release-notes tool

Expand Down
Original file line number Diff line number Diff line change
@@ -1,4 +1,4 @@
## This file is used as input to deployment script, which ammends it as needed.
## This file is used as input to deployment script, which amends it as needed.
## More specifically, we need to add peer hosts for each but the elected peer.

coreos:
Expand Down
Original file line number Diff line number Diff line change
@@ -1,4 +1,4 @@
## This file is used as input to deployment script, which ammends it as needed.
## This file is used as input to deployment script, which amends it as needed.
## More specifically, we need to add environment files for as many nodes as we
## are going to deploy.

Expand Down