diff --git a/website/source/docs/agent/basics.html.markdown b/website/source/docs/agent/basics.html.markdown
index 5740f3b84f..f97ca3fbde 100644
--- a/website/source/docs/agent/basics.html.markdown
+++ b/website/source/docs/agent/basics.html.markdown
@@ -72,7 +72,7 @@ There are several important components that `consul agent` outputs:
agent. This is also the address other applications can use over [RPC to control Consul](/docs/agent/rpc.html).
* **Cluster Addr**: This is the address and ports used for communication between
- Consul agents in a cluster. Every Consul agent in a cluster does not have to
+ Consul agents in a cluster. Not all Consul agents in a cluster have to
use the same port, but this address **MUST** be reachable by all other nodes.
## Stopping an Agent
@@ -105,7 +105,7 @@ this lifecycle is useful to building a mental model of an agent's interactions
with a cluster, and how the cluster treats a node.
When an agent is first started, it does not know about any other node in the cluster.
-To discover it's peers, it must _join_ the cluster. This is done with the `join`
+To discover its peers, it must _join_ the cluster. This is done with the `join`
command or by providing the proper configuration to auto-join on start. Once a node
joins, this information is gossiped to the entire cluster, meaning all nodes will
eventually be aware of each other. If the agent is a server, existing servers will
@@ -115,18 +115,18 @@ In the case of a network failure, some nodes may be unreachable by other nodes.
In this case, unreachable nodes are marked as _failed_. It is impossible to distinguish
between a network failure and an agent crash, so both cases are handled the same.
Once a node is marked as failed, this information is updated in the service catalog.
-There is some nuance here relating, since this update is only possible if the
-servers can still [form a quorum](/docs/internals/consensus.html). Once the network
-failure recovers, or a crashed agent restarts, the cluster will repair itself,
-and unmark a node as failed. The health check in the catalog will also be updated
-to reflect this.
+There is some nuance here since this update is only possible if the servers can
+still [form a quorum](/docs/internals/consensus.html). Once the network recovers,
+or a crashed agent restarts, the cluster will repair itself, and unmark
+a node as failed. The health check in the catalog will also be updated to reflect
+this.
-When a node _leaves_, it specifies it's intent to do so, and so the cluster
+When a node _leaves_, it specifies its intent to do so, and so the cluster
marks that node as having _left_. Unlike the _failed_ case, all of the
services provided by a node are immediately deregistered. If the agent was
a server, replication to it will stop. To prevent an accumulation
of dead nodes, Consul will automatically reap _failed_ nodes out of the
-catalog as well. This is currently done on a non-configurable interval
-which defaults to 72 hours. Reaping is similar to leaving, causing all
-associated services to be deregistered.
+catalog as well. This is currently done on a non-configurable interval of
+72 hours. Reaping is similar to leaving, causing all associated services
+to be deregistered.
diff --git a/website/source/docs/agent/checks.html.markdown b/website/source/docs/agent/checks.html.markdown
index a5a4bdb566..8a31a04620 100644
--- a/website/source/docs/agent/checks.html.markdown
+++ b/website/source/docs/agent/checks.html.markdown
@@ -16,12 +16,12 @@ or added at runtime over the HTTP interface.
There are two different kinds of checks:
* Script + Interval - These checks depend on invoking an external application
- which does the health check and exits with an appropriate exit code, potentially
+ that does the health check and exits with an appropriate exit code, potentially
generating some output. A script is paired with an invocation interval (e.g.
every 30 seconds). This is similar to the Nagios plugin system.
- * TTL - These checks retain their last known state for a given TTL. The state
- of the check must be updated periodically over the HTTP interface. If an
+ * Time to Live (TTL) - These checks retain their last known state for a given TTL.
+ The state of the check must be updated periodically over the HTTP interface. If an
external system fails to update the status within a given TTL, the check is
set to the failed state. This mechanism is used to allow an application to
directly report its health. For example, a web app can periodically curl the
@@ -75,7 +75,7 @@ check can be registered dynamically using the [HTTP API](/docs/agent/http.html).
## Check Scripts
A check script is generally free to do anything to determine the status
-of the check. The only limitations placed are the exit codes must convey
+of the check. The only limitations placed are that the exit codes must convey
a specific meaning. Specifically:
* Exit code 0 - Check is passing
diff --git a/website/source/docs/agent/dns.html.markdown b/website/source/docs/agent/dns.html.markdown
index ab6995a35d..d61603253a 100644
--- a/website/source/docs/agent/dns.html.markdown
+++ b/website/source/docs/agent/dns.html.markdown
@@ -8,22 +8,23 @@ description: |-
# DNS Interface
-One of the primary query interfaces for Consul is using DNS.
+One of the primary query interfaces for Consul is DNS.
The DNS interface allows applications to make use of service
discovery without any high-touch integration with Consul. For
-example, instead of making any HTTP API requests to Consul,
+example, instead of making HTTP API requests to Consul,
a host can use the DNS server directly and just do a name lookup
for "redis.service.east-aws.consul".
This query automatically translates to a lookup of nodes that
-provide the redis service, located in the "east-aws" datacenter,
-with no failing health checks. It's that simple!
+provide the redis service, are located in the "east-aws" datacenter,
+and have no failing health checks. It's that simple!
There are a number of [configuration options](/docs/agent/options.html) that
are important for the DNS interface. They are `client_addr`, `ports.dns`, `recursors`,
`domain`, and `dns_config`. By default Consul will listen on 127.0.0.1:8600 for DNS queries
-in the "consul." domain, without support for DNS recursion. All queries are case-insensitive, a
-name lookup for `PostgreSQL.node.dc1.consul` will find all nodes named `postgresql`, no matter of case.
+in the "consul." domain, without support for DNS recursion. All queries are case-insensitive: a
+name lookup for `PostgreSQL.node.dc1.consul` will find all nodes named `postgresql`,
+regardless of case.
There are a few ways to use the DNS interface. One option is to use a custom
DNS resolver library and point it at Consul. Another option is to set Consul
@@ -76,24 +77,24 @@ consul. 0 IN SOA ns.consul. postmaster.consul. 1392836399 3600 600 86400 0
## Service Lookups
A service lookup is the alternate type of query. It is used to query for service
-providers and supports two mode of lookup, a strict RCF style lookup and the
-standard lookup.
+providers and supports two lookup methods: standard lookup, and strict
+[RFC 2782](https://tools.ietf.org/html/rfc2782) lookup.
-### Standard Style Lookup
+### Standard Lookup
-The format of a standard service lookup is like the following:
+The format of a standard service lookup is:
[tag.].service[.datacenter][.domain]
As with node lookups, the `datacenter` is optional, as is the `tag`. If no tag is
provided, then no filtering is done on tag. So, if we want to find any redis service
-providers in our local datacenter, we could lookup "redis.service.consul.", however
+providers in our local datacenter, we could lookup "redis.service.consul.", while
if we care about the PostgreSQL master in a particular datacenter, we could lookup
"master.postgresql.service.dc2.consul."
The DNS query system makes use of health check information to prevent routing
to unhealthy nodes. When a service query is made, any services failing their health
-check, or failing a node system check will be omitted from the results. To allow
+check, or failing a node system check, will be omitted from the results. To allow
for simple load balancing, the set of nodes returned is also randomized each time.
These simple mechanisms make it easy to use DNS along with application level retries
as a simple foundation for an auto-healing service oriented architecture.
@@ -124,13 +125,13 @@ consul.service.consul. 0 IN SRV 1 1 8300 foobar.node.dc1.consul.
foobar.node.dc1.consul. 0 IN A 10.1.10.12
```
-### RFC-2782 Style Lookup
+### RFC 2782 Lookup
-The format for RFC style lookups uses the following format:
+The format for RFC 2782 SRV lookups is:
_._.service[.datacenter][.domain]
-Per [RFC-2782](https://www.ietf.org/rfc/rfc2782.txt), SRV queries should use
+Per [RFC 2782](https://tools.ietf.org/html/rfc2782), SRV queries should use
underscores (_) as a prefix to the `service` and `protocol` values in a query to
prevent DNS collisions. The `protocol` value can be any of the tags for a
service or if the service has no tags, the value "tcp" should be used. If "tcp"
@@ -139,7 +140,7 @@ is specified as the protocol, the query will not perform any tag filtering.
Other than the query format and default "tcp" protocol/tag value, the behavior
of the RFC style lookup is the same as the standard style of lookup.
-Using the RCF style lookup, If you registered the service "rabbitmq" on port
+Using RFC 2782 lookup, If you registered the service "rabbitmq" on port
5672 and tagged it with "amqp" you would query the SRV record as
"_rabbitmq._amqp.service.consul" as illustrated in the example below:
@@ -168,7 +169,7 @@ rabbitmq.node1.dc1.consul. 0 IN A 10.1.11.20
When the DNS query is performed using UDP, Consul will truncate the results
without setting the truncate bit. This is to prevent a redundant lookup over
-TCP which generate additional load. If the lookup is done over TCP, the results
+TCP that generates additional load. If the lookup is done over TCP, the results
are not truncated.
## Caching
diff --git a/website/source/docs/agent/watches.html.markdown b/website/source/docs/agent/watches.html.markdown
index 5ab4707217..261eef08da 100644
--- a/website/source/docs/agent/watches.html.markdown
+++ b/website/source/docs/agent/watches.html.markdown
@@ -15,7 +15,7 @@ executable. As an example, you could watch the status of health checks and
notify an external system when a check is critical.
Watches are implemented using blocking queries in the [HTTP API](/docs/agent/http.html).
-Agent's automatically make the proper API calls to watch for changes,
+Agents automatically make the proper API calls to watch for changes,
and inform a handler when the data view has updated.
Watches can be configured as part of the [agent's configuration](/docs/agent/options.html),
@@ -36,7 +36,7 @@ The watch specification specifies the view of data to be monitored.
Once that view is updated the specified handler is invoked. The handler
can be any executable.
-A handler should read it's input from stdin, and expect to read
+A handler should read its input from stdin, and expect to read
JSON formatted data. The format of the data depends on the type of the
watch. Each watch type documents the format type, and because they
map directly to an HTTP API, handlers should expect the input to
@@ -280,7 +280,7 @@ An example of the output of this command:
### Type: checks
The "checks" watch type is used to monitor the checks of a given
-service or in a specific state. It optionally takes the "service"
+service or those in a specific state. It optionally takes the "service"
parameter to filter to a specific service, or "state" to filter
to a specific state. By default, it will watch all checks.
diff --git a/website/source/docs/compatibility.html.markdown b/website/source/docs/compatibility.html.markdown
index bc6aca1396..1d5516568d 100644
--- a/website/source/docs/compatibility.html.markdown
+++ b/website/source/docs/compatibility.html.markdown
@@ -10,8 +10,8 @@ description: |-
We expect Consul to run in large clusters as long-running agents. Because
upgrading agents in this sort of environment relies heavily on protocol
-compatibility, this page makes it clear on our promise to keeping different
-Consul versions compatible with each other.
+compatibility, this page makes clear our promise to keep different Consul
+versions compatible with each other.
We promise that every subsequent release of Consul will remain backwards
compatible with _at least_ one prior version. Concretely: version 0.5 can
diff --git a/website/source/docs/internals/acl.html.markdown b/website/source/docs/internals/acl.html.markdown
index 08522a7077..e2547513e8 100644
--- a/website/source/docs/internals/acl.html.markdown
+++ b/website/source/docs/internals/acl.html.markdown
@@ -11,7 +11,7 @@ description: |-
Consul provides an optional Access Control List (ACL) system which can be used to control
access to data and APIs. The ACL system is a
[Capability-based system](http://en.wikipedia.org/wiki/Capability-based_security) that relies
-on tokens which can have fine grained rules applied to them. It is very similar to
+on tokens to which fine grained rules can be applied. It is very similar to
[AWS IAM](http://aws.amazon.com/iam/) in many ways.
## ACL Design
@@ -30,10 +30,10 @@ perform all actions.
The token ID is passed along with each RPC request to the servers. Agents
[can be configured](/docs/agent/options.html) with `acl_token` to provide a default token,
but the token can also be specified by a client on a [per-request basis](/docs/agent/http.html).
-ACLs are new as of Consul 0.4, meaning versions prior do not provide a token.
+ACLs are new as of Consul 0.4, meaning prior versions do not provide a token.
This is handled by the special "anonymous" token. Anytime there is no token provided,
-the rules defined by that token are automatically applied. This lets policy be enforced
-on legacy clients.
+the rules defined by that token are automatically applied. This allows
+policy to be enforced on legacy clients.
Enforcement is always done by the server nodes. All servers must be [configured
to provide](/docs/agent/options.html) an `acl_datacenter`, which enables
@@ -47,7 +47,7 @@ all the tokens.
When a request is made to any non-authoritative server with a token, it must
be resolved into the appropriate policy. This is done by reading the token
from the authoritative server and caching a configurable `acl_ttl`. The implication
-of caching is that the cache TTL is an upper-bound on the staleness of policy
+of caching is that the cache TTL is an upper bound on the staleness of policy
that is enforced. It is possible to set a zero TTL, but this has adverse
performance impacts, as every request requires refreshing the policy.
diff --git a/website/source/docs/internals/architecture.html.markdown b/website/source/docs/internals/architecture.html.markdown
index ff0cb46695..068bfc0f50 100644
--- a/website/source/docs/internals/architecture.html.markdown
+++ b/website/source/docs/internals/architecture.html.markdown
@@ -23,20 +23,20 @@ Before describing the architecture, we provide a glossary of terms to help
clarify what is being discussed:
* Agent - An agent is the long running daemon on every member of the Consul cluster.
-It is started by running `consul agent`. The agent is able to run in either *client*,
+It is started by running `consul agent`. The agent is able to run in either *client*
or *server* mode. Since all nodes must be running an agent, it is simpler to refer to
-the node as either being a client or server, but there are other instances of the agent. All
+the node as being either a client or server, but there are other instances of the agent. All
agents can run the DNS or HTTP interfaces, and are responsible for running checks and
keeping services in sync.
* Client - A client is an agent that forwards all RPCs to a server. The client is relatively
-stateless. The only background activity a client performs is taking part of LAN gossip pool.
-This has a minimal resource overhead and consumes only a small amount of network bandwidth.
+stateless. The only background activity a client performs is taking part in the LAN gossip
+pool. This has a minimal resource overhead and consumes only a small amount of network
+bandwidth.
-* Server - An agent that is server mode. When in server mode, there is an expanded set
-of responsibilities including participating in the Raft quorum, maintaining cluster state,
-responding to RPC queries, WAN gossip to other datacenters, and forwarding queries to leaders
-or remote datacenters.
+* Server - A server is an agent with an expanded set of responsibilities including
+participating in the Raft quorum, maintaining cluster state, responding to RPC queries,
+WAN gossip to other datacenters, and forwarding queries to leaders or remote datacenters.
* Datacenter - A datacenter seems obvious, but there are subtle details such as multiple
availability zones in EC2. We define a datacenter to be a networking environment that is
@@ -47,13 +47,13 @@ the public internet.
the elected leader as well as agreement on the ordering of transactions. Since these
transactions are applied to a FSM, we implicitly include the consistency of a replicated
state machine. Consensus is described in more detail on [Wikipedia](http://en.wikipedia.org/wiki/Consensus_(computer_science)),
-as well as our [implementation here](/docs/internals/consensus.html).
+and our implementation is described [here](/docs/internals/consensus.html).
* Gossip - Consul is built on top of [Serf](http://www.serfdom.io/), which provides a full
[gossip protocol](http://en.wikipedia.org/wiki/Gossip_protocol) that is used for multiple purposes.
Serf provides membership, failure detection, and event broadcast mechanisms. Our use of these
is described more in the [gossip documentation](/docs/internals/gossip.html). It is enough to know
-gossip involves random node-to-node communication, primarily over UDP.
+that gossip involves random node-to-node communication, primarily over UDP.
* LAN Gossip - Refers to the LAN gossip pool, which contains nodes that are all
located on the same local area network or datacenter.
@@ -62,8 +62,8 @@ located on the same local area network or datacenter.
servers are primarily located in different datacenters and typically communicate
over the internet or wide area network.
-* RPC - RPC is short for a Remote Procedure Call. This is a request / response mechanism
-allowing a client to make a request from a server.
+* RPC - Remote Procedure Call. This is a request / response mechanism allowing a
+client to make a request of a server.
## 10,000 foot view
@@ -73,7 +73,7 @@ From a 10,000 foot altitude the architecture of Consul looks like this:
![Consul Architecture](consul-arch.png)
-Lets break down this image and describe each piece. First of all we can see
+Let's break down this image and describe each piece. First of all we can see
that there are two datacenters, one and two respectively. Consul has first
class support for multiple datacenters and expects this to be the common case.
@@ -85,9 +85,9 @@ and they can easily scale into the thousands or tens of thousands.
All the nodes that are in a datacenter participate in a [gossip protocol](/docs/internals/gossip.html).
This means there is a gossip pool that contains all the nodes for a given datacenter. This serves
-a few purposes: first, there is no need to configure clients with the addresses of servers,
+a few purposes: first, there is no need to configure clients with the addresses of servers;
discovery is done automatically. Second, the work of detecting node failures
-is not placed on the servers but is distributed. This makes the failure detection much more
+is not placed on the servers but is distributed. This makes failure detection much more
scalable than naive heartbeating schemes. Thirdly, it is used as a messaging layer to notify
when important events such as leader election take place.
@@ -97,8 +97,8 @@ processing all queries and transactions. Transactions must also be replicated to
as part of the [consensus protocol](/docs/internals/consensus.html). Because of this requirement,
when a non-leader server receives an RPC request it forwards it to the cluster leader.
-The server nodes also operate as part of a WAN gossip. This pool is different from the LAN pool,
-as it is optimized for the higher latency of the internet, and is expected to only contain
+The server nodes also operate as part of a WAN gossip pool. This pool is different from the LAN pool,
+as it is optimized for the higher latency of the internet, and is expected to contain only
other Consul server nodes. The purpose of this pool is to allow datacenters to discover each
other in a low touch manner. Bringing a new datacenter online is as easy as joining the existing
WAN gossip. Because the servers are all operating in this pool, it also enables cross-datacenter requests.
@@ -110,8 +110,8 @@ connection caching and multiplexing, cross-datacenter requests are relatively fa
## Getting in depth
-At this point we've covered the high level architecture of Consul, but there are much
-more details to each of the sub-systems. The [consensus protocol](/docs/internals/consensus.html) is
+At this point we've covered the high level architecture of Consul, but there are many
+more details for each of the sub-systems. The [consensus protocol](/docs/internals/consensus.html) is
documented in detail, as is the [gossip protocol](/docs/internals/gossip.html). The [documentation](/docs/internals/security.html)
for the security model and protocols used are also available.
diff --git a/website/source/docs/internals/security.html.markdown b/website/source/docs/internals/security.html.markdown
index 5c8e2f195f..d9a22aae88 100644
--- a/website/source/docs/internals/security.html.markdown
+++ b/website/source/docs/internals/security.html.markdown
@@ -10,8 +10,8 @@ description: |-
Consul relies on both a lightweight gossip mechanism and an RPC system
to provide various features. Both of the systems have different security
-mechanisms that stem from their designs. However, the goals
-of Consuls security are to provide [confidentiality, integrity and authentication](http://en.wikipedia.org/wiki/Information_security).
+mechanisms that stem from their designs. However, the overall goal
+of Consul's security model is to provide [confidentiality, integrity and authentication](http://en.wikipedia.org/wiki/Information_security).
The [gossip protocol](/docs/internals/gossip.html) is powered by [Serf](http://www.serfdom.io/),
which uses a symmetric key, or shared secret, cryptosystem. There are more
@@ -19,10 +19,11 @@ details on the security of [Serf here](http://www.serfdom.io/docs/internals/secu
The RPC system supports using end-to-end TLS, with optional client authentication.
[TLS](http://en.wikipedia.org/wiki/Transport_Layer_Security) is a widely deployed asymmetric
-cryptosystem, and is the foundation of security on the Internet.
+cryptosystem, and is the foundation of security on the Web, as well as
+some other critical parts of the Internet.
This means Consul communication is protected against eavesdropping, tampering,
-or spoofing. This makes it possible to run Consul over untrusted networks such
+and spoofing. This makes it possible to run Consul over untrusted networks such
as EC2 and other shared hosting providers.
~> **Advanced Topic!** This page covers the technical details of
diff --git a/website/source/docs/internals/sessions.html.markdown b/website/source/docs/internals/sessions.html.markdown
index 62f451e643..7daea5ae87 100644
--- a/website/source/docs/internals/sessions.html.markdown
+++ b/website/source/docs/internals/sessions.html.markdown
@@ -64,9 +64,9 @@ systems to be built that require an operator to intervene in the
case of a failure, but preclude the possibility of a split-brain.
The final nuance is that sessions may provide a `lock-delay`. This
-is a time duration, between 0 and 60 second. When a session invalidation
+is a time duration, between 0 and 60 seconds. When a session invalidation
takes place, Consul prevents any of the previously held locks from
-being re-acquired for the `lock-delay` interval; this is a safe guard
+being re-acquired for the `lock-delay` interval; this is a safeguard
inspired by Google's Chubby. The purpose of this delay is to allow
the potentially still live leader to detect the invalidation and stop
processing requests that may lead to inconsistent state. While not a
@@ -79,7 +79,7 @@ mechanism by providing a zero delay value.
Integration between the Key/Value store and sessions are the primary
place where sessions are used. A session must be created prior to use,
-and is then referred to by it's ID.
+and is then referred to by its ID.
The Key/Value API is extended to support an `acquire` and `release` operation.
The `acquire` operation acts like a Check-And-Set operation, except it
@@ -93,7 +93,7 @@ since the request will fail if given an invalid session. A critical note is
that the lock can be released without being the creator of the session.
This is by design, as it allows operators to intervene and force terminate
a session if necessary. As mentioned above, a session invalidation will also
-cause all held locks to be released. When a lock is released, the `LockIndex`,
+cause all held locks to be released. When a lock is released, the `LockIndex`
does not change, however the `Session` is cleared and the `ModifyIndex` increments.
These semantics (heavily borrowed from Chubby), allow the tuple of (Key, LockIndex, Session)
@@ -103,7 +103,7 @@ is incremented on each `acquire`, even if the same session re-acquires a lock,
the `sequencer` will be able to detect a stale request. Similarly, if a session is
invalided, the Session corresponding to the given `LockIndex` will be blank.
-To make clear, this locking system is purely *advisory*. There is no enforcement
+To be clear, this locking system is purely *advisory*. There is no enforcement
that clients must acquire a lock to perform any operation. Any client can
read, write, and delete a key without owning the corresponding lock. It is not
the goal of Consul to protect against misbehaving clients.
diff --git a/website/source/intro/getting-started/agent.html.markdown b/website/source/intro/getting-started/agent.html.markdown
index cb3508e48a..27dced736b 100644
--- a/website/source/intro/getting-started/agent.html.markdown
+++ b/website/source/intro/getting-started/agent.html.markdown
@@ -8,8 +8,8 @@ description: |-
# Run the Consul Agent
-After Consul is installed, the agent must be run. The agent can either run
-in a server or client mode. Each datacenter must have at least one server,
+After Consul is installed, the agent must be run. The agent can run either
+in server or client mode. Each datacenter must have at least one server,
although 3 or 5 is recommended. A single server deployment is _**highly**_ discouraged
as data loss is inevitable in a failure scenario. [This guide](/docs/guides/bootstrapping.html)
covers bootstrapping a new datacenter. All other agents run in client mode, which
@@ -114,10 +114,11 @@ By gracefully leaving, Consul notifies other cluster members that the
node _left_. If you had forcibly killed the agent process, other members
of the cluster would have detected that the node _failed_. When a member leaves,
its services and checks are removed from the catalog. When a member fails,
-its health is simply marked as critical, but is not removed from the catalog.
+its health is simply marked as critical, but it is not removed from the catalog.
Consul will automatically try to reconnect to _failed_ nodes, which allows it
to recover from certain network conditions, while _left_ nodes are no longer contacted.
Additionally, if an agent is operating as a server, a graceful leave is important
to avoid causing a potential availability outage affecting the [consensus protocol](/docs/internals/consensus.html).
-See the [guides section](/docs/guides/index.html) to safely add and remove servers.
+See the [guides section](/docs/guides/index.html) for how to safely add
+and remove servers.
diff --git a/website/source/intro/getting-started/checks.html.markdown b/website/source/intro/getting-started/checks.html.markdown
index a072c2a647..e4f8210b11 100644
--- a/website/source/intro/getting-started/checks.html.markdown
+++ b/website/source/intro/getting-started/checks.html.markdown
@@ -19,9 +19,8 @@ two node cluster running.
## Defining Checks
Similarly to a service, a check can be registered either by providing a
-[check definition](/docs/agent/checks.html)
-, or by making the appropriate calls to the
-[HTTP API](/docs/agent/http.html).
+[check definition](/docs/agent/checks.html), or by making the
+appropriate calls to the [HTTP API](/docs/agent/http.html).
We will use the check definition, because just like services, definitions
are the most common way to setup checks.
diff --git a/website/source/intro/getting-started/join.html.markdown b/website/source/intro/getting-started/join.html.markdown
index bf0f4bacfb..eed68b475e 100644
--- a/website/source/intro/getting-started/join.html.markdown
+++ b/website/source/intro/getting-started/join.html.markdown
@@ -14,12 +14,12 @@ Consul, but didn't show how this could be extended to a scalable production
service discovery infrastructure. On this page, we'll create our first
real cluster with multiple members.
-When starting a Consul agent, it begins without knowledge of any other node, and is
-an isolated cluster of one. To learn about other cluster members, the agent must
-_join_ an existing cluster. To join an existing cluster, it only needs to know
-about a _single_ existing member. After it joins, the agent will gossip with this
-member and quickly discover the other members in the cluster. A Consul
-agent can join any other agent, it doesn't have to be an agent in server mode.
+When a Consul agent is started, it begins without knowledge of any other node,
+and is an isolated cluster of one. To learn about other cluster members, the
+agent must _join_ an existing cluster. To join an existing cluster, it only
+needs to know about a _single_ existing member. After it joins, the agent will
+gossip with this member and quickly discover the other members in the cluster.
+A Consul agent can join any other agent, not just agents in server mode.
## Starting the Agents
@@ -78,7 +78,7 @@ agent-one 172.20.20.10:8301 alive role=consul,dc=dc1,vsn=2,vsn_min=1,vsn_m
agent-two 172.20.20.11:8301 alive role=node,dc=dc1,vsn=2,vsn_min=1,vsn_max=2
```
--> **Remember:** To join a cluster, a Consul agent needs to only
+-> **Remember:** To join a cluster, a Consul agent only needs to
learn about one existing member. After joining the cluster, the
agents gossip with each other to propagate full membership information.
diff --git a/website/source/intro/getting-started/kv.html.markdown b/website/source/intro/getting-started/kv.html.markdown
index 7c2249ce8d..e2ecd7a6c1 100644
--- a/website/source/intro/getting-started/kv.html.markdown
+++ b/website/source/intro/getting-started/kv.html.markdown
@@ -62,7 +62,7 @@ $ curl http://localhost:8500/v1/kv/?recurse
Here we have created 3 keys, each with the value of "test". Note that the
`Value` field returned is base64 encoded to allow non-UTF8
characters. For the "web/key2" key, we set a `flag` value of 42. All keys
-support setting a 64bit integer flag value. This is opaque to Consul but can
+support setting a 64-bit integer flag value. This is opaque to Consul but can
be used by clients for any purpose.
After setting the values, we then issued a GET request to retrieve multiple
@@ -115,5 +115,5 @@ returning the current, unchanged value. This can be used to efficiently wait for
key modifications. Additionally, this same technique can be used to wait for a list
of keys, waiting only until any of the keys has a newer modification time.
-This is only a few example of what the API supports. For full documentation, please
-reference the [HTTP API](/docs/agent/http.html).
+These are only a few examples of what the API supports. For full
+documentation, please see the [HTTP API](/docs/agent/http.html).
diff --git a/website/source/intro/getting-started/services.html.markdown b/website/source/intro/getting-started/services.html.markdown
index 9b5e212bc9..a2839af5a0 100644
--- a/website/source/intro/getting-started/services.html.markdown
+++ b/website/source/intro/getting-started/services.html.markdown
@@ -67,9 +67,9 @@ service using either the DNS or HTTP API.
Let's first query it using the DNS API. For the DNS API, the DNS name
for services is `NAME.service.consul`. All DNS names are always in the
-`consul` namespace. The `service` subdomain on that tells Consul we're
-querying services, and the `NAME` is the name of the service. For the
-web service we registered, that would be `web.service.consul`:
+`consul` namespace. The `service` subdomain tells Consul we're querying
+services, and the `NAME` is the name of the service. For the web service
+we registered, that would be `web.service.consul`:
```text
$ dig @127.0.0.1 -p 8600 web.service.consul