Skip to content

Commit

Permalink
Fix Typos in Documentation, Comments, and Logs (apache#12027)
Browse files Browse the repository at this point in the history
* Fix Typos in Documentation, Comments, and Logs

* Relace url with URL

* Upper case Pulsar

* Fix typo in test assertion
  • Loading branch information
michaeljmarshall authored Sep 29, 2021
1 parent 41ea8b9 commit 52d1482
Show file tree
Hide file tree
Showing 100 changed files with 147 additions and 147 deletions.
4 changes: 2 additions & 2 deletions .github/ISSUE_TEMPLATE/flaky_test.md
Original file line number Diff line number Diff line change
Expand Up @@ -12,8 +12,8 @@ Instructions for reporting a flaky test using this issue template:
1. Replace [test class] in title and body with the test class name
2. Replace [test method] in title and body with the test method that failed. Multiple methods are flaky, remove the content that refers to the test method.
3. Replace "url here" with a url to an example failure. In the Github Actions workflow run logs, you can right click on the line number to copy a link to the line. Example of such url is https://github.com/apache/pulsar/pull/8892/checks?check_run_id=1531075794#step:9:377 . The logs are available for a limited amount of time (usually for a few weeks).
4. Replace "relevant parts of the exception stacktrace here" with the a few lines of the stack trace that shows at leat the exception message and the line of test code where the stacktrace occured.
5. Replace "full exception stacktrace here" with the full exception stacktrace from logs. This section will be hidded by default.
4. Replace "relevant parts of the exception stacktrace here" with the a few lines of the stack trace that shows at least the exception message and the line of test code where the stacktrace occurred.
5. Replace "full exception stacktrace here" with the full exception stacktrace from logs. This section will be hidden by default.
6. Remove all unused fields / content to unclutter the reported issue. Remove this comment too.
-->
Expand Down
2 changes: 1 addition & 1 deletion bin/pulsar
Original file line number Diff line number Diff line change
Expand Up @@ -50,7 +50,7 @@ then
. "$PULSAR_HOME/conf/bkenv.sh"
fi

# Check pulsar env and load pulser_env.sh
# Check pulsar env and load pulsar_env.sh
if [ -f "$PULSAR_HOME/conf/pulsar_env.sh" ]
then
. "$PULSAR_HOME/conf/pulsar_env.sh"
Expand Down
2 changes: 1 addition & 1 deletion conf/pulsar_env.sh
Original file line number Diff line number Diff line change
Expand Up @@ -66,6 +66,6 @@ PULSAR_EXTRA_OPTS=${PULSAR_EXTRA_OPTS:-" -Dpulsar.allocator.exit_on_oom=true -Di
#Folder where the Bookie server PID file should be stored
#PULSAR_PID_DIR=

#Wait time before forcefully kill the pulser server instance, if the stop is not successful
#Wait time before forcefully kill the pulsar server instance, if the stop is not successful
#PULSAR_STOP_TIMEOUT=

2 changes: 1 addition & 1 deletion conf/pulsar_tools_env.sh
Original file line number Diff line number Diff line change
Expand Up @@ -56,5 +56,5 @@ PULSAR_EXTRA_OPTS="${PULSAR_EXTRA_OPTS} ${PULSAR_MEM} ${PULSAR_GC} ${PULSAR_GC_L
#Folder where the Bookie server PID file should be stored
#PULSAR_PID_DIR=

#Wait time before forcefully kill the pulser server instance, if the stop is not successful
#Wait time before forcefully kill the pulsar server instance, if the stop is not successful
#PULSAR_STOP_TIMEOUT=
2 changes: 1 addition & 1 deletion deployment/terraform-ansible/templates/pulsar_env.sh
Original file line number Diff line number Diff line change
Expand Up @@ -56,5 +56,5 @@ PULSAR_EXTRA_OPTS="${PULSAR_EXTRA_OPTS} ${PULSAR_MEM} ${PULSAR_GC} -Dio.netty.le
#Folder where the Bookie server PID file should be stored
#PULSAR_PID_DIR=

#Wait time before forcefully kill the pulser server instance, if the stop is not successful
#Wait time before forcefully kill the pulsar server instance, if the stop is not successful
#PULSAR_STOP_TIMEOUT=
2 changes: 1 addition & 1 deletion faq.md
Original file line number Diff line number Diff line change
Expand Up @@ -25,7 +25,7 @@ There are few parameters to look at :
4. Initially the default is to have 4 “bundles” for a namespace
5. When the traffic increases on a given bundle, it will be split in 2 and reassigned to a different broker
6. There are some adjustable thresholds that can be used to control when the split happens, based on number of topics/partitions, messages in/out, bytes in/out, etc..
7. It’s also possible to specify a higher number of bundles when creating a namepsac
7. It’s also possible to specify a higher number of bundles when creating a namespace
8. There are the load-manager threshold that control when a broker should offload some of the bundles to other brokers

### What is the lifecycle of subscription?
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -22,7 +22,7 @@
import org.apache.pulsar.common.policies.data.BrokerStatus;

/**
* Basic defintion of an auto-failover policy.
* Basic definition of an auto-failover policy.
*/
public abstract class AutoFailoverPolicy {

Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -133,7 +133,7 @@ private static String[] splitParameter(List<String> params, int n) {

String[] parts = params.get(0).split("/");
if (parts.length != n) {
throw new ParameterException("Paramter format is incorrect");
throw new ParameterException("Parameter format is incorrect");
}

return parts;
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -83,7 +83,7 @@ void runCmd() throws Exception {
}
}

@Parameters(commandDescription = "Get the assignments of the functions accross the worker cluster")
@Parameters(commandDescription = "Get the assignments of the functions across the worker cluster")
class GetFunctionAssignments extends BaseCommand {


Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -158,7 +158,7 @@ void run() throws PulsarAdminException {
String namespace = validateNamespace(params);
if (numBundles < 0 || numBundles > MAX_BUNDLES) {
throw new ParameterException(
"Invalid number of bundles. Number of numbles has to be in the range of (0, 2^32].");
"Invalid number of bundles. Number of bundles has to be in the range of (0, 2^32].");
}

NamespaceName namespaceName = NamespaceName.get(namespace);
Expand Down Expand Up @@ -1175,7 +1175,7 @@ private class SetPersistence extends CliCommand {
private int bookkeeperWriteQuorum;

@Parameter(names = { "-a",
"--bookkeeper-ack-quorum" }, description = "Number of acks (garanteed copies) to wait for each entry", required = true)
"--bookkeeper-ack-quorum" }, description = "Number of acks (guaranteed copies) to wait for each entry", required = true)
private int bookkeeperAckQuorum;

@Parameter(names = { "-r",
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -548,7 +548,7 @@ void run() throws PulsarAdminException {

try {
MessageId lastMessageId = getPersistentTopics().terminateTopicAsync(persistentTopic).get();
System.out.println("Topic succesfully terminated at " + lastMessageId);
System.out.println("Topic successfully terminated at " + lastMessageId);
} catch (InterruptedException | ExecutionException e) {
throw new PulsarAdminException(e);
}
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -808,7 +808,7 @@ void run() throws PulsarAdminException {

try {
MessageId lastMessageId = getTopics().terminateTopicAsync(persistentTopic).get();
System.out.println("Topic succesfully terminated at " + lastMessageId);
System.out.println("Topic successfully terminated at " + lastMessageId);
} catch (InterruptedException | ExecutionException e) {
throw new PulsarAdminException(e);
}
Expand Down Expand Up @@ -1668,7 +1668,7 @@ private class SetPersistence extends CliCommand {
private int bookkeeperWriteQuorum;

@Parameter(names = { "-a",
"--bookkeeper-ack-quorum" }, description = "Number of acks (garanteed copies) to wait for each entry", required = true)
"--bookkeeper-ack-quorum" }, description = "Number of acks (guaranteed copies) to wait for each entry", required = true)
private int bookkeeperAckQuorum;

@Parameter(names = { "-r",
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -181,7 +181,7 @@ public class WorkerConfig implements Serializable, PulsarConfiguration {
private String functionsDirectory = "./functions";
@FieldContext(
category = CATEGORY_FUNC_METADATA_MNG,
doc = "The pulsar topic used for storing function metadata"
doc = "The Pulsar topic used for storing function metadata"
)
private String functionMetadataTopicName;
@FieldContext(
Expand All @@ -191,32 +191,32 @@ public class WorkerConfig implements Serializable, PulsarConfiguration {
private Boolean useCompactedMetadataTopic = false;
@FieldContext(
category = CATEGORY_FUNC_METADATA_MNG,
doc = "The web service url for function workers"
doc = "The web service URL for function workers"
)
private String functionWebServiceUrl;
@FieldContext(
category = CATEGORY_FUNC_METADATA_MNG,
doc = "The pulser binary service url that function metadata manager talks to"
doc = "The Pulsar binary service URL that function metadata manager talks to"
)
private String pulsarServiceUrl;
@FieldContext(
category = CATEGORY_FUNC_METADATA_MNG,
doc = "The pulsar web service url that function metadata manager talks to"
doc = "The Pulsar web service URL that function metadata manager talks to"
)
private String pulsarWebServiceUrl;
@FieldContext(
category = CATEGORY_FUNC_METADATA_MNG,
doc = "The pulsar topic used for cluster coordination"
doc = "The Pulsar topic used for cluster coordination"
)
private String clusterCoordinationTopicName;
@FieldContext(
category = CATEGORY_FUNC_METADATA_MNG,
doc = "The pulsar namespace for storing metadata topics"
doc = "The Pulsar namespace for storing metadata topics"
)
private String pulsarFunctionsNamespace;
@FieldContext(
category = CATEGORY_FUNC_METADATA_MNG,
doc = "The pulsar cluster name. Used for creating pulsar namespace during worker initialization"
doc = "The Pulsar cluster name. Used for creating Pulsar namespace during worker initialization"
)
private String pulsarFunctionsCluster;
@FieldContext(
Expand All @@ -231,12 +231,12 @@ public class WorkerConfig implements Serializable, PulsarConfiguration {
private String downloadDirectory;
@FieldContext(
category = CATEGORY_STATE,
doc = "The service url of state storage"
doc = "The service URL of state storage"
)
private String stateStorageServiceUrl;
@FieldContext(
category = CATEGORY_FUNC_RUNTIME_MNG,
doc = "The pulsar topic used for storing function assignment informations"
doc = "The Pulsar topic used for storing function assignment informations"
)
private String functionAssignmentTopicName;
@FieldContext(
Expand Down Expand Up @@ -304,7 +304,7 @@ public boolean isBrokerClientAuthenticationEnabled() {
private String bookkeeperClientAuthenticationPlugin;
@FieldContext(
category = CATEGORY_CLIENT_SECURITY,
doc = "BookKeeper auth plugin implementatation specifics parameters name and values"
doc = "BookKeeper auth plugin implementation specifics parameters name and values"
)
private String bookkeeperClientAuthenticationParametersName;
@FieldContext(
Expand Down Expand Up @@ -413,7 +413,7 @@ public Boolean isInitializedDlogMetadata() {
return this.initializedDlogMetadata;
};

/******** security settings for pulsar broker client **********/
/******** security settings for Pulsar broker client **********/

@FieldContext(
category = CATEGORY_CLIENT_SECURITY,
Expand Down
6 changes: 3 additions & 3 deletions site2/docs/administration-zk-bk.md
Original file line number Diff line number Diff line change
Expand Up @@ -65,7 +65,7 @@ If you deploy a [single-cluster](#single-cluster-pulsar-instance) instance, you

If your Pulsar instance consists of just one cluster, then you can deploy a configuration store on the same machines as the local ZooKeeper quorum but run on different TCP ports.

To deploy a ZooKeeper configuration store in a single-cluster instance, add the same ZooKeeper servers that the local quorom uses to the configuration file in [`conf/global_zookeeper.conf`](reference-configuration.md#configuration-store) using the same method for [local ZooKeeper](#local-zookeeper), but make sure to use a different port (2181 is the default for ZooKeeper). The following is an example that uses port 2184 for a three-node ZooKeeper cluster:
To deploy a ZooKeeper configuration store in a single-cluster instance, add the same ZooKeeper servers that the local quorum uses to the configuration file in [`conf/global_zookeeper.conf`](reference-configuration.md#configuration-store) using the same method for [local ZooKeeper](#local-zookeeper), but make sure to use a different port (2181 is the default for ZooKeeper). The following is an example that uses port 2184 for a three-node ZooKeeper cluster:

```properties
clientPort=2184
Expand Down Expand Up @@ -276,7 +276,7 @@ Use the [`set-persistence`](reference-pulsar-admin.md#namespaces-set-persistence

Flag | Description | Default
:----|:------------|:-------
`-a`, `--bookkeeper-ack-quorom` | The number of acks (guaranteed copies) to wait on for each entry | 0
`-a`, `--bookkeeper-ack-quorum` | The number of acks (guaranteed copies) to wait on for each entry | 0
`-e`, `--bookkeeper-ensemble` | The number of [bookies](reference-terminology.md#bookie) to use for topics in the namespace | 0
`-w`, `--bookkeeper-write-quorum` | The number of writes to make for each entry | 0
`-r`, `--ml-mark-delete-max-rate` | Throttling rate for mark-delete operations (0 means no throttle) | 0
Expand All @@ -285,7 +285,7 @@ The following is an example:

```shell
$ pulsar-admin namespaces set-persistence my-tenant/my-ns \
--bookkeeper-ack-quorom 3 \
--bookkeeper-ack-quorum 3 \
--bookeeper-ensemble 2
```

Expand Down
2 changes: 1 addition & 1 deletion site2/docs/client-libraries-node.md
Original file line number Diff line number Diff line change
Expand Up @@ -128,7 +128,7 @@ Pulsar Node.js producers have the following methods available:
| Method | Description | Return type |
| :----- | :---------- | :---------- |
| `send(Object)` | Publishes a [message](#messages) to the producer's topic. When the message is successfully acknowledged by the Pulsar broker, or an error is thrown, the Promise object whose result is the message ID runs executor function. | `Promise<Object>` |
| `flush()` | Sends message from send queue to Pulser broker. When the message is successfully acknowledged by the Pulsar broker, or an error is thrown, the Promise object runs executor function. | `Promise<null>` |
| `flush()` | Sends message from send queue to Pulsar broker. When the message is successfully acknowledged by the Pulsar broker, or an error is thrown, the Promise object runs executor function. | `Promise<null>` |
| `close()` | Closes the producer and releases all resources allocated to it. Once `close()` is called, no more messages are accepted from the publisher. This method returns a Promise object. It runs the executor function when all pending publish requests are persisted by Pulsar. If an error is thrown, no pending writes are retried. | `Promise<null>` |
| `getProducerName()` | Getter method of the producer name. | `string` |
| `getTopic()` | Getter method of the name of the topic. | `string` |
Expand Down
2 changes: 1 addition & 1 deletion site2/docs/deploy-bare-metal-multi-cluster.md
Original file line number Diff line number Diff line change
Expand Up @@ -128,7 +128,7 @@ If you deploy a single-cluster instance, you do not need a separate cluster for

If your Pulsar instance consists of just one cluster, then you can deploy a configuration store on the same machines as the local ZooKeeper quorum but run on different TCP ports.

To deploy a ZooKeeper configuration store in a single-cluster instance, add the same ZooKeeper servers that the local quorom.You need to use the configuration file in [`conf/global_zookeeper.conf`](reference-configuration.md#configuration-store) using the same method for [local ZooKeeper](#local-zookeeper), but make sure to use a different port (2181 is the default for ZooKeeper). The following is an example that uses port 2184 for a three-node ZooKeeper cluster:
To deploy a ZooKeeper configuration store in a single-cluster instance, add the same ZooKeeper servers that the local quorum. You need to use the configuration file in [`conf/global_zookeeper.conf`](reference-configuration.md#configuration-store) using the same method for [local ZooKeeper](#local-zookeeper), but make sure to use a different port (2181 is the default for ZooKeeper). The following is an example that uses port 2184 for a three-node ZooKeeper cluster:

```properties
clientPort=2184
Expand Down
2 changes: 1 addition & 1 deletion site2/docs/functions-worker.md
Original file line number Diff line number Diff line change
Expand Up @@ -119,7 +119,7 @@ To run function-worker separately, you have to configure the following parameter
#### Function metadata parameter

- `pulsarServiceUrl`: The Pulsar service URL for your broker cluster.
- `pulsarWebServiceUrl`: The Pulser web service URL for your broker cluster.
- `pulsarWebServiceUrl`: The Pulsar web service URL for your broker cluster.
- `pulsarFunctionsCluster`: Set the value to your Pulsar cluster name (same as the `clusterName` setting in the broker configuration).

If authentication is enabled for your broker cluster, you *should* configure the authentication plugin and parameters for the functions worker to communicate with the brokers.
Expand Down
2 changes: 1 addition & 1 deletion site2/website-next/docs/functions-worker.md
Original file line number Diff line number Diff line change
Expand Up @@ -130,7 +130,7 @@ To run function-worker separately, you have to configure the following parameter
#### Function metadata parameter

- `pulsarServiceUrl`: The Pulsar service URL for your broker cluster.
- `pulsarWebServiceUrl`: The Pulser web service URL for your broker cluster.
- `pulsarWebServiceUrl`: The Pulsar web service URL for your broker cluster.
- `pulsarFunctionsCluster`: Set the value to your Pulsar cluster name (same as the `clusterName` setting in the broker configuration).

If authentication is enabled for your broker cluster, you *should* configure the authentication plugin and parameters for the functions worker to communicate with the brokers.
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -132,7 +132,7 @@ To run function-worker separately, you have to configure the following parameter
#### Function metadata parameter

- `pulsarServiceUrl`: The Pulsar service URL for your broker cluster.
- `pulsarWebServiceUrl`: The Pulser web service URL for your broker cluster.
- `pulsarWebServiceUrl`: The Pulsar web service URL for your broker cluster.
- `pulsarFunctionsCluster`: Set the value to your Pulsar cluster name (same as the `clusterName` setting in the broker configuration).

If authentication is enabled for your broker cluster, you *should* configure the authentication plugin and parameters for the functions worker to communicate with the brokers.
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -131,7 +131,7 @@ To run function-worker separately, you have to configure the following parameter
#### Function metadata parameter

- `pulsarServiceUrl`: The Pulsar service URL for your broker cluster.
- `pulsarWebServiceUrl`: The Pulser web service URL for your broker cluster.
- `pulsarWebServiceUrl`: The Pulsar web service URL for your broker cluster.
- `pulsarFunctionsCluster`: Set the value to your Pulsar cluster name (same as the `clusterName` setting in the broker configuration).

If authentication is enabled for your broker cluster, you *should* configure the authentication plugin and parameters for the functions worker to communicate with the brokers.
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -66,7 +66,7 @@ If you're deploying a [single-cluster](#single-cluster-pulsar-instance) instance

If your Pulsar instance will consist of just one cluster, then you can deploy a configuration store on the same machines as the local ZooKeeper quorum but running on different TCP ports.

To deploy a ZooKeeper configuration store in a single-cluster instance, add the same ZooKeeper servers used by the local quorom to the configuration file in [`conf/global_zookeeper.conf`](reference-configuration.md#configuration-store) using the same method for [local ZooKeeper](#local-zookeeper), but make sure to use a different port (2181 is the default for ZooKeeper). Here's an example that uses port 2184 for a three-node ZooKeeper cluster:
To deploy a ZooKeeper configuration store in a single-cluster instance, add the same ZooKeeper servers used by the local quorum to the configuration file in [`conf/global_zookeeper.conf`](reference-configuration.md#configuration-store) using the same method for [local ZooKeeper](#local-zookeeper), but make sure to use a different port (2181 is the default for ZooKeeper). Here's an example that uses port 2184 for a three-node ZooKeeper cluster:

```properties
clientPort=2184
Expand Down Expand Up @@ -251,7 +251,7 @@ Use the [`set-persistence`](reference-pulsar-admin.md#namespaces-set-persistence

Flag | Description | Default
:----|:------------|:-------
`-a`, `--bookkeeper-ack-quorom` | The number of acks (guaranteed copies) to wait on for each entry | 0
`-a`, `--bookkeeper-ack-quorum` | The number of acks (guaranteed copies) to wait on for each entry | 0
`-e`, `--bookkeeper-ensemble` | The number of [bookies](reference-terminology.md#bookie) to use for topics in the namespace | 0
`-w`, `--bookkeeper-write-quorum` | How many writes to make for each entry | 0
`-r`, `--ml-mark-delete-max-rate` | Throttling rate for mark-delete operations (0 means no throttle) | 0
Expand All @@ -260,7 +260,7 @@ Flag | Description | Default

```shell
$ pulsar-admin namespaces set-persistence my-tenant/my-ns \
--bookkeeper-ack-quorom 3 \
--bookkeeper-ack-quorum 3 \
--bookeeper-ensemble 2
```

Expand Down
Loading

0 comments on commit 52d1482

Please sign in to comment.