summaryrefslogtreecommitdiff
path: root/sql/sys_vars.cc
diff options
context:
space:
mode:
authorOtto Kekalainen <ottoke@amazon.com>2023-03-11 11:07:48 -0800
committerDaniel Black <daniel@mariadb.org>2023-03-24 12:54:05 +1100
commit50c8ef01fc63e32acd38788107ae4029d0f1b9ce (patch)
tree1ae7f3757935d720d3975c76766f11f494777717 /sql/sys_vars.cc
parentfaeca0008bd4ef90b1c9b41171d47b7892deac5e (diff)
downloadmariadb-git-50c8ef01fc63e32acd38788107ae4029d0f1b9ce.tar.gz
Fix trivial spelling errors
- agressively -> aggressively - exising -> existing - occured -> occurred - releated -> related - seperated -> separated - sucess -> success - use use -> use All new code of the whole pull request, including one or several files that are either new files or modified ones, are contributed under the BSD-new license. I am contributing on behalf of my employer Amazon Web Services, Inc.
Diffstat (limited to 'sql/sys_vars.cc')
-rw-r--r--sql/sys_vars.cc4
1 files changed, 2 insertions, 2 deletions
diff --git a/sql/sys_vars.cc b/sql/sys_vars.cc
index 1ed3d61bcf8..b1d2b4919f2 100644
--- a/sql/sys_vars.cc
+++ b/sql/sys_vars.cc
@@ -2063,7 +2063,7 @@ Sys_gtid_strict_mode(
"stops with an error if it encounters an event that would cause it to "
"generate an out-of-order binlog if executed. "
"When ON the same server-id semisync-replicated transactions that "
- "duplicate exising ones in binlog are ignored without error "
+ "duplicate existing ones in binlog are ignored without error "
"and slave interruption.",
GLOBAL_VAR(opt_gtid_strict_mode),
CMD_LINE(OPT_ARG), DEFAULT(FALSE));
@@ -2828,7 +2828,7 @@ static Sys_var_ulong Sys_optimizer_search_depth(
static Sys_var_ulong Sys_optimizer_extra_pruning_depth(
"optimizer_extra_pruning_depth",
"If the optimizer needs to enumerate join prefix of this size or "
- "larger, then it will try agressively prune away the search space.",
+ "larger, then it will try aggressively prune away the search space.",
SESSION_VAR(optimizer_extra_pruning_depth), CMD_LINE(REQUIRED_ARG),
VALID_RANGE(0, MAX_TABLES+1), DEFAULT(8), BLOCK_SIZE(1));