summaryrefslogtreecommitdiff
Commit message (Collapse)AuthorAgeFilesLines
* Merge branch '10.7' into 10.8Oleksandr Byelkin2023-01-31140-715/+4106
|\
| * Merge branch '10.6' into 10.7Oleksandr Byelkin2023-01-31140-715/+4106
| |\
| | * Merge branch '10.5' into 10.6Oleksandr Byelkin2023-01-31133-705/+4071
| | |\
| | | * MDEV-30940: Revert "binlog.innodb_rc_insert_before_delete is disabled with ↵mariadb-10.5.19Andrei2023-01-302-1/+3
| | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | MDEV-30490" This reverts commit b2ea57e899b50cb428b6b58a21de5cfe1b191576, as well as edits binlog.innodb_rc_insert_before_delete.test to be safely runnable with any preceding test. Note: manual 10.5 -> 10.6 merge is required to the test.
| | | * binlog.innodb_rc_insert_before_delete is disabled with MDEV-30490Andrei2023-01-301-0/+1
| | | |
| | | * Merge branch '10.4' into 10.5Oleksandr Byelkin2023-01-306-12/+59
| | | |\
| | | | * Fix number of rows passing in case of EQ_REFmariadb-10.4.28Oleksandr Byelkin2023-01-301-3/+5
| | | | |
| | | | * MDEV-30218: Incorrect optimization for rowid_filtering, correctionSergei Petrunia2023-01-283-15/+73
| | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | Enable use of Rowid Filter optimization with eq_ref access. Use the following assumptions: - Assume index-only access cost is 50% of non-index-only access cost. - Take into account that "Eq_ref access cache" reduces the number of lookups eq_ref access will make. = This means the number of Rowid Filter checks is reduced also = Eq_ref access cost is computed using that assumption (see prev_record_reads() call), so we should use it in all cost ' computations.
| | | | * Merge branch '10.3' into 10.4Oleksandr Byelkin2023-01-2861-293/+1743
| | | | |\
| | | | | * MDEV-30010 post-push: fixing test results.mariadb-10.3.38Andrei2023-01-282-8/+52
| | | | | |
| | | * | | binlog.innodb_rc_insert_before_delete is disabled with MDEV-30490Andrei2023-01-281-0/+1
| | | | | |
| | | * | | Merge branch '10.4' into 10.5Oleksandr Byelkin2023-01-27114-649/+3235
| | | |\ \ \
| | | | * | | MDEV-30218: Incorrect optimization for rowid_filtering, correctionSergei Petrunia2023-01-263-15/+73
| | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | Enable use of Rowid Filter optimization with eq_ref access. Use the following assumptions: - Assume index-only access cost is 50% of non-index-only access cost. - Take into account that "Eq_ref access cache" reduces the number of lookups eq_ref access will make. = This means the number of Rowid Filter checks is reduced also = Eq_ref access cost is computed using that assumption (see prev_record_reads() call), so we should use it in all cost ' computations.
| | | | * | | Merge branch '10.3' into 10.4Oleksandr Byelkin2023-01-2661-293/+1699
| | | | |\ \ \ | | | | | |/ / | | | | |/| / | | | | | |/
| | | | | * MDEV-30457 Windows, signtool error "No file digest algorithm specified."Vladislav Vaintroub2023-01-241-1/+1
| | | | | | | | | | | | | | | | | | | | | | | | | | | | | | Add /fd parameter. It is now mandatory for the recent versions of signtool
| | | | | * MDEV-30323 Some DDLs like ANALYZE can complete on parallel slave out of orderbb-10.3-andreiAndrei2023-01-244-3/+237
| | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | ANALYZE was observed to race over a preceding in binlog order DML in updating the binlog and slave gtid states. Tagging ANALYZE and other admin class commands in binlog by the fixes of MDEV-17515 left a flaw allowing such race leading to the gtid mode out-of-order error. This is fixed now to observe by ADMIN commands the ordered access to the slave gtid status variables and binlog.
| | | | | * MDEV-30010 Slave (additional info): Commit failed due to failure of an ↵Andrei2023-01-242-0/+97
| | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | earlier commit on which this one depends Error_code: 1964 This commit merely adds is a Read-Committed version MDEV-30225 test solely to prove the RC isolation yields ROW binlog format as it is supposed to per docs.
| | | | | * MDEV-30081 Crash with splitting from constant mergeable derived tableIgor Babaev2023-01-243-1/+108
| | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | This bug manifested itself in very rare situations when splitting optimization was applied to a materialized derived table with group clause by key over a constant meargeable derived table that was in inner part of an outer join. In this case the used tables for the key to access the split table incorrectly was evaluated to a not empty table map. Approved by Oleksandr Byelkin <sanja@mariadb.com>
| | | | | * MDEV-29639: Seconds_Behind_Master is incorrect for Delayed, Parallel ReplicasBrandon Nesterenko2023-01-246-9/+231
| | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | Problem ======== On a parallel, delayed replica, Seconds_Behind_Master will not be calculated until after MASTER_DELAY seconds have passed and the event has finished executing, resulting in potentially very large values of Seconds_Behind_Master (which could be much larger than the MASTER_DELAY parameter) for the entire duration the event is delayed. This contradicts the documented MASTER_DELAY behavior, which specifies how many seconds to withhold replicated events from execution. Solution ======== After a parallel replica idles, the first event after idling should immediately update last_master_timestamp with the time that it began execution on the primary. Reviewed By =========== Andrei Elkin <andrei.elkin@mariadb.com>
| | | | | * Added comments re JOIN::all_fields, JOIN::fields_listSergei Petrunia2023-01-241-2/+20
| | | | | |
| | | | | * Update 10.3 HELP tablesIan Gilfillan2023-01-241-46/+46
| | | | | |
| | | | | * MDEV-30248 Infinite sequence of recursive calls when processing embedded CTEIgor Babaev2023-01-235-81/+311
| | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | This patch fixes the patch for bug MDEV-30248 that unsatisfactorily resolved the problem of resolution of references to CTE. In some cases when such a reference has the same table name as the name of one of CTEs containing this reference the reference could be resolved incorrectly that led to an invalid select tree where units could be mutually dependent. This in its turn could lead to an infinite sequence of recursive calls or to falls into infinite loops. The patch also removes LEX::resolve_references_to_cte_in_hanging_cte() as with the new code for resolution of CTE references the call of this function is not needed anymore. Approved by Oleksandr Byelkin <sanja@mariadb.com>
| | | | | * MDEV-15178: Filesort::make_sortorder: Assertion `pos->field != __null |Sergei Petrunia2023-01-236-19/+188
| | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | (Initial patch by Varun Gupta. Amended and added comments). When the query has both 1. Aggregate functions that require sorting data by group, and 2. Window functions we need to use two temporary tables. The first temp.table will hold the join output. Then it is passed to filesort(). Reading it in sorted order allows to compute the aggregate functions. Then, we need to write their values into the second temp. table. Then, Window Function computation step can pass that to filesort() and read them in the order it needs. Failure to create the second temp. table would cause an assertion failure: window function could would not find where to get the values of the aggregate functions.
| | | | | * Fix connect bson.cpp warningVicențiu Ciorbaru2023-01-201-1/+1
| | | | | | | | | | | | | | | | | | | | | | | | The ptyp variable is unused.
| | | | | * Fix mroonga warning of use-after-freeVicențiu Ciorbaru2023-01-201-2/+2
| | | | | |
| | | | | * Minimize unsafe C functions usage - replace strcat() and strcpy() (and ↵Mikhail Chalov2023-01-2023-162/+221
| | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | strncat() and strncpy()) with custom safe_strcat() and safe_strcpy() functions The MariaDB code base uses strcat() and strcpy() in several places. These are known to have memory safety issues and their usage is discouraged. Common security scanners like Flawfinder flags them. In MariaDB we should start using modern and safer variants on these functions. This is similar to memory issues fixes in 19af1890b56c6c147c296479bb6a4ad00fa59dbb and 9de9f105b5cb88249acc39af73d32af337d6fd5f but now replace use of strcat() and strcpy() with safer options strncat() and strncpy(). However, add '\0' forcefully to make sure the result string is correct since for these two functions it is not guaranteed what new string will be null-terminated. Example: size_t dest_len = sizeof(g->Message); strncpy(g->Message, "Null json tree", dest_len); strncat(g->Message, ":", sizeof(g->Message) - strlen(g->Message)); size_t wrote_sz = strlen(g->Message); size_t cur_len = wrote_sz >= dest_len ? dest_len - 1 : wrote_sz; g->Message[cur_len] = '\0'; All new code of the whole pull request, including one or several files that are either new files or modified ones, are contributed under the BSD-new license. I am contributing on behalf of my employer Amazon Web Services -- Reviewer and co-author Vicențiu Ciorbaru <vicentiu@mariadb.org> -- Reviewer additions: * The initial function implementation was flawed. Replaced with a simpler and also correct version. * Simplified code by making use of snprintf instead of chaining strcat. * Simplified code by removing dynamic string construction in the first place and using static strings if possible. See connect storage engine changes.
| | | | | * MDEV-30052 Crash with a query containing nested WINDOW clausesIgor Babaev2023-01-206-17/+130
| | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | Use SELECT_LEX to save lists for ORDER BY and GROUP BY before parsing WINDOW clauses / specifications. This is needed for proper parsing of a nested WINDOW clause when a WINDOW clause is used in a subquery contained in another WINDOW clause. Fix assignment of empty SQL_I_List to another one (in case of empty list next shoud point on first).
| | | | | * Add my_afree after my_alloca in early return caseEric Herman2023-01-201-0/+3
| | | | | | | | | | | | | | | | | | | | | | | | | | | | | | The code already had a call to `my_afree` in the normal return case, but failed to do so in the early return case.
| | | | | * MDEV-23335 MariaBackup Incremental Does Not Reflect Dropped/Created DatabasesAlexander Barkov2023-01-193-1/+146
| | | | | |
| | | | * | CREDITS: re-instate Tencent CloudDaniel Black2023-01-251-0/+1
| | | | | | | | | | | | | | | | | | | | | | | | This was an accidential deletion looking at an outdated list.
| | | | * | MDEV-30370 Fixing spider hang when server abortsYuchen Pei2023-01-253-4/+25
| | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | This is Kentoku's patch for MDEV-22979 (e6e41f04f4e + 22a0097727f), which fixes 30370. It changes the wait to a timed wait for the first sts thread, which waits on server start to execute the init queries for spider. It also flips the flag init_command to false when the sts thread is being freed. With these changes the sts thread can check the flag regularly and abort the init_queries when it finds out the init_command is false. This avoids the deadlock that causes the problem in MDEV-30370. It also fixes MDEV-22979 for 10.4, but not 10.5. I have not tested higher versions for MDEV-22979. A test has also been done on MDEV-29904 to avoid regression, given MDEV-27233 is a similar problem and its patch caused the regression. The test passes for 10.4-11.0. However, this adhoc test only works consistently when placed in the main testsuite. We should not place spider tests in the main suite, so we do not include it in this commit. A patch for MDEV-27912 should fix this problem and allow a proper test for MDEV-29904. See comments in the jira ticket MDEV-30370/29904 for the adhoc testcase used for this commit.
| | | | * | Update 10.4 HELP tablesIan Gilfillan2023-01-241-46/+46
| | | | | |
| | | | * | MDEV-29461 AddressSanitizer: stack-buffer-overflow in strxmovSergei Golubchik2023-01-201-1/+1
| | | | | |
| | | | * | MDEV-26817 runtime error: index 24320 out of bounds for type ↵Sergei Golubchik2023-01-201-1/+1
| | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | 'json_string_char_classes [128] *and* ASAN: global-buffer-overflow on address ... READ of size 4 on SELECT JSON_VALID protect from out-of-bound array access it was already done in all other places, this one was the only one missed
| | | | * | MDEV-29199 Unique hash key is ignored upon INSERT ... SELECT into non-empty ↵Sergei Golubchik2023-01-203-97/+211
| | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | MyISAM table disable bulk insert optimization if long uniques are used, because they need to read the table (index_read) after every inserted now. And bulk insert optimization might disable indexes. bulk insert is already disabled in other cases when there are chances that the table will be read duing the bulk insert.
| | | | * | MDEV-27631 Assertion `global_status_var.global_memory_used == 0' failed in ↵Sergei Golubchik2023-01-203-20/+43
| | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | mysqld_exit plugin_vars_free_values() was walking plugin sysvars and thus did not free memory of plugin PLUGIN_VAR_NOSYSVAR vars. * change it to walk all plugin vars * add the pluginname_ prefix to NOSYSVARS var names too, so that plugin_vars_free_values() would be able to find their bookmarks
| | | | * | Merge branch 'codership-10.4-fix-sst_received' into bb-10.4-MDEV-30419Jan Lindström2023-01-203-7/+26
| | | | |\ \
| | | | | * | MDEV-29294 Assertion `functype() == ((Item_cond *) new_item)->functype()' ↵Oleg Smirnov2023-01-206-26/+268
| | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | failed in Item_cond::remove_eq_conds on SELECT Item_singlerow_subselect may be converted to Item_cond during optimization. So there is a possibility of constructing nested Item_cond_and or Item_cond_or which is not allowed (such conditions must be flattened). This commit checks if such kind of optimization has been applied and flattens the condition if needed
| | | | | * | MDEV-27653 long uniques don't work with unicode collationsAlexander Barkov2023-01-200-0/+0
| | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | There are no source code changes in this commit! This is an empty follow-up commit for 284ac6f2b73650f138064c97a96c8e1d8846550b to comment what was done, as the patch itself did not have change comments. Problems solved in this patch: 1. The function calc_hash_for_unique() erroneously takes into account the string length, so equal strings (in terms of the collation) with different lengths got different hash value. For example: - LATIN LETTER A - 1 byte - LATIN LETTER A WITH ACUTE - 2 bytes are equal in utf8_general_ci, but as their lengths are different, calc_hash_for_unique() returned different hash values. 2. calc_hash_for_unique() also erroneously used val_str() result to calculate hashes. This may not be correct for some data types, e.g. TIMESTAMP, as its string value depends on the session environment (e.g. @@time_zone). Change summary: Instead of doing Item::val_str(), we should always call Field::hash() of the underlying Field. It properly handles both cases (equal strings with different lengths, as well as tricky data types like TIMESTAMP). Detailed change description: Non-functional changes (make the code cleaner): - Adding a helper class Hasher, to pass hash parts nr1 and nr2 through function arguments easier. - Splitting virtual Field::hash() into non-virtual wrapper Field::hash() and virtual Field::hash_not_null(). This helps to get rid of duplicate code handling SQL NULL, as it was equal in all Field_xxx implementations. - Adding a new method THD::my_ok_with_recreate_info(). Actual fix changes (make new tables work properly): - Adding a virtual method Item::hash_not_null() This helps to handle hashes on full fields (Item_field) and hashes on prefix fields (Item_func_left(Item_field)) in a polymorphic way. Implementing overrides for Item_field and Item_func_left. - Rewriting Item_func_hash::val_int() to use Item::hash_not_null(), instead of the combination of val_str() and alc_hash_for_unique(). Backward compatibility changes (make old tables work in the new server): - Adding a new class Item_func_hash_mariadb_100403. Moving the old version of Item_func_hash::val_int() into Item_func_hash_mariadb_100403::val_int(). The old class Item_func_hash_mariadb_100403 is still needed, to open old tables before upgrade is done. - Adding TABLE_SHARE::old_long_hash_function() and handler::check_long_hash_compatibility() to test if a table is using an old hash function. - Adding a helper method TABLE_SHARE::make_long_hash_func() to instantiate either Item_func_hash_mariadb_100403 (for old not upgraded tables) or Item_func_hash (for new tables). Upgrade changes (make old tables upgrade in the new server properly): Upgrading an old table to a new hash can be done using either of these two statements: ALTER IGNORE TABLE t1 FORCE; REPAIR TABLE t1; !!! These statements find and filter out erreneous duplicates!!! The table after these statements will have less records if there were erroneous duplicates (such and A and A WITH ACUTE). The information about filtered out records is reported in both statements. - Adding a new class Recreate_info to return out information about copied and duplucate rows from these functions: - mysql_alter_table() - mysql_recreate_table() - admin_recreate_table() This helps to print a warning during REPAIR: MariaDB [test]> repair table mdev27653_100422_text; +----------------------------+--------+----------+------------------------------------+ | Table | Op | Msg_type | Msg_text | +----------------------------+--------+----------+------------------------------------+ | test.mdev27653_100422_text | repair | Warning | Number of rows changed from 2 to 1 | | test.mdev27653_100422_text | repair | status | OK | +----------------------------+--------+----------+------------------------------------+ 2 rows in set (0.018 sec)
| | | | | * | Correct assert_grep.inc params in galera gcache testsDaniele Sciascia2023-01-202-4/+4
| | | | | | |
| | | | | * | MDEV-26541 Make UBSAN builds work with spider again.Yuchen Pei2023-01-205-107/+215
| | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | When built with ubsan and trying to load the spider plugin, the hidden visibility of mysqld compiling flag causes ha_spider.so to be missing the symbol ha_partition. This commit fixes that, as well as some memcpy null pointer issues when built with ubsan. Signed-off-by: Yuchen Pei <yuchen.pei@mariadb.com>
| | | | | * | MDEV-27653 long uniques don't work with unicode collationsAlexander Barkov2023-01-2027-66/+668
| | | | | | |
| | | | | * | MDEV-30419 Fix unhandled exception thrown from wsrep-libTeemu Ollakka2023-01-193-7/+26
| | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | Updated wsrep-lib to version in which server_state wait_until_state() and sst_received() were changed to report errors via return codes instead of throwing exceptions. Added error handling accordingly. Tested manually that failure in sst_received() which was caused by server misconfiguration (unknown configuration variable in server configuration) does not cause crash due to uncaught exception.
| | | | * | | MDEV-29294 Assertion `functype() == ((Item_cond *) new_item)->functype()' ↵Oleg Smirnov2023-01-206-26/+268
| | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | failed in Item_cond::remove_eq_conds on SELECT Item_singlerow_subselect may be converted to Item_cond during optimization. So there is a possibility of constructing nested Item_cond_and or Item_cond_or which is not allowed (such conditions must be flattened). This commit checks if such kind of optimization has been applied and flattens the condition if needed
| | | | * | | MDEV-27653 long uniques don't work with unicode collationsAlexander Barkov2023-01-200-0/+0
| | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | There are no source code changes in this commit! This is an empty follow-up commit for 284ac6f2b73650f138064c97a96c8e1d8846550b to comment what was done, as the patch itself did not have change comments. Problems solved in this patch: 1. The function calc_hash_for_unique() erroneously takes into account the string length, so equal strings (in terms of the collation) with different lengths got different hash value. For example: - LATIN LETTER A - 1 byte - LATIN LETTER A WITH ACUTE - 2 bytes are equal in utf8_general_ci, but as their lengths are different, calc_hash_for_unique() returned different hash values. 2. calc_hash_for_unique() also erroneously used val_str() result to calculate hashes. This may not be correct for some data types, e.g. TIMESTAMP, as its string value depends on the session environment (e.g. @@time_zone). Change summary: Instead of doing Item::val_str(), we should always call Field::hash() of the underlying Field. It properly handles both cases (equal strings with different lengths, as well as tricky data types like TIMESTAMP). Detailed change description: Non-functional changes (make the code cleaner): - Adding a helper class Hasher, to pass hash parts nr1 and nr2 through function arguments easier. - Splitting virtual Field::hash() into non-virtual wrapper Field::hash() and virtual Field::hash_not_null(). This helps to get rid of duplicate code handling SQL NULL, as it was equal in all Field_xxx implementations. - Adding a new method THD::my_ok_with_recreate_info(). Actual fix changes (make new tables work properly): - Adding a virtual method Item::hash_not_null() This helps to handle hashes on full fields (Item_field) and hashes on prefix fields (Item_func_left(Item_field)) in a polymorphic way. Implementing overrides for Item_field and Item_func_left. - Rewriting Item_func_hash::val_int() to use Item::hash_not_null(), instead of the combination of val_str() and alc_hash_for_unique(). Backward compatibility changes (make old tables work in the new server): - Adding a new class Item_func_hash_mariadb_100403. Moving the old version of Item_func_hash::val_int() into Item_func_hash_mariadb_100403::val_int(). The old class Item_func_hash_mariadb_100403 is still needed, to open old tables before upgrade is done. - Adding TABLE_SHARE::old_long_hash_function() and handler::check_long_hash_compatibility() to test if a table is using an old hash function. - Adding a helper method TABLE_SHARE::make_long_hash_func() to instantiate either Item_func_hash_mariadb_100403 (for old not upgraded tables) or Item_func_hash (for new tables). Upgrade changes (make old tables upgrade in the new server properly): Upgrading an old table to a new hash can be done using either of these two statements: ALTER IGNORE TABLE t1 FORCE; REPAIR TABLE t1; !!! These statements find and filter out erreneous duplicates!!! The table after these statements will have less records if there were erroneous duplicates (such and A and A WITH ACUTE). The information about filtered out records is reported in both statements. - Adding a new class Recreate_info to return out information about copied and duplucate rows from these functions: - mysql_alter_table() - mysql_recreate_table() - admin_recreate_table() This helps to print a warning during REPAIR: MariaDB [test]> repair table mdev27653_100422_text; +----------------------------+--------+----------+------------------------------------+ | Table | Op | Msg_type | Msg_text | +----------------------------+--------+----------+------------------------------------+ | test.mdev27653_100422_text | repair | Warning | Number of rows changed from 2 to 1 | | test.mdev27653_100422_text | repair | status | OK | +----------------------------+--------+----------+------------------------------------+ 2 rows in set (0.018 sec)
| | | | * | | Correct assert_grep.inc params in galera gcache testsDaniele Sciascia2023-01-202-4/+4
| | | | | | |
| | | | * | | MDEV-26541 Make UBSAN builds work with spider again.Yuchen Pei2023-01-205-107/+215
| | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | When built with ubsan and trying to load the spider plugin, the hidden visibility of mysqld compiling flag causes ha_spider.so to be missing the symbol ha_partition. This commit fixes that, as well as some memcpy null pointer issues when built with ubsan. Signed-off-by: Yuchen Pei <yuchen.pei@mariadb.com>
| | | | * | | MDEV-27653 long uniques don't work with unicode collationsAlexander Barkov2023-01-1927-66/+668
| | | | |/ /
| | | * | | MDEV-30404: Inconsistent updates of PAGE_MAX_TRX_ID on ROW_FORMAT=COMPRESSED ↵Marko Mäkelä2023-01-261-6/+3
| | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | | pages page_copy_rec_list_start(): Do not update the PAGE_MAX_TRX_ID on the compressed copy of the page. The modification is supposed to be logged as part of page_zip_compress() or page_zip_reorganize(). If the page cannot be compressed (due to running out of space), then page_zip_decompress() must be able to roll back the changes. This fixes a regression that was introduced in commit 56f6dab1d0e5a464ea49c1e5efb0032a0f5cea3e (MDEV-21174).
| | | * | | MDEV-23855 fixup: Remove SRV_MASTER_CHECKPOINT_INTERVALMarko Mäkelä2023-01-251-12/+1
| | | | | |