[转帖]Use TiFlash

use,tiflash · 浏览次数 : 0

小编点评

## Generate Content with Simple Ordering Here's how to generate content with simple ordering: **1. Define Data Type:** * Use appropriate data type for each variable. * For example, use `decimal(7,4)` for a decimal number with 4 decimal places. **2. Define Ordering:** * Use the `ORDER BY` clause to specify the order of variables. * Order can be based on specific variable values or directly with keywords. **3. Define Grouping and Ordering:** * Use `GROUP BY` clause to group data based on specific variables. * Within the `GROUP BY` clause, use `ORDER BY` to order data within each group. **4. Use Ordering Clauses:** * Use the `ORDER BY` clause within the main query to define the order of results. * You can combine different ordering clauses using `ORDER BY` clauses. **5. Example:** ```sql -- Define data type decimal(7,4) price_decimal; -- Define order order by price_decimal DESC; -- Define grouping group by price_decimal; -- Define ordering within group order by price_decimal ASC; -- Use ordering clauses select price_decimal, order_by_price_decimal from table_name order by price_decimal DESC; ``` **Note:** * This approach assumes you have control over data type and variable names. * Modify the `ORDER BY` clauses as needed based on your specific requirements. * Remember to use appropriate data types for each variable. By following these steps, you can generate content with simple ordering and achieve desired results.

正文

https://docs.pingcap.com/tidb/v5.0/use-tiflash

After TiFlash is deployed, data replication does not automatically begin. You need to manually specify the tables to be replicated.

You can either use TiDB to read TiFlash replicas for medium-scale analytical processing, or use TiSpark to read TiFlash replicas for large-scale analytical processing, which is based on your own needs. See the following sections for details:

Create TiFlash replicas for tables

After TiFlash is connected to the TiKV cluster, data replication by default does not begin. You can send a DDL statement to TiDB through a MySQL client to create a TiFlash replica for a specific table:

ALTER TABLE table_name SET TIFLASH REPLICA count;

The parameter of the above command is described as follows:

  • count indicates the number of replicas. When the value is 0, the replica is deleted.

If you execute multiple DDL statements on the same table, only the last statement is ensured to take effect. In the following example, two DDL statements are executed on the table tpch50, but only the second statement (to delete the replica) takes effect.

Create two replicas for the table:

ALTER TABLE `tpch50`.`lineitem` SET TIFLASH REPLICA 2;

Delete the replica:

ALTER TABLE `tpch50`.`lineitem` SET TIFLASH REPLICA 0;

Notes:

  • If the table t is replicated to TiFlash through the above DDL statements, the table created using the following statement will also be automatically replicated to TiFlash:

    CREATE TABLE table_name like t;
  • For versions earlier than v4.0.6, if you create the TiFlash replica before using TiDB Lightning to import the data, the data import will fail. You must import data to the table before creating the TiFlash replica for the table.

  • If TiDB and TiDB Lightning are both v4.0.6 or later, no matter a table has TiFlash replica(s) or not, you can import data to that table using TiDB Lightning. Note that this might slow the TiDB Lightning procedure, which depends on the NIC bandwidth on the lightning host, the CPU and disk load of the TiFlash node, and the number of TiFlash replicas.

  • It is recommended that you do not replicate more than 1,000 tables because this lowers the PD scheduling performance. This limit will be removed in later versions.

Check replication progress

You can check the status of the TiFlash replicas of a specific table using the following statement. The table is specified using the WHERE clause. If you remove the WHERE clause, you will check the replica status of all tables.

SELECT * FROM information_schema.tiflash_replica WHERE TABLE_SCHEMA = '<db_name>' and TABLE_NAME = '<table_name>';

In the result of above statement:

  • AVAILABLE indicates whether the TiFlash replicas of this table are available or not. 1 means available and 0 means unavailable. Once the replicas become available, this status does not change. If you use DDL statements to modify the number of replicas, the replication status will be recalculated.
  • PROGRESS means the progress of the replication. The value is between 0.0 and 1.01 means at least one replica is replicated.

Speed up TiFlash replication

Before TiFlash replicas are added, each TiKV instance performs a full table scan and sends the scanned data to TiFlash as a "snapshot" to create replicas. By default, TiFlash replicas are added slowly with fewer resources usage in order to minimize the impact on the online service. If there are spare CPU and disk IO resources in your TiKV and TiFlash nodes, you can accelerate TiFlash replication by performing the following steps.

  1. Temporarily increase the snapshot write speed limit for each TiKV and TiFlash instance by adjusting the TiFlash Proxy and TiKV configuration. For example, when using TiUP to manage configurations, the configuration is as below:

    tikv: server.snap-max-write-bytes-per-sec: 300MiB # Default to 100MiB. tiflash-learner: raftstore.snap-handle-pool-size: 10 # Default to 2. Can be adjusted to >= node's CPU num * 0.6. raftstore.apply-low-priority-pool-size: 10 # Default to 1. Can be adjusted to >= node's CPU num * 0.6. server.snap-max-write-bytes-per-sec: 300MiB # Default to 100MiB.

    The configuration change takes effect after restarting the TiFlash and TiKV instances. The TiKV configuration can be also changed online by using the Dynamic Config SQL statement, which takes effect immediately without restarting TiKV instances:

    SET CONFIG tikv `server.snap-max-write-bytes-per-sec` = '300MiB';

    After adjusting the preceding configurations, you cannot observe the acceleration for now, as the replication speed is still restricted by the PD limit globally.

  2. Use PD Control to progressively ease the new replica speed limit.

    The default new replica speed limit is 30, which means, approximately 30 Regions add TiFlash replicas every minute. Executing the following command will adjust the limit to 60 for all TiFlash instances, which doubles the original speed:

    tiup ctl:v<CLUSTER_VERSION> pd -u http://<PD_ADDRESS>:2379 store limit all engine tiflash 60 add-peer

    In the preceding command, you need to replace <CLUSTER_VERSION> with the actual cluster version and <PD_ADDRESS>:2379 with the address of any PD node. For example:

    tiup ctl:v6.1.1 pd -u http://192.168.1.4:2379 store limit all engine tiflash 60 add-peer

    Within a few minutes, you will observe a significant increase in CPU and disk IO resource usage of the TiFlash nodes, and TiFlash should create replicas faster. At the same time, the TiKV nodes' CPU and disk IO resource usage increases as well.

    If the TiKV and TiFlash nodes still have spare resources at this point and the latency of your online service does not increase significantly, you can further ease the limit, for example, triple the original speed:

    tiup ctl:v<CLUSTER_VERSION> pd -u http://<PD_ADDRESS>:2379 store limit all engine tiflash 90 add-peer
  3. After the TiFlash replication is complete, revert to the default configuration to reduce the impact on online services.

    Execute the following PD Control command to restore the default new replica speed limit:

    tiup ctl:v<CLUSTER_VERSION> pd -u http://<PD_ADDRESS>:2379 store limit all engine tiflash 30 add-peer

    Comment out the changed configuration in TiUP to restore the default snapshot write speed limit:

    # tikv: # server.snap-max-write-bytes-per-sec: 300MiB # tiflash-learner: # raftstore.snap-handle-pool-size: 10 # raftstore.apply-low-priority-pool-size: 10 # server.snap-max-write-bytes-per-sec: 300MiB

Set available zones

When configuring replicas, if you need to distribute TiFlash replicas to multiple data centers for disaster recovery, you can configure available zones by following the steps below:

  1. Specify labels for TiFlash nodes in the cluster configuration file.

    tiflash_servers: - host: 172.16.5.81 config: flash.proxy.labels: zone=z1 - host: 172.16.5.82 config: flash.proxy.labels: zone=z1 - host: 172.16.5.85 config: flash.proxy.labels: zone=z2
  2. After starting a cluster, specify the labels when creating replicas.

    ALTER TABLE table_name SET TIFLASH REPLICA count LOCATION LABELS location_labels;

    For example:

    ALTER TABLE t SET TIFLASH REPLICA 2 LOCATION LABELS "zone";
  3. PD schedules the replicas based on the labels. In this example, PD respectively schedules two replicas of the table t to two available zones. You can use pd-ctl to view the scheduling.

    > tiup ctl:v<CLUSTER_VERSION> pd -u http://<PD_ADDRESS>:2379 store ... "address": "172.16.5.82:23913", "labels": [ { "key": "engine", "value": "tiflash"}, { "key": "zone", "value": "z1" } ], "region_count": 4, ... "address": "172.16.5.81:23913", "labels": [ { "key": "engine", "value": "tiflash"}, { "key": "zone", "value": "z1" } ], "region_count": 5, ... "address": "172.16.5.85:23913", "labels": [ { "key": "engine", "value": "tiflash"}, { "key": "zone", "value": "z2" } ], "region_count": 9, ...

For more information about scheduling replicas by using labels, see Schedule Replicas by Topology LabelsMultiple Data Centers in One City Deployment, and Three Data Centers in Two Cities Deployment.

Use TiDB to read TiFlash replicas

TiDB provides three ways to read TiFlash replicas. If you have added a TiFlash replica without any engine configuration, the CBO (cost-based optimization) mode is used by default.

Smart selection

For tables with TiFlash replicas, the TiDB optimizer automatically determines whether to use TiFlash replicas based on the cost estimation. You can use the desc or explain analyze statement to check whether or not a TiFlash replica is selected. For example:

desc select count(*) from test.t;
+--------------------------+---------+--------------+---------------+--------------------------------+ | id | estRows | task | access object | operator info | +--------------------------+---------+--------------+---------------+--------------------------------+ | StreamAgg_9 | 1.00 | root | | funcs:count(1)->Column#4 | | └─TableReader_17 | 1.00 | root | | data:TableFullScan_16 | | └─TableFullScan_16 | 1.00 | cop[tiflash] | table:t | keep order:false, stats:pseudo | +--------------------------+---------+--------------+---------------+--------------------------------+ 3 rows in set (0.00 sec)
explain analyze select count(*) from test.t;
+--------------------------+---------+---------+--------------+---------------+----------------------------------------------------------------------+--------------------------------+-----------+------+ | id | estRows | actRows | task | access object | execution info | operator info | memory | disk | +--------------------------+---------+---------+--------------+---------------+----------------------------------------------------------------------+--------------------------------+-----------+------+ | StreamAgg_9 | 1.00 | 1 | root | | time:83.8372ms, loops:2 | funcs:count(1)->Column#4 | 372 Bytes | N/A | | └─TableReader_17 | 1.00 | 1 | root | | time:83.7776ms, loops:2, rpc num: 1, rpc time:83.5701ms, proc keys:0 | data:TableFullScan_16 | 152 Bytes | N/A | | └─TableFullScan_16 | 1.00 | 1 | cop[tiflash] | table:t | time:43ms, loops:1 | keep order:false, stats:pseudo | N/A | N/A | +--------------------------+---------+---------+--------------+---------------+----------------------------------------------------------------------+--------------------------------+-----------+------+

cop[tiflash] means that the task will be sent to TiFlash for processing. If you have not selected a TiFlash replica, you can try to update the statistics using the analyze table statement, and then check the result using the explain analyze statement.

Note that if a table has only a single TiFlash replica and the related node cannot provide service, queries in the CBO mode will repeatedly retry. In this situation, you need to specify the engine or use the manual hint to read data from the TiKV replica.

Engine isolation

Engine isolation is to specify that all queries use a replica of the specified engine by configuring the corresponding variable. The optional engines are "tikv", "tidb" (indicates the internal memory table area of TiDB, which stores some TiDB system tables and cannot be actively used by users), and "tiflash", with the following two configuration levels:

  • TiDB instance-level, namely, INSTANCE level. Add the following configuration item in the TiDB configuration file:

    [isolation-read] engines = ["tikv", "tidb", "tiflash"]

    The INSTANCE-level default configuration is ["tikv", "tidb", "tiflash"].

  • SESSION level. Use the following statement to configure:

    set @@session.tidb_isolation_read_engines = "engine list separated by commas";

    or

    set SESSION tidb_isolation_read_engines = "engine list separated by commas";

    The default configuration of the SESSION level inherits from the configuration of the TiDB INSTANCE level.

The final engine configuration is the session-level configuration, that is, the session-level configuration overrides the instance-level configuration. For example, if you have configured "tikv" in the INSTANCE level and "tiflash" in the SESSION level, then the TiFlash replicas are read. If the final engine configuration is "tikv" and "tiflash", then the TiKV and TiFlash replicas are both read, and the optimizer automatically selects a better engine to execute.

 
Note

Because TiDB Dashboard and other components need to read some system tables stored in the TiDB memory table area, it is recommended to always add the "tidb" engine to the instance-level engine configuration.

If the queried table does not have a replica of the specified engine (for example, the engine is configured as "tiflash" but the table does not have a TiFlash replica), the query returns an error.

Manual hint

Manual hint can force TiDB to use specified replicas for specific table(s) on the premise of satisfying engine isolation. Here is an example of using the manual hint:

select /*+ read_from_storage(tiflash[table_name]) */ ... from table_name;

If you set an alias to a table in a query statement, you must use the alias in the statement that includes a hint for the hint to take effect. For example:

select /*+ read_from_storage(tiflash[alias_a,alias_b]) */ ... from table_name_1 as alias_a, table_name_2 as alias_b where alias_a.column_1 = alias_b.column_2;

In the above statements, tiflash[] prompts the optimizer to read the TiFlash replicas. You can also use tikv[] to prompt the optimizer to read the TiKV replicas as needed. For hint syntax details, refer to READ_FROM_STORAGE.

If the table specified by a hint does not have a replica of the specified engine, the hint is ignored and a warning is reported. In addition, a hint only takes effect on the premise of engine isolation. If the engine specified in a hint is not in the engine isolation list, the hint is also ignored and a warning is reported.

 
Note

The MySQL client of 5.7.7 or earlier versions clears optimizer hints by default. To use the hint syntax in these early versions, start the client with the --comments option, for example, mysql -h 127.0.0.1 -P 4000 -uroot --comments.

The relationship of smart selection, engine isolation, and manual hint

In the above three ways of reading TiFlash replicas, engine isolation specifies the overall range of available replicas of engines; within this range, manual hint provides statement-level and table-level engine selection that is more fine-grained; finally, CBO makes the decision and selects a replica of an engine based on cost estimation within the specified engine list.

 
Note

Before v4.0.3, the behavior of reading from TiFlash replica in a non-read-only SQL statement (for example, INSERT INTO ... SELECTSELECT ... FOR UPDATEUPDATE ...DELETE ...) is undefined. In v4.0.3 and later versions, internally TiDB ignores the TiFlash replica for a non-read-only SQL statement to guarantee the data correctness. That is, for smart selection, TiDB automatically chooses the non-TiFlash replica; for engine isolation that specifies TiFlash replica only, TiDB reports an error; and for manual hint, TiDB ignores the hint.

Use TiSpark to read TiFlash replicas

Currently, you can use TiSpark to read TiFlash replicas in a method similar to the engine isolation in TiDB. This method is to configure the spark.tispark.isolation_read_engines parameter. The parameter value defaults to tikv,tiflash, which means that TiDB reads data from TiFlash or from TiKV according to CBO's selection. If you set the parameter value to tiflash, it means that TiDB forcibly reads data from TiFlash.

Notes

When this parameter is set to tiflash, only the TiFlash replicas of all tables involved in the query are read and these tables must have TiFlash replicas; for tables that do not have TiFlash replicas, an error is reported. When this parameter is set to tikv, only the TiKV replica is read.

You can configure this parameter in one of the following ways:

  • Add the following item in the spark-defaults.conf file:

    spark.tispark.isolation_read_engines tiflash
  • Add --conf spark.tispark.isolation_read_engines=tiflash in the initialization command when initializing Spark shell or Thrift server.

  • Set spark.conf.set("spark.tispark.isolation_read_engines", "tiflash") in Spark shell in a real-time manner.

  • Set set spark.tispark.isolation_read_engines=tiflash in Thrift server after the server is connected via beeline.

Supported push-down calculations

TiFlash supports the push-down of the following operators:

  • TableScan: Reads data from tables.
  • Selection: Filters data.
  • HashAgg: Performs data aggregation based on the Hash Aggregation algorithm.
  • StreamAgg: Performs data aggregation based on the Stream Aggregation algorithm. SteamAgg only supports the aggregation without the GROUP BY condition.
  • TopN: Performs the TopN calculation.
  • Limit: Performs the limit calculation.
  • Project: Performs the projection calculation.
  • HashJoin: Performs the join calculation based on the Hash Join algorithm, but with the following conditions:
    • The operator can be pushed down only in the MPP mode.
    • The operator must have the join condition with the equivalent value.
    • The push-down of Full Outer Join is not supported.

In TiDB, operators are organized in a tree structure. For an operator to be pushed down to TiFlash, all of the following prerequisites must be met:

  • All of its child operators can be pushed down to TiFlash.
  • If an operator contains expressions (most of the operators contain expressions), all expressions of the operator can be pushed down to TiFlash.

Currently, TiFlash supports the following push-down expressions:

  • Mathematical functions: +, -, /, *, >=, <=, =, !=, <, >, round(int), round(double), abs, floor(int), ceil(int), ceiling(int)
  • Logical functions: and, or, not, case when, if, ifnull, isnull, in
  • Bitwise operations: bitand, bitor, bigneg, bitxor
  • String functions: substr, char_length, replace, concat, concat_ws, left, right
  • Date functions: date_format, timestampdiff, from_unixtime, unix_timestamp(int), unix_timestamp(decimal), str_to_date(date), str_to_date(datetime), datediff, year, month, day, extract(datetime)
  • JSON function: json_length
  • Conversion functions: cast(int as double), cast(int as decimal), cast(int as string), cast(int as time), cast(double as int), cast(double as decimal), cast(double as string), cast(double as time), cast(string as int), cast(string as double), cast(string as decimal), cast(string as time), cast(decimal as int), cast(decimal as string), cast(decimal as time), cast(time as int), cast(time as decimal), cast(time as string)
  • Aggregate functions: min, max, sum, count, avg, approx_count_distinct

In addition, expressions that contain the Time/Bit/Set/Enum/Geometry type cannot be pushed down to TiFlash.

If a query encounters unsupported push-down calculations, TiDB needs to complete the remaining calculations, which might greatly affect the TiFlash acceleration effect. The currently unsupported operators and expressions might be supported in future versions.

Use the MPP mode

TiFlash supports using the MPP mode to execute queries, which introduces cross-node data exchange (data shuffle process) into the computation. The MPP mode is enabled by default and can be disabled by setting the value of the global/session variable tidb_allow_mpp to 0 or OFF.

set @@session.tidb_allow_mpp=0

MPP mode supports these physical algorithms: Broadcast Hash Join, Shuffled Hash Join, and Shuffled Hash Aggregation. The optimizer automatically determines which algorithm to be used in a query. To check the specific query execution plan, you can execute the EXPLAIN statement. If the result of the EXPLAIN statement shows ExchangeSender and ExchangeReceiver operators, it indicates that the MPP mode has taken effect.

The following statement takes the table structure in the TPC-H test set as an example:

explain select count(*) from customer c join nation n on c.c_nationkey=n.n_nationkey; +------------------------------------------+------------+-------------------+---------------+----------------------------------------------------------------------------+ | id | estRows | task | access object | operator info | +------------------------------------------+------------+-------------------+---------------+----------------------------------------------------------------------------+ | HashAgg_23 | 1.00 | root | | funcs:count(Column#16)->Column#15 | | └─TableReader_25 | 1.00 | root | | data:ExchangeSender_24 | | └─ExchangeSender_24 | 1.00 | batchCop[tiflash] | | ExchangeType: PassThrough | | └─HashAgg_12 | 1.00 | batchCop[tiflash] | | funcs:count(1)->Column#16 | | └─HashJoin_17 | 3000000.00 | batchCop[tiflash] | | inner join, equal:[eq(tpch.nation.n_nationkey, tpch.customer.c_nationkey)] | | ├─ExchangeReceiver_21(Build) | 25.00 | batchCop[tiflash] | | | | │ └─ExchangeSender_20 | 25.00 | batchCop[tiflash] | | ExchangeType: Broadcast | | │ └─TableFullScan_18 | 25.00 | batchCop[tiflash] | table:n | keep order:false | | └─TableFullScan_22(Probe) | 3000000.00 | batchCop[tiflash] | table:c | keep order:false | +------------------------------------------+------------+-------------------+---------------+----------------------------------------------------------------------------+ 9 rows in set (0.00 sec)

In the example execution plan, the ExchangeReceiver and ExchangeSender operators are included. The execution plan indicates that after the nation table is read, the ExchangeSender operator broadcasts the table to each node, the HashJoin and HashAgg operations are performed on the nation table and the customer table, and then the results are returned to TiDB.

TiFlash provides the following two global/session variables to control whether to use Broadcast Hash Join:

  • tidb_broadcast_join_threshold_size: The unit of the value is bytes. If the table size (in the unit of bytes) is less than the value of the variable, the Broadcast Hash Join algorithm is used. Otherwise, the Shuffled Hash Join algorithm is used.
  • tidb_broadcast_join_threshold_count: The unit of the value is rows. If the objects of the join operation belong to a subquery, the optimizer cannot estimate the size of the subquery result set, so the size is determined by the number of rows in the result set. If the estimated number of rows in the subquery is less than the value of this variable, the Broadcast Hash Join algorithm is used. Otherwise, the Shuffled Hash Join algorithm is used.

Known issues of MPP

In the current version, TiFlash uses the start_ts of a query as the unique key of the query. In most cases, the start_ts of each query can uniquely identify a query, but in the following cases, different queries have the same start_ts:

  • All queries in the same transaction have the same start_ts.
  • When you use tidb_snapshot to read data at a specific historical time point, the same time point is manually specified.

When start_ts cannot uniquely represent the MPP query, if TiFlash detects that different queries have the same start_ts at a given time, TiFlash might report an error. Typical error cases are as follows:

  • When multiple queries with the same start_ts are sent to TiFlash at the same time, you might encounter the task has been registered error.
  • When multiple simple queries with LIMIT are executed continuously in the same transaction, once the LIMIT condition is met, TiDB sends a cancel request to TiFlash to cancel the query. This request also uses start_ts to identify the query to be canceled. If there are other queries with the same start_ts in TiFlash, these queries might be canceled by mistake. An example of this issue can be found in #43426.

This issue is fixed in TiDB v6.6.0. It is recommended to use the latest LTS version.

Notes

Currently, TiFlash does not support some features. These features might be incompatible with the native TiDB:

  • In the TiFlash computation layer:

    • Checking overflowed numerical values is not supported. For example, adding two maximum values of the BIGINT type 9223372036854775807 + 9223372036854775807. The expected behavior of this calculation in TiDB is to return the ERROR 1690 (22003): BIGINT value is out of range error. However, if this calculation is performed in TiFlash, an overflow value of -2 is returned without any error.

    • The window function is not supported.

    • Reading data from TiKV is not supported.

    • Currently, the sum function in TiFlash does not support the string-type argument. But TiDB cannot identify whether any string-type argument has been passed into the sum function during the compiling. Therefore, when you execute statements similar to select sum(string_col) from t, TiFlash returns the [FLASH:Coprocessor:Unimplemented] CastStringAsReal is not supported. error. To avoid such an error in this case, you need to modify this SQL statement to select sum(cast(string_col as double)) from t.

    • Currently, TiFlash's decimal division calculation is incompatible with that of TiDB. For example, when dividing decimal, TiFlash performs the calculation always using the type inferred from the compiling. However, TiDB performs this calculation using a type that is more precise than that inferred from the compiling. Therefore, some SQL statements involving the decimal division return different execution results when executed in TiDB + TiKV and in TiDB + TiFlash. For example:

      mysql> create table t (a decimal(3,0), b decimal(10, 0)); Query OK, 0 rows affected (0.07 sec) mysql> insert into t values (43, 1044774912); Query OK, 1 row affected (0.03 sec) mysql> alter table t set tiflash replica 1; Query OK, 0 rows affected (0.07 sec) mysql> set session tidb_isolation_read_engines='tikv'; Query OK, 0 rows affected (0.00 sec) mysql> select a/b, a/b + 0.0000000000001 from t where a/b; +--------+-----------------------+ | a/b | a/b + 0.0000000000001 | +--------+-----------------------+ | 0.0000 | 0.0000000410001 | +--------+-----------------------+ 1 row in set (0.00 sec) mysql> set session tidb_isolation_read_engines='tiflash'; Query OK, 0 rows affected (0.00 sec) mysql> select a/b, a/b + 0.0000000000001 from t where a/b; Empty set (0.01 sec)

      In the example above, a/b's inferred type from the compiling is Decimal(7,4) both in TiDB and in TiFlash. Constrained by Decimal(7,4)a/b's returned type should be 0.0000. In TiDB, a/b's runtime precision is higher than Decimal(7,4), so the original table data is not filtered by the where a/b condition. However, in TiFlash, the calculation of a/b uses Decimal(7,4) as the result type, so the original table data is filtered by the where a/b condition.

  • The MPP mode in TiFlash does not support the following features:

    • The partitioned table is not supported. For queries on the partitioned tables, the MPP mode is not chosen by default.
    • If the new_collations_enabled_on_first_bootstrap configuration item's value is true, the MPP mode does not support the string-type join key or the string column type in the group by aggregation. For these two query types, the MPP mode is not chosen by default.

与[转帖]Use TiFlash相似的内容:

[转帖]Use TiFlash

https://docs.pingcap.com/tidb/v5.0/use-tiflash 8 Contributors After TiFlash is deployed, data replication does not automatically begin. You need to ma

[转帖]使用 TiDB 读取 TiFlash

https://docs.pingcap.com/zh/tidb/stable/use-tidb-to-read-tiflash 本文档介绍如何使用 TiDB 读取 TiFlash 副本。 TiDB 提供三种读取 TiFlash 副本的方式。如果添加了 TiFlash 副本,而没有做任何 engin

[转帖]USE方法

https://www.jianshu.com/p/2d05e94e15b7 USE方法全称"Utilization Saturation and Errors Method",主要用于分析系统性能问题,可以指导用户快速识别资源瓶颈以及错误的方法。正如USE方法的名字所表示的含义,USE方法主要关注

[转帖]The USE Method

Boeing 707 Emergency Checklist(1969) The Utilization Saturation and Errors (USE) Method is a methodology for analyzing the performance of any system.

[转帖]MySQL 8.0: When to use utf8mb3 over utf8mb4?

https://dev.mysql.com/blog-archive/mysql-8-0-when-to-use-utf8mb3-over-utf8mb4/ MySQL 8.0: When to use utf8mb3 over utf8mb4? Posted on May 19, 2017 by 

[转帖]使用MAT命令行工具生成堆dump分析文件

https://www.cnblogs.com/hellxz/p/use_mat_linux_command_line_generate_reports.html 写作目标 Java程序运行过程中,难免会出现 OOM,往往是在 JVM 启动参数中添加出现 OOM 时输出堆 dump(又名:堆转储、堆

[转帖]没 K8s 用不了 Chaos Mesh?试试 Chaosd

https://cn.pingcap.com/blog/cannot-use-chaosmesh-without-k8s-then-try-chaosd Chaosd 是什么? 相信大家对 Chaos Mesh 已经比较了解了:支持多种类型的混沌实验,有 Dashboard web 界面直接管理实验

[转帖][译] 星巴克不使用两阶段提交(2004)

http://arthurchiao.art/blog/starbucks-do-not-use-two-phase-commit-zh/ 译者序 本文翻译自 2004 年的一篇文章: Starbucks Does Not Use Two-Phase Commit. 由于译者水平有限,本文不免存在遗

【转帖】What are segfault rip/rsp numbers and how to use them

https://stackoverflow.com/questions/1456899/what-are-segfault-rip-rsp-numbers-and-how-to-use-them 11 When my linux application crashes, it produces a

[转帖]坑:jmeter代理服务器录制脚本出现target controller is configured to "use recording Controller" but no such controller exists...

配置好代理服务器后,运行代理服务器 run 报错: target controller is configured to "use recording Controller" but no such controller exists,ensure you add a Recording Contr