Merge branch 'master' into atomic_metadata5

This commit is contained in:
alesapin 2020-06-19 11:09:03 +03:00
commit 67062f45be
23 changed files with 322 additions and 153 deletions

View File

@ -1,6 +1,6 @@
---
toc_priority: 31
toc_title: "\u61D2\u60F0"
toc_title: "延时引擎"
---
# 延时引擎Lazy {#lazy}

View File

@ -1,19 +1,19 @@
---
machine_translated: true
machine_translated_rev: 72537a2d527c63c07aa5d2361a8829f3895cf2bd
toc_priority: 34
toc_title: JDBC
toc_title: JDBC表引擎
---
# JDBC {#table-engine-jdbc}
允许ClickHouse通过以下方式连接到外部数据库 [JDBC](https://en.wikipedia.org/wiki/Java_Database_Connectivity).
允许CH通过 [JDBC](https://en.wikipedia.org/wiki/Java_Database_Connectivity) 连接到外部数据库。
要实现JDBC连接ClickHouse使用单独的程序 [ツ暗ェツ氾环催ツ団ツ法ツ人](https://github.com/alex-krash/clickhouse-jdbc-bridge) 这应该作为守护进程运行。
该引擎支持 [可为空](../../../sql-reference/data-types/nullable.md) 数据类型
要实现JDBC连接CH需要使用以后台进程运行的程序 [clickhouse-jdbc-bridge](https://github.com/alex-krash/clickhouse-jdbc-bridge)
## 创建表 {#creating-a-table}
该引擎支持 [Nullable](../../../sql-reference/data-types/nullable.md) 数据类型。
## 建表 {#creating-a-table}
``` sql
CREATE TABLE [IF NOT EXISTS] [db.]table_name
@ -23,20 +23,22 @@ CREATE TABLE [IF NOT EXISTS] [db.]table_name
ENGINE = JDBC(dbms_uri, external_database, external_table)
```
**发动机参数**
**引擎参数**
- `dbms_uri`URI of an external DBMS.
- `dbms_uri`外部DBMS的uri.
格式: `jdbc:<driver_name>://<host_name>:<port>/?user=<username>&password=<password>`.
Mysql的示例: `jdbc:mysql://localhost:3306/?user=root&password=root`.
MySQL示例: `jdbc:mysql://localhost:3306/?user=root&password=root`.
- `external_database`Database in an external DBMS.
- `external_database`外部DBMS的数据库名.
- `external_table`Name of the table in `external_database`.
- `external_table``external_database`中的外部表名.
## 用法示例 {#usage-example}
通过直接与它的控制台客户端连接在MySQL服务器中创建一个表:
通过mysql控制台客户端来创建表
Creating a table in MySQL server by connecting directly with its console client:
``` text
mysql> CREATE TABLE `test`.`test` (
@ -59,7 +61,7 @@ mysql> select * from test;
1 row in set (0,00 sec)
```
在ClickHouse服务器中创建表并从中选择数据:
在CH服务端创建表并从中查询数据
``` sql
CREATE TABLE jdbc_table
@ -83,8 +85,8 @@ FROM jdbc_table
└────────┴──────────────┴───────┴────────────────┘
```
## 另请参阅 {#see-also}
## 参见 {#see-also}
- [JDBC表函数](../../../sql-reference/table-functions/jdbc.md).
[原始文](https://clickhouse.tech/docs/en/operations/table_engines/jdbc/) <!--hide-->
[原始文](https://clickhouse.tech/docs/en/operations/table_engines/jdbc/) <!--hide-->

View File

@ -1,17 +1,16 @@
---
machine_translated: true
machine_translated_rev: 72537a2d527c63c07aa5d2361a8829f3895cf2bd
toc_priority: 38
toc_title: GraphiteMergeTree
---
# GraphiteMergeTree {#graphitemergetree}
此引擎专为细化和聚合/平均rollup) [石墨](http://graphite.readthedocs.io/en/latest/index.html) 戴达 对于想要使用ClickHouse作为Graphite的数据存储的开发人员来说这可能会有所帮助
该引擎用来对 [Graphite](http://graphite.readthedocs.io/en/latest/index.html)数据进行瘦身及汇总。对于想使用CH来存储Graphite数据的开发者来说可能有用
您可以使用任何ClickHouse表引擎来存储石墨数据如果你不需要汇总但如果你需要一个汇总使用 `GraphiteMergeTree`. 该引擎减少了存储量并提高了Graphite查询的效率。
引擎继承从属性 [MergeTree](mergetree.md).
如果不需要对Graphite数据做汇总那么可以使用任意的CH表引擎但若需要那就采用 `GraphiteMergeTree` 引擎。它能减少存储空间同时能提高Graphite数据的查询效率。
该引擎继承自 [MergeTree](../../../engines/table-engines/mergetree-family/mergetree.md).
## 创建表 {#creating-table}
@ -30,36 +29,32 @@ CREATE TABLE [IF NOT EXISTS] [db.]table_name [ON CLUSTER cluster]
[SETTINGS name=value, ...]
```
请参阅的详细说明 [CREATE TABLE](../../../sql-reference/statements/create.md#create-table-query) 查询。
建表语句的详细说明请参见 [创建表](../../../sql-reference/statements/create.md#create-table-query)
Graphite数据的表应具有以下数据的列:
含有Graphite数据集的表应该包含以下的数据列
- 指标名称(Graphite sensor),数据类型:`String`
- 指标的时间度量,数据类型: `DateTime`
- 指标的值,数据类型:任意数值类型
- 指标的版本号,数据类型: 任意数值类型
- 公制名称(石墨传感器)。 数据类型: `String`.
CH以最大的版本号保存行记录若版本号相同保留最后写入的数据。
- 测量度量的时间。 数据类型: `DateTime`.
以上列必须设置在汇总参数配置中。
- 度量值。 数据类型:任何数字。
- 指标的版本。 数据类型:任何数字。
**GraphiteMergeTree 参数**
- `config_section` - 配置文件中标识汇总规则的节点名称
如果版本相同ClickHouse会保存版本最高或最后写入的行。 其他行在数据部分合并期间被删除。
**建表语句**
应在汇总配置中设置这些列的名称。
**GraphiteMergeTree参数**
- `config_section` — Name of the section in the configuration file, where are the rules of rollup set.
**查询子句**
当创建一个 `GraphiteMergeTree` 表,相同 [条款](mergetree.md#table_engine-mergetree-creating-a-table) 是必需的,因为当创建 `MergeTree` 桌子
在创建 `GraphiteMergeTree` 表时,需要采用和 [clauses](../../../engines/table-engines/mergetree-family/mergetree.md#table_engine-mergetree-creating-a-table) 相同的语句,就像创建 `MergeTree` 一样。
<details markdown="1">
<summary>不推荐使用的创建表的方法</summary>
<summary>已废弃的建表语句</summary>
!!! attention "注意"
不要在新项目中使用此方法,如果可能的话,请将旧项目切换到上述方法
!!! 注意 "Attention"
请不要在新项目中使用;如有可能,请将旧的项目按上述的方法进行替换。
``` sql
CREATE TABLE [IF NOT EXISTS] [db.]table_name [ON CLUSTER cluster]
@ -73,31 +68,30 @@ CREATE TABLE [IF NOT EXISTS] [db.]table_name [ON CLUSTER cluster]
) ENGINE [=] GraphiteMergeTree(date-column [, sampling_expression], (primary, key), index_granularity, config_section)
```
所有参数除外 `config_section` 具有相同的含义 `MergeTree`.
除了`config_section`,其它所有参数和`MergeTree`的相应参数一样.
- `config_section` Name of the section in the configuration file, where are the rules of rollup set.
- `config_section`配置文件中设置汇总规则的节点
</details>
## 汇总配置 {#rollup-configuration}
## 汇总配置的参数 {#rollup-configuration}
汇总的配置参数由服务器配置的 [graphite\_rollup](../../../operations/server-configuration-parameters/settings.md#server_configuration_parameters-graphite) 参数定义。参数名称可以是任意的。允许为多个不同表创建多组配置并使用。
汇总的设置由 [graphite\_rollup](../../../operations/server-configuration-parameters/settings.md#server_configuration_parameters-graphite) 服务器配置中的参数。 参数的名称可以是any。 您可以创建多个配置并将它们用于不同的表。
汇总配置结构:
汇总配置的结构如下:
所需的列
模式Patterns
required-columns
patterns
### 必填列 {#required-columns}
### 所需的列 {#required-columns}
- `path_column_name` — 保存指标名称的列名 (Graphite sensor). 默认值: `Path`.
- `time_column_name` — 保存指标时间度量的列名. Default value: `Time`.
- `value_column_name` — The name of the column storing the value of the metric at the time set in `time_column_name`.默认值: `Value`.
- `version_column_name` - 保存指标的版本号列. 默认值: `Timestamp`.
- `path_column_name` — The name of the column storing the metric name (Graphite sensor). Default value: `Path`.
- `time_column_name` — The name of the column storing the time of measuring the metric. Default value: `Time`.
- `value_column_name` — The name of the column storing the value of the metric at the time set in `time_column_name`. 默认值: `Value`.
- `version_column_name` — The name of the column storing the version of the metric. Default value: `Timestamp`.
### 模式 {#patterns}
的结构 `patterns` 科:
### 模式Patterns {#patterns}
`patterns` 的结构:
``` text
pattern
@ -120,21 +114,20 @@ default
...
```
!!! warning "注意"
模式必须严格排序:
!!! 注意 "Attention"
模式必须严格按顺序配置:
1. 不含`function` or `retention`的Patterns
1. 同时含有`function` and `retention`的Patterns
1. `default`的Patterns.
1. Patterns without `function` or `retention`.
1. Patterns with both `function` and `retention`.
1. Pattern `default`.
CH在处理行记录时会检查 `pattern`节点的规则。每个 `pattern`(含`default`)节点可以包含 `function` 用于聚合操作,或`retention`参数,或者两者都有。如果指标名称和 `regexp`相匹配,相应 `pattern`的规则会生效;否则,使用 `default` 节点的规则。
在处理行时ClickHouse会检查以下内容中的规则 `pattern` 部分。 每个 `pattern` (包括 `default`)部分可以包含 `function` 聚合参数, `retention` 参数或两者兼而有之。 如果指标名称匹配 `regexp`,从规则 `pattern` 部分sections节的应用;否则,从规则 `default` 部分被使用。
`pattern``default` 节点的字段设置:
字段为 `pattern``default` 科:
- `regexp` A pattern for the metric name.
- `age` The minimum age of the data in seconds.
- `precision` How precisely to define the age of the data in seconds. Should be a divisor for 86400 (seconds in a day).
- `function` The name of the aggregating function to apply to data whose age falls within the range `[age, age + precision]`.
- `regexp` 指标名的pattern.
- `age` 数据的最小存活时间(按秒算).
- `precision` 按秒来衡量数据存活时间时的精确程度. 必须能被86400整除 (一天的秒数).
- `function` 对于存活时间在 `[age, age + precision]`之内的数据,需要使用的聚合函数
### 配置示例 {#configuration-example}
@ -171,4 +164,4 @@ default
</graphite_rollup>
```
[原始文](https://clickhouse.tech/docs/en/operations/table_engines/graphitemergetree/) <!--hide-->
[原始文](https://clickhouse.tech/docs/en/operations/table_engines/graphitemergetree/) <!--hide-->

View File

@ -1,35 +1,31 @@
---
machine_translated: true
machine_translated_rev: 72537a2d527c63c07aa5d2361a8829f3895cf2bd
toc_priority: 46
toc_title: GenerateRandom
toc_title: 随机数生成
---
# Generaterandom {#table_engines-generate}
# 随机数生成表引擎 {#table_engines-generate}
GenerateRandom表引擎为给定的表架构生成随机数据。
随机数生成表引擎为指定的表模式生成随机数
使用示例:
- 测试时生成可复写的大表
- 为复杂测试生成随机输入
- 在测试中使用填充可重复的大表。
- 为模糊测试生成随机输入。
## 在ClickHouse服务器中的使用 {#usage-in-clickhouse-server}
## CH服务端的用法 {#usage-in-clickhouse-server}
``` sql
ENGINE = GenerateRandom(random_seed, max_string_length, max_array_length)
```
`max_array_length``max_string_length` 参数指定所有的最大长度
数组列和字符串相应地在生成的数据中。
生成数据时,通过`max_array_length` 设置array列的最大长度 `max_string_length`设置string数据的最大长度
生成表引擎仅支持 `SELECT` 查询。
该引擎仅支持 `SELECT` 查询语句.
它支持所有 [数据类型](../../../sql-reference/data-types/index.md) 可以存储在一个表中,除了 `LowCardinality``AggregateFunction`.
该引擎支持能在表中存储的所有数据类型 [DataTypes](../../../sql-reference/data-types/index.md) ,除了 `LowCardinality``AggregateFunction`.
**示例:**
## 示例 {#example}
**1.** 设置 `generate_engine_table` 表:
**1.** 设置 `generate_engine_table` 引擎表:
``` sql
CREATE TABLE generate_engine_table (name String, value UInt32) ENGINE = GenerateRandom(1, 5, 3)
@ -49,13 +45,13 @@ SELECT * FROM generate_engine_table LIMIT 3
└──────┴────────────┘
```
## 实细节 {#details-of-implementation}
## 实细节 {#details-of-implementation}
- 不支持:
- 以下特性不支持:
- `ALTER`
- `SELECT ... SAMPLE`
- `INSERT`
- 指数
- 复制
- Indices
- Replication
[原始文](https://clickhouse.tech/docs/en/operations/table_engines/generate/) <!--hide-->
[原始文](https://clickhouse.tech/docs/en/operations/table_engines/generate/) <!--hide-->

View File

@ -27,7 +27,7 @@ namespace ErrorCodes
DataTypePtr DataTypeFactory::get(const String & full_name) const
{
ParserIdentifierWithOptionalParameters parser;
ParserDataType parser;
ASTPtr ast = parseQuery(parser, full_name.data(), full_name.data() + full_name.size(), "data type", 0, DBMS_DEFAULT_MAX_PARSER_DEPTH);
return get(ast);
}

View File

@ -379,14 +379,14 @@ bool DataTypeString::equals(const IDataType & rhs) const
static DataTypePtr create(const ASTPtr & arguments)
{
if (arguments)
if (arguments && !arguments->children.empty())
{
if (arguments->children.size() > 1)
throw Exception("String data type family mustnt have more than one argument - size in characters", ErrorCodes::NUMBER_OF_ARGUMENTS_DOESNT_MATCH);
throw Exception("String data type family mustn't have more than one argument - size in characters", ErrorCodes::NUMBER_OF_ARGUMENTS_DOESNT_MATCH);
const auto * argument = arguments->children[0]->as<ASTLiteral>();
if (!argument || argument->value.getType() != Field::Types::UInt64 || argument->value.get<UInt64>() == 0)
throw Exception("FixedString data type family may have only a number (positive integer) as its argument", ErrorCodes::UNEXPECTED_AST_STRUCTURE);
throw Exception("String data type family may have only a number (positive integer) as its argument", ErrorCodes::UNEXPECTED_AST_STRUCTURE);
}
return std::make_shared<DataTypeString>();
@ -415,5 +415,19 @@ void registerDataTypeString(DataTypeFactory & factory)
factory.registerAlias("MEDIUMBLOB", "String", DataTypeFactory::CaseInsensitive);
factory.registerAlias("LONGBLOB", "String", DataTypeFactory::CaseInsensitive);
factory.registerAlias("BYTEA", "String", DataTypeFactory::CaseInsensitive); /// PostgreSQL
factory.registerAlias("CHARACTER LARGE OBJECT", "String", DataTypeFactory::CaseInsensitive);
factory.registerAlias("CHARACTER VARYING", "String", DataTypeFactory::CaseInsensitive);
factory.registerAlias("CHAR LARGE OBJECT", "String", DataTypeFactory::CaseInsensitive);
factory.registerAlias("CHAR VARYING", "String", DataTypeFactory::CaseInsensitive);
factory.registerAlias("NATIONAL CHAR", "String", DataTypeFactory::CaseInsensitive);
factory.registerAlias("NATIONAL CHARACTER", "String", DataTypeFactory::CaseInsensitive);
factory.registerAlias("NATIONAL CHARACTER LARGE OBJECT", "String", DataTypeFactory::CaseInsensitive);
factory.registerAlias("NATIONAL CHARACTER VARYING", "String", DataTypeFactory::CaseInsensitive);
factory.registerAlias("NATIONAL CHAR VARYING", "String", DataTypeFactory::CaseInsensitive);
factory.registerAlias("NCHAR VARYING", "String", DataTypeFactory::CaseInsensitive);
factory.registerAlias("NCHAR LARGE OBJECT", "String", DataTypeFactory::CaseInsensitive);
factory.registerAlias("BINARY LARGE OBJECT", "String", DataTypeFactory::CaseInsensitive);
factory.registerAlias("BINARY VARYING", "String", DataTypeFactory::CaseInsensitive);
}
}

View File

@ -63,6 +63,8 @@ void registerDataTypeNumbers(DataTypeFactory & factory)
factory.registerAlias("REAL", "Float32", DataTypeFactory::CaseInsensitive);
factory.registerAlias("SINGLE", "Float32", DataTypeFactory::CaseInsensitive); /// MS Access
factory.registerAlias("DOUBLE", "Float64", DataTypeFactory::CaseInsensitive);
factory.registerAlias("DOUBLE PRECISION", "Float64", DataTypeFactory::CaseInsensitive);
}
}

View File

@ -18,8 +18,8 @@ namespace ErrorCodes
namespace
{
constexpr const char * any = "any";
constexpr const char * anyLast = "anyLast";
constexpr auto * any = "any";
constexpr auto * anyLast = "anyLast";
}
ASTPtr * getExactChild(const ASTPtr & ast, const size_t ind)
@ -30,12 +30,12 @@ ASTPtr * getExactChild(const ASTPtr & ast, const size_t ind)
}
///recursive searching of identifiers
void changeAllIdentifiers(ASTPtr & ast, size_t ind, std::string& mode)
void changeAllIdentifiers(ASTPtr & ast, size_t ind, const std::string & name)
{
const char * name = mode.c_str();
ASTPtr * exact_child = getExactChild(ast, ind);
if (!exact_child)
return;
if ((*exact_child)->as<ASTIdentifier>())
{
///put new any
@ -43,14 +43,15 @@ void changeAllIdentifiers(ASTPtr & ast, size_t ind, std::string& mode)
*exact_child = makeASTFunction(name);
(*exact_child)->as<ASTFunction>()->arguments->children.push_back(old_ast);
}
else if ((*exact_child)->as<ASTFunction>() &&
!AggregateFunctionFactory::instance().isAggregateFunctionName((*exact_child)->as<ASTFunction>()->name))
else if ((*exact_child)->as<ASTFunction>())
{
if (AggregateFunctionFactory::instance().isAggregateFunctionName((*exact_child)->as<ASTFunction>()->name))
throw Exception("Aggregate function " + (*exact_child)->as<ASTFunction>()->name +
" is found inside aggregate function " + name + " in query", ErrorCodes::ILLEGAL_AGGREGATION);
for (size_t i = 0; i < (*exact_child)->as<ASTFunction>()->arguments->children.size(); i++)
changeAllIdentifiers(*exact_child, i, mode);
else if ((*exact_child)->as<ASTFunction>() &&
AggregateFunctionFactory::instance().isAggregateFunctionName((*exact_child)->as<ASTFunction>()->name))
throw Exception("Aggregate function " + (*exact_child)->as<ASTFunction>()->name +
" is found inside aggregate function " + name + " in query", ErrorCodes::ILLEGAL_AGGREGATION);
changeAllIdentifiers(*exact_child, i, name);
}
}
@ -62,18 +63,20 @@ void AnyInputMatcher::visit(ASTPtr & current_ast, Data data)
return;
auto * function_node = current_ast->as<ASTFunction>();
if (function_node && (function_node->name == any || function_node->name == anyLast)
&& !function_node->arguments->children.empty() && function_node->arguments->children[0] &&
function_node->arguments->children[0]->as<ASTFunction>())
if (!function_node || function_node->arguments->children.empty())
return;
const auto & function_argument = function_node->arguments->children[0];
if ((function_node->name == any || function_node->name == anyLast)
&& function_argument && function_argument->as<ASTFunction>())
{
std::string mode = function_node->name;
auto name = function_node->name;
///cut any or anyLast
if (function_node->arguments->children[0]->as<ASTFunction>() &&
!function_node->arguments->children[0]->as<ASTFunction>()->arguments->children.empty())
if (!function_argument->as<ASTFunction>()->arguments->children.empty())
{
current_ast = (function_node->arguments->children[0])->clone();
current_ast = function_argument->clone();
for (size_t i = 0; i < current_ast->as<ASTFunction>()->arguments->children.size(); ++i)
changeAllIdentifiers(current_ast, i, mode);
changeAllIdentifiers(current_ast, i, name);
}
}
}

View File

@ -41,29 +41,31 @@ bool onlyConstsInside(const ASTFunction * func_node)
bool inappropriateNameInside(const ASTFunction * func_node, const char * inter_func_name)
{
return (func_node->arguments->children[0]->as<ASTFunction>() &&
inter_func_name != func_node->arguments->children[0]->as<ASTFunction>()->name) ||
strcmp(inter_func_name, func_node->arguments->children[0]->as<ASTFunction>()->name.c_str()) != 0) ||
(func_node->arguments->children.size() == 2 &&
func_node->arguments->children[1]->as<ASTFunction>() &&
inter_func_name != func_node->arguments->children[1]->as<ASTFunction>()->name);
strcmp(inter_func_name, func_node->arguments->children[1]->as<ASTFunction>()->name.c_str()) != 0);
}
bool isInappropriate(const ASTPtr & node, const char * inter_func_name)
{
return !node->as<ASTFunction>() || inter_func_name != node->as<ASTFunction>()->name;
return !node->as<ASTFunction>() || (strcmp(inter_func_name, node->as<ASTFunction>()->name.c_str()) != 0);
}
ASTFunction * getInternalFunction(const ASTFunction * f_n)
{
const auto * function_args = f_n->arguments->as<ASTExpressionList>();
if (!function_args || function_args->children.size() != 1)
throw Exception("Wrong number of arguments for function" + f_n->name + "(" + toString(function_args->children.size()) + " instead of 1)",
throw Exception("Wrong number of arguments for function " + f_n->name + "(" + toString(function_args->children.size()) + " instead of 1)",
ErrorCodes::NUMBER_OF_ARGUMENTS_DOESNT_MATCH);
return f_n->arguments->children[0]->as<ASTFunction>();
}
ASTFunction * treeFiller(ASTFunction * old_tree, const ASTs & nodes_array, size_t size, const char * name)
ASTFunction * treeFiller(ASTFunction * old_tree, const ASTs & nodes_array, size_t size, const char * name, bool flag)
{
if (flag)
--size;
for (size_t i = 0; i < size; ++i)
{
old_tree->arguments->children = {};
@ -94,19 +96,23 @@ std::pair<ASTs, ASTs> tryGetConst(const char * name, const ASTs & arguments)
not_const.push_back(arg);
}
if ((name == plus || name == mul) && const_num.size() + not_const.size() != 2)
{
if ((strcmp(name, plus) == 0 || strcmp(name, mul) == 0) && const_num.size() + not_const.size() != 2)
throw Exception("Wrong number of arguments for function 'plus' or 'multiply' (" + toString(const_num.size() + not_const.size()) + " instead of 2)",
ErrorCodes::NUMBER_OF_ARGUMENTS_DOESNT_MATCH);
}
return {const_num, not_const};
}
std::pair<ASTs, ASTs> findAllConsts(const ASTFunction * func_node, const char * inter_func_name)
{
if (!func_node->arguments)
return {};
if (func_node->arguments->children.empty())
{
if (strcmp(func_node->name.c_str(), plus) == 0 || strcmp(func_node->name.c_str(), mul) == 0)
throw Exception("Wrong number of arguments for function" + func_node->name + "(0 instead of 2)",
ErrorCodes::NUMBER_OF_ARGUMENTS_DOESNT_MATCH);
else
return {};
}
if (onlyConstsInside(func_node))
return tryGetConst(func_node->name.c_str(), func_node->arguments->children);
@ -139,7 +145,7 @@ std::pair<ASTs, ASTs> findAllConsts(const ASTFunction * func_node, const char *
std::pair<ASTs, ASTs> fl = tryGetConst(func_node->name.c_str(), func_node->arguments->children);
ASTs first_lvl_consts = fl.first;
ASTs first_lvl_not_consts = fl.second;
if (!first_lvl_not_consts[0]->as<ASTFunction>())
if (first_lvl_not_consts.empty() || !first_lvl_not_consts[0]->as<ASTFunction>())
return {first_lvl_consts, first_lvl_not_consts};
std::pair<ASTs, ASTs> ans = findAllConsts(first_lvl_not_consts[0]->as<ASTFunction>(), inter_func_name);
@ -176,17 +182,21 @@ void buildTree(ASTFunction * cur_node, const char * func_name, const char * intr
ASTs cons_val = tree_comp.first;
ASTs non_cons = tree_comp.second;
bool not_const_empty = non_cons.empty();
cur_node->name = intro_func;
cur_node = treeFiller(cur_node, cons_val, cons_val.size(), intro_func);
cur_node = treeFiller(cur_node, cons_val, cons_val.size(), intro_func, not_const_empty);
cur_node->name = func_name;
if (non_cons.size() == 1)
if (non_cons.empty())
cur_node->arguments->children.push_back(cons_val[cons_val.size() - 1]);
else if (non_cons.size() == 1)
cur_node->arguments->children.push_back(non_cons[0]);
else
{
cur_node->arguments->children.push_back(makeASTFunction(intro_func));
cur_node = cur_node->arguments->children[0]->as<ASTFunction>();
cur_node = treeFiller(cur_node, non_cons, non_cons.size() - 2, intro_func);
cur_node = treeFiller(cur_node, non_cons, non_cons.size() - 2, intro_func, not_const_empty);
cur_node->arguments->children = {non_cons[non_cons.size() - 2], non_cons[non_cons.size() - 1]};
}
}

View File

@ -352,7 +352,7 @@ bool ParserCastExpression::parseImpl(Pos & pos, ASTPtr & node, Expected & expect
&& ParserToken(TokenType::OpeningRoundBracket).ignore(pos, expected)
&& ParserExpression().parse(pos, expr_node, expected)
&& ParserKeyword("AS").ignore(pos, expected)
&& ParserIdentifierWithOptionalParameters().parse(pos, type_node, expected)
&& ParserDataType().parse(pos, type_node, expected)
&& ParserToken(TokenType::ClosingRoundBracket).ignore(pos, expected))
{
/// Convert to canonical representation in functional form: CAST(expr, 'type')
@ -1233,7 +1233,7 @@ bool ParserSubstitution::parseImpl(Pos & pos, ASTPtr & node, Expected & expected
++pos;
auto old_pos = pos;
ParserIdentifierWithOptionalParameters type_parser;
ParserDataType type_parser;
if (!type_parser.ignore(pos, expected))
{
expected.add(pos, "substitution type");

View File

@ -48,13 +48,6 @@ protected:
};
class ParserBareWord : public IParserBase
{
protected:
const char * getName() const override { return "bare word"; }
bool parseImpl(Pos & pos, ASTPtr & node, Expected & expected) override;
};
/** An identifier, possibly containing a dot, for example, x_yz123 or `something special` or Hits.EventTime,
* possibly with UUID clause like `db name`.`table name` UUID 'xxxxxxxx-xxxx-xxxx-xxxx-xxxxxxxxxxxx'
*/
@ -363,7 +356,7 @@ protected:
bool brackets_can_be_omitted;
};
/** Data type or table engine, possibly with parameters. For example, UInt8 or see examples from ParserIdentifierWithParameters
/** Table engine, possibly with parameters. See examples from ParserIdentifierWithParameters
* Parse result is ASTFunction, with or without arguments.
*/
class ParserIdentifierWithOptionalParameters : public IParserBase

View File

@ -53,12 +53,7 @@ bool ParserNestedTable::parseImpl(Pos & pos, ASTPtr & node, Expected & expected)
bool ParserIdentifierWithParameters::parseImpl(Pos & pos, ASTPtr & node, Expected & expected)
{
ParserFunction function_or_array;
if (function_or_array.parse(pos, node, expected))
return true;
ParserNestedTable nested;
return nested.parse(pos, node, expected);
return ParserFunction().parse(pos, node, expected);
}
bool ParserNameTypePairList::parseImpl(Pos & pos, ASTPtr & node, Expected & expected)
@ -85,7 +80,7 @@ bool ParserIndexDeclaration::parseImpl(Pos & pos, ASTPtr & node, Expected & expe
ParserKeyword s_granularity("GRANULARITY");
ParserIdentifier name_p;
ParserIdentifierWithOptionalParameters ident_with_optional_params_p;
ParserDataType data_type_p;
ParserExpression expression_p;
ParserUnsignedInteger granularity_p;
@ -103,7 +98,7 @@ bool ParserIndexDeclaration::parseImpl(Pos & pos, ASTPtr & node, Expected & expe
if (!s_type.ignore(pos, expected))
return false;
if (!ident_with_optional_params_p.parse(pos, type, expected))
if (!data_type_p.parse(pos, type, expected))
return false;
if (!s_granularity.ignore(pos, expected))

View File

@ -8,6 +8,7 @@
#include <Parsers/ASTIdentifier.h>
#include <Parsers/ASTLiteral.h>
#include <Parsers/CommonParsers.h>
#include <Parsers/ParserDataType.h>
#include <Poco/String.h>
@ -24,10 +25,9 @@ protected:
};
/** Parametric type or Storage. For example:
* FixedString(10) or
* Partitioned(Log, ChunkID) or
* Nested(UInt32 CounterID, FixedString(2) UserAgentMajor)
/** Storage engine or Codec. For example:
* Memory()
* ReplicatedMergeTree('/path', 'replica')
* Result of parsing - ASTFunction with or without parameters.
*/
class ParserIdentifierWithParameters : public IParserBase
@ -47,14 +47,12 @@ protected:
/** The name and type are separated by a space. For example, URL String. */
using ParserNameTypePair = IParserNameTypePair<ParserIdentifier>;
/** Name and type separated by a space. The name can contain a dot. For example, Hits.URL String. */
using ParserCompoundNameTypePair = IParserNameTypePair<ParserCompoundIdentifier>;
template <typename NameParser>
bool IParserNameTypePair<NameParser>::parseImpl(Pos & pos, ASTPtr & node, Expected & expected)
{
NameParser name_parser;
ParserIdentifierWithOptionalParameters type_parser;
ParserDataType type_parser;
ASTPtr name, type;
if (name_parser.parse(pos, name, expected)
@ -115,7 +113,7 @@ template <typename NameParser>
bool IParserColumnDeclaration<NameParser>::parseImpl(Pos & pos, ASTPtr & node, Expected & expected)
{
NameParser name_parser;
ParserIdentifierWithOptionalParameters type_parser;
ParserDataType type_parser;
ParserKeyword s_default{"DEFAULT"};
ParserKeyword s_null{"NULL"};
ParserKeyword s_not{"NOT"};

View File

@ -0,0 +1,90 @@
#include <Parsers/ParserDataType.h>
#include <Parsers/ExpressionElementParsers.h>
#include <Parsers/CommonParsers.h>
#include <Parsers/ASTFunction.h>
#include <Parsers/ASTIdentifier.h>
#include <Parsers/ParserCreateQuery.h>
namespace DB
{
bool ParserDataType::parseImpl(Pos & pos, ASTPtr & node, Expected & expected)
{
ParserNestedTable nested;
if (nested.parse(pos, node, expected))
return true;
String type_name;
ParserIdentifier name_parser;
ASTPtr identifier;
if (!name_parser.parse(pos, identifier, expected))
return false;
tryGetIdentifierNameInto(identifier, type_name);
String type_name_upper = Poco::toUpper(type_name);
String type_name_suffix;
/// Special cases for compatibility with SQL standard. We can parse several words as type name
/// only for certain first words, otherwise we don't know how many words to parse
if (type_name_upper == "NATIONAL")
{
if (ParserKeyword("CHARACTER LARGE OBJECT").ignore(pos))
type_name_suffix = "CHARACTER LARGE OBJECT";
else if (ParserKeyword("CHARACTER VARYING").ignore(pos))
type_name_suffix = "CHARACTER VARYING";
else if (ParserKeyword("CHAR VARYING").ignore(pos))
type_name_suffix = "CHAR VARYING";
else if (ParserKeyword("CHARACTER").ignore(pos))
type_name_suffix = "CHARACTER";
else if (ParserKeyword("CHAR").ignore(pos))
type_name_suffix = "CHAR";
}
else if (type_name_upper == "BINARY" ||
type_name_upper == "CHARACTER" ||
type_name_upper == "CHAR" ||
type_name_upper == "NCHAR")
{
if (ParserKeyword("LARGE OBJECT").ignore(pos))
type_name_suffix = "LARGE OBJECT";
else if (ParserKeyword("VARYING").ignore(pos))
type_name_suffix = "VARYING";
}
else if (type_name_upper == "DOUBLE")
{
if (ParserKeyword("PRECISION").ignore(pos))
type_name_suffix = "PRECISION";
}
if (!type_name_suffix.empty())
type_name = type_name_upper + " " + type_name_suffix;
auto function_node = std::make_shared<ASTFunction>();
function_node->name = type_name;
if (pos->type != TokenType::OpeningRoundBracket)
{
node = function_node;
return true;
}
++pos;
/// Parse optional parameters
ParserList args_parser(std::make_unique<ParserExpression>(), std::make_unique<ParserToken>(TokenType::Comma));
ASTPtr expr_list_args;
if (!args_parser.parse(pos, expr_list_args, expected))
return false;
if (pos->type != TokenType::ClosingRoundBracket)
return false;
++pos;
function_node->arguments = expr_list_args;
function_node->children.push_back(function_node->arguments);
node = function_node;
return true;
}
}

View File

@ -0,0 +1,18 @@
#pragma once
#include <Parsers/IParserBase.h>
namespace DB
{
/// Parses data type as ASTFunction
/// Examples: Int8, Array(Nullable(FixedString(16))), DOUBLE PRECISION, Nested(UInt32 CounterID, FixedString(2) UserAgentMajor)
class ParserDataType : public IParserBase
{
protected:
const char * getName() const override { return "data type"; }
bool parseImpl(Pos & pos, ASTPtr & node, Expected & expected) override;
};
}

View File

@ -2,6 +2,7 @@
#include <Parsers/ExpressionElementParsers.h>
#include <Parsers/ExpressionListParsers.h>
#include <Parsers/ParserDataType.h>
namespace DB
{
@ -9,7 +10,7 @@ namespace DB
bool ParserDictionaryAttributeDeclaration::parseImpl(Pos & pos, ASTPtr & node, Expected & expected)
{
ParserIdentifier name_parser;
ParserIdentifierWithOptionalParameters type_parser;
ParserDataType type_parser;
ParserKeyword s_default{"DEFAULT"};
ParserKeyword s_expression{"EXPRESSION"};
ParserKeyword s_hierarchical{"HIERARCHICAL"};

View File

@ -79,6 +79,7 @@ SRCS(
ParserCreateSettingsProfileQuery.cpp
ParserCreateUserQuery.cpp
ParserDescribeTableQuery.cpp
ParserDataType.cpp
ParserDictionary.cpp
ParserDictionaryAttributeDeclaration.cpp
ParserDropAccessEntityQuery.cpp

View File

@ -0,0 +1,3 @@
CREATE TABLE default.multiword_types\n(\n `a` Float64,\n `b` Float64,\n `c` String DEFAULT \'str\',\n `d` String,\n `e` String COMMENT \'comment\',\n `f` String,\n `g` String,\n `h` String DEFAULT toString(a) COMMENT \'comment\',\n `i` String,\n `j` String,\n `k` String,\n `l` String,\n `m` String,\n `n` String,\n `o` String,\n `p` String\n)\nENGINE = Memory
Tuple(Float64, Float64, String, String, String, String, String, String, String, String, String, String, String, String, String, String)
42 42

View File

@ -0,0 +1,29 @@
DROP TABLE IF EXISTS multiword_types;
CREATE TABLE multiword_types (
a DOUBLE,
b DOUBLE PRECISION,
c CHAR DEFAULT 'str',
d CHAR VARYING,
e CHAR LARGE OBJECT COMMENT 'comment',
f CHARACTER VARYING(123),
g ChArAcTeR large OBJECT,
h nchar varying (456) default toString(a) comment 'comment',
i NCHAR LARGE OBJECT,
j BINARY LARGE OBJECT,
k BINARY VARYING,
l NATIONAL CHAR,
m NATIONAL CHARACTER,
n NATIONAL CHARACTER LARGE OBJECT,
o NATIONAL CHARACTER VARYING,
p NATIONAL CHAR VARYING
) ENGINE=Memory;
SHOW CREATE TABLE multiword_types;
INSERT INTO multiword_types(a) VALUES (1);
SELECT toTypeName((*,)) FROM multiword_types;
SELECT CAST('42' AS DOUBLE PRECISION), CAST(42, 'NATIONAL CHARACTER VARYING');
DROP TABLE multiword_types;

View File

@ -0,0 +1,2 @@
9
SELECT any(number) + (any(number) * 2)\nFROM numbers(3, 10)

View File

@ -0,0 +1,4 @@
SET optimize_any_input=1;
SET enable_debug_queries=1;
SELECT any(number + number * 2) FROM numbers(3, 10);
ANALYZE SELECT any(number + number * 2) FROM numbers(3, 10);

View File

@ -0,0 +1,15 @@
SET optimize_arithmetic_operations_in_aggregate_functions = 1;
SELECT max(multiply(1)); -- { serverError 42 }
SELECT min(multiply(2));-- { serverError 42 }
SELECT sum(multiply(3)); -- { serverError 42 }
SELECT max(plus(1)); -- { serverError 42 }
SELECT min(plus(2)); -- { serverError 42 }
SELECT sum(plus(3)); -- { serverError 42 }
SELECT max(multiply()); -- { serverError 42 }
SELECT min(multiply(1, 2 ,3)); -- { serverError 42 }
SELECT sum(plus() + multiply()); -- { serverError 42 }
SELECT sum(plus(multiply(42, 3), multiply(42))); -- { serverError 42 }