2020-04-03 13:23:32 +00:00
---
2020-04-08 14:22:25 +00:00
machine_translated: true
machine_translated_rev: b111334d6614a02564cf32f379679e9ff970d9b1
toc_priority: 45
toc_title: hdfs
2020-04-03 13:23:32 +00:00
---
# hdfs {#hdfs}
2020-04-08 14:22:25 +00:00
从HDFS中的文件创建表。 此表函数类似于 [url ](url.md ) 和 [文件 ](file.md ) 一些的。
2020-04-03 13:23:32 +00:00
``` sql
hdfs(URI, format, structure)
```
2020-04-08 14:22:25 +00:00
**输入参数**
2020-04-03 13:23:32 +00:00
2020-04-08 14:22:25 +00:00
- `URI` — The relative URI to the file in HDFS. Path to file support following globs in readonly mode: `*` , `?` , `{abc,def}` 和 `{N..M}` 哪里 `N` , `M` — numbers, \``'abc', 'def'` — strings.
- `format` — The [格式 ](../../interfaces/formats.md#formats ) 的文件。
2020-04-03 13:23:32 +00:00
- `structure` — Structure of the table. Format `'column1_name column1_type, column2_name column2_type, ...'` .
2020-04-08 14:22:25 +00:00
**返回值**
2020-04-03 13:23:32 +00:00
2020-04-08 14:22:25 +00:00
具有指定结构的表,用于读取或写入指定文件中的数据。
2020-04-03 13:23:32 +00:00
2020-04-08 14:22:25 +00:00
**示例**
2020-04-03 13:23:32 +00:00
2020-04-08 14:22:25 +00:00
表从 `hdfs://hdfs1:9000/test` 并从中选择前两行:
2020-04-03 13:23:32 +00:00
``` sql
SELECT *
FROM hdfs('hdfs://hdfs1:9000/test', 'TSV', 'column1 UInt32, column2 UInt32, column3 UInt32')
LIMIT 2
```
``` text
┌─column1─┬─column2─┬─column3─┐
│ 1 │ 2 │ 3 │
│ 3 │ 2 │ 1 │
└─────────┴─────────┴─────────┘
```
2020-04-08 14:22:25 +00:00
**路径中的水珠**
2020-04-03 13:23:32 +00:00
2020-04-08 14:22:25 +00:00
多个路径组件可以具有globs。 对于正在处理的文件应该存在并匹配到整个路径模式(不仅后缀或前缀)。
2020-04-03 13:23:32 +00:00
2020-04-08 14:22:25 +00:00
- `*` — Substitutes any number of any characters except `/` 包括空字符串。
2020-04-03 13:23:32 +00:00
- `?` — Substitutes any single character.
- `{some_string,another_string,yet_another_one}` — Substitutes any of strings `'some_string', 'another_string', 'yet_another_one'` .
2020-04-08 14:22:25 +00:00
- `{N..M}` — Substitutes any number in range from N to M including both borders.
2020-04-03 13:23:32 +00:00
2020-04-08 14:22:25 +00:00
建筑与 `{}` 类似于 [远程表功能 ](../../sql_reference/table_functions/remote.md )).
2020-04-03 13:23:32 +00:00
2020-04-08 14:22:25 +00:00
**示例**
2020-04-03 13:23:32 +00:00
2020-04-08 14:22:25 +00:00
1. 假设我们在HDFS上有几个具有以下Uri的文件:
2020-04-03 13:23:32 +00:00
- ‘ hdfs://hdfs1:9000/some\_dir/some\_file\_1’
- ‘ hdfs://hdfs1:9000/some\_dir/some\_file\_2’
- ‘ hdfs://hdfs1:9000/some\_dir/some\_file\_3’
- ‘ hdfs://hdfs1:9000/another\_dir/some\_file\_1’
- ‘ hdfs://hdfs1:9000/another\_dir/some\_file\_2’
- ‘ hdfs://hdfs1:9000/another\_dir/some\_file\_3’
2020-04-08 14:22:25 +00:00
1. 查询这些文件中的行数:
2020-04-03 13:23:32 +00:00
<!-- -->
``` sql
SELECT count(*)
FROM hdfs('hdfs://hdfs1:9000/{some,another}_dir/some_file_{1..3}', 'TSV', 'name String, value UInt32')
```
2020-04-08 14:22:25 +00:00
1. 查询这两个目录的所有文件中的行数:
2020-04-03 13:23:32 +00:00
<!-- -->
``` sql
SELECT count(*)
FROM hdfs('hdfs://hdfs1:9000/{some,another}_dir/*', 'TSV', 'name String, value UInt32')
```
2020-04-08 14:22:25 +00:00
!!! warning "警告"
如果您的文件列表包含带前导零的数字范围,请单独使用带大括号的构造或使用 `?` .
2020-04-03 13:23:32 +00:00
2020-04-08 14:22:25 +00:00
**示例**
2020-04-03 13:23:32 +00:00
2020-04-08 14:22:25 +00:00
从名为 `file000` , `file001` , … , `file999` :
2020-04-03 13:23:32 +00:00
``` sql
SELECT count(*)
FROM hdfs('hdfs://hdfs1:9000/big_dir/file{0..9}{0..9}{0..9}', 'CSV', 'name String, value UInt32')
```
2020-04-08 14:22:25 +00:00
## 虚拟列 {#virtual-columns}
2020-04-03 13:23:32 +00:00
- `_path` — Path to the file.
- `_file` — Name of the file.
2020-04-08 14:22:25 +00:00
**另请参阅**
2020-04-03 13:23:32 +00:00
2020-04-08 14:22:25 +00:00
- [虚拟列 ](https://clickhouse.tech/docs/en/operations/table_engines/#table_engines-virtual_columns )
2020-04-03 13:23:32 +00:00
2020-04-08 14:22:25 +00:00
[原始文章 ](https://clickhouse.tech/docs/en/query_language/table_functions/hdfs/ ) <!--hide-->