mirror of
https://github.com/ClickHouse/ClickHouse.git
synced 2024-11-27 01:51:59 +00:00
1.1 KiB
1.1 KiB
toc_priority | toc_title |
---|---|
68 | External Disks for Storing Data |
External Disks for Storing Data
Data, processed in ClickHouse, is usually stored in the local file system — on the same machine with the ClickHouse server. That requires large-capacity disks, which can be expensive enough. To avoid that you can store the data remotely — on Amazon s3 disks or in the Hadoop Distributed File System (HDFS).
To work with data stored on Amazon s3
disks use s3 table engine, and to work with data in the Hadoop Distributed File System — HDFS table engine.
Zero-copy Replication
ClickHouse supports zero-copy replication for s3
and HDFS
disks, which means that if the data is stored remotely on several machines and needs to be synchronized, then only the metadata is replicated (paths to the data parts), but not the data itself.