mirror of
https://github.com/ClickHouse/ClickHouse.git
synced 2024-12-18 20:32:43 +00:00
67 lines
1.7 KiB
Markdown
67 lines
1.7 KiB
Markdown
---
|
|
sidebar_label: WikiStat
|
|
---
|
|
|
|
# WikiStat
|
|
|
|
The dataset contains 0.5 trillion records.
|
|
|
|
See the video from FOSDEM 2023: https://www.youtube.com/watch?v=JlcI2Vfz_uk
|
|
|
|
And the presentation: https://presentations.clickhouse.com/fosdem2023/
|
|
|
|
Data source: https://dumps.wikimedia.org/other/pageviews/
|
|
|
|
Getting the list of links:
|
|
```
|
|
for i in {2015..2023}; do
|
|
for j in {01..12}; do
|
|
echo "${i}-${j}" >&2
|
|
curl -sSL "https://dumps.wikimedia.org/other/pageviews/$i/$i-$j/" \
|
|
| grep -oE 'pageviews-[0-9]+-[0-9]+\.gz'
|
|
done
|
|
done | sort | uniq | tee links.txt
|
|
```
|
|
|
|
Downloading the data:
|
|
```
|
|
sed -r 's!pageviews-([0-9]{4})([0-9]{2})[0-9]{2}-[0-9]+\.gz!https://dumps.wikimedia.org/other/pageviews/\1/\1-\2/\0!' \
|
|
links.txt | xargs -P3 wget --continue
|
|
```
|
|
|
|
(it will take about 3 days)
|
|
|
|
Creating a table:
|
|
|
|
``` sql
|
|
CREATE TABLE wikistat
|
|
(
|
|
time DateTime CODEC(Delta, ZSTD(3)),
|
|
project LowCardinality(String),
|
|
subproject LowCardinality(String),
|
|
path String CODEC(ZSTD(3)),
|
|
hits UInt64 CODEC(ZSTD(3)),
|
|
size UInt64 CODEC(ZSTD(3))
|
|
)
|
|
ENGINE = MergeTree
|
|
ORDER BY (path, time);
|
|
```
|
|
|
|
Loading the data:
|
|
|
|
```
|
|
clickhouse-local --query "
|
|
WITH replaceRegexpOne(_path, '^.+pageviews-(\\d{4})(\\d{2})(\\d{2})-(\\d{2})(\\d{2})(\\d{2}).gz$', '\1-\2-\3 \4-\5-\6')::DateTime AS time,
|
|
extractGroups(line, '^([^ \\.]+)(\\.[^ ]+)? +([^ ]+) +(\\d+) +(\\d+)$') AS values
|
|
SELECT
|
|
time,
|
|
values[1] AS project,
|
|
values[2] AS subproject,
|
|
values[3] AS path,
|
|
(values[4])::UInt64 AS hits,
|
|
(values[5])::UInt64 AS size
|
|
FROM file('pageviews*.gz', LineAsString)
|
|
WHERE length(values) = 5 FORMAT Native
|
|
" | clickhouse-client --query "INSERT INTO wikistat FORMAT Native"
|
|
```
|