mirror of
https://github.com/ClickHouse/ClickHouse.git
synced 2024-11-24 16:42:05 +00:00
[Docs] Fix style check errors
This commit is contained in:
parent
20385161e4
commit
0dd4735e07
@ -7,13 +7,13 @@ description: Analyzing Stack Overflow data with ClickHouse
|
||||
|
||||
# Analyzing Stack Overflow data with ClickHouse
|
||||
|
||||
This dataset contains every Post, User, Vote, Comment, Badge, PostHistory, and PostLink that has occurred on Stack Overflow.
|
||||
This dataset contains every `Post`, `User`, `Vote`, `Comment`, `Badge, `PostHistory`, and `PostLink` that has occurred on Stack Overflow.
|
||||
|
||||
Users can either download pre-prepared Parquet versions of the data, containing every post up to April 2024, or download the latest data in XML format and load this. Stack Overflow provide updates to this data periodically - historically every 3 months.
|
||||
|
||||
The following diagram shows the schema for the available tables assuming Parquet format.
|
||||
|
||||
![Stackoverflow schema](./images/stackoverflow.png)
|
||||
![Stack Overflow schema](./images/stackoverflow.png)
|
||||
|
||||
A description of the schema of this data can be found [here](https://meta.stackexchange.com/questions/2677/database-schema-documentation-for-the-public-data-dump-and-sede).
|
||||
|
||||
@ -159,7 +159,7 @@ INSERT INTO stackoverflow.badges SELECT * FROM s3('https://datasets-documentatio
|
||||
0 rows in set. Elapsed: 6.635 sec. Processed 51.29 million rows, 797.05 MB (7.73 million rows/s., 120.13 MB/s.)
|
||||
```
|
||||
|
||||
### Postlinks
|
||||
### `PostLinks`
|
||||
|
||||
```sql
|
||||
CREATE TABLE stackoverflow.postlinks
|
||||
@ -178,7 +178,7 @@ INSERT INTO stackoverflow.postlinks SELECT * FROM s3('https://datasets-documenta
|
||||
0 rows in set. Elapsed: 1.534 sec. Processed 6.55 million rows, 129.70 MB (4.27 million rows/s., 84.57 MB/s.)
|
||||
```
|
||||
|
||||
### PostHistory
|
||||
### `PostHistory`
|
||||
|
||||
```sql
|
||||
CREATE TABLE stackoverflow.posthistory
|
||||
@ -222,7 +222,7 @@ These files are upto 35GB and can take around 30 mins to download depending on i
|
||||
|
||||
### Convert to JSON
|
||||
|
||||
At the time of writing, ClickHouse does not have native support for XML as an input format. To load the data into ClickHouse we first convert to NdJSON.
|
||||
At the time of writing, ClickHouse does not have native support for XML as an input format. To load the data into ClickHouse we first convert to NDJSON.
|
||||
|
||||
To convert XML to JSON we recommend the [`xq`](https://github.com/kislyuk/yq) linux tool, a simple `jq` wrapper for XML documents.
|
||||
|
||||
@ -301,7 +301,7 @@ Peak memory usage: 224.03 MiB.
|
||||
|
||||
### User with the most answers (active accounts)
|
||||
|
||||
Account requires a UserId.
|
||||
Account requires a `UserId`.
|
||||
|
||||
```sql
|
||||
SELECT
|
||||
|
@ -572,6 +572,7 @@ MySQLDump
|
||||
MySQLThreads
|
||||
NATS
|
||||
NCHAR
|
||||
NDJSON
|
||||
NEKUDOTAYIM
|
||||
NEWDATE
|
||||
NEWDECIMAL
|
||||
@ -714,6 +715,8 @@ PlantUML
|
||||
PointDistKm
|
||||
PointDistM
|
||||
PointDistRads
|
||||
PostHistory
|
||||
PostLink
|
||||
PostgreSQLConnection
|
||||
PostgreSQLThreads
|
||||
Postgres
|
||||
@ -2496,6 +2499,7 @@ sqlite
|
||||
sqrt
|
||||
src
|
||||
srcReplicas
|
||||
stackoverflow
|
||||
stacktrace
|
||||
stacktraces
|
||||
startsWith
|
||||
@ -2834,6 +2838,7 @@ userver
|
||||
utils
|
||||
uuid
|
||||
uuidv
|
||||
vCPU
|
||||
varPop
|
||||
varPopStable
|
||||
varSamp
|
||||
|
Loading…
Reference in New Issue
Block a user