Skip to content

Commit a66af84

Browse files
authored
chore: fix loading data for benchmark (#17889)
1 parent c2ce223 commit a66af84

File tree

3 files changed

+11
-9
lines changed

3 files changed

+11
-9
lines changed

benchmark/clickbench/benchmark_local.sh

+2
Original file line numberDiff line numberDiff line change
@@ -70,6 +70,8 @@ token=$(curl -X PUT "http://169.254.169.254/latest/api/token" -H "X-aws-ec2-meta
7070
instance_type=$(curl -H "X-aws-ec2-metadata-token: $token" http://169.254.169.254/latest/meta-data/instance-type)
7171
echo "Instance type: ${instance_type}"
7272

73+
echo "CREATE STAGE IF NOT EXISTS datasets url = 's3://repo.databend.rs' CONNECTION =(region='us-east-2');" | bendsql
74+
7375
# Load data
7476
echo "Loading data..."
7577
load_start=$(date +%s)

benchmark/clickbench/hits/load.sql

+1-1
Original file line numberDiff line numberDiff line change
@@ -1,5 +1,5 @@
11
COPY INTO hits
2-
FROM 'https://datasets.databend.org/hits_100m_obfuscated_v1.tsv.xz' FILE_FORMAT = (
2+
FROM @datasets/hits/hits_100m_obfuscated_v1.tsv.xz FILE_FORMAT =(
33
type = TSV compression = XZ field_delimiter = '\t' record_delimiter = '\n' skip_header = 0
44
);
55
ANALYZE TABLE hits;

benchmark/clickbench/tpch/load.sql

+8-8
Original file line numberDiff line numberDiff line change
@@ -1,40 +1,40 @@
11
COPY INTO customer
2-
FROM 's3://repo.databend.com/tpch100/customer/' CONNECTION =(REGION = 'us-east-2') PATTERN = 'customer.tbl.*' FILE_FORMAT =(
2+
FROM @datasets/tpch/customer/ PATTERN = 'customer.tbl.*' FILE_FORMAT =(
33
type = 'CSV' field_delimiter = '|' record_delimiter = '\n' skip_header = 1
44
);
55
ANALYZE TABLE customer;
66
COPY INTO lineitem
7-
FROM 's3://repo.databend.com/tpch100/lineitem/' CONNECTION =(REGION = 'us-east-2') PATTERN = 'lineitem.tbl.*' FILE_FORMAT =(
7+
FROM @datasets/tpch/lineitem/ PATTERN = 'lineitem.tbl.*' FILE_FORMAT =(
88
type = 'CSV' field_delimiter = '|' record_delimiter = '\n' skip_header = 0
99
);
1010
ANALYZE TABLE lineitem;
1111
COPY INTO nation
12-
FROM 's3://repo.databend.com/tpch100/nation.tbl' CONNECTION =(REGION = 'us-east-2') FILE_FORMAT =(
12+
FROM @datasets/tpch/nation.tbl FILE_FORMAT =(
1313
type = 'CSV' field_delimiter = '|' record_delimiter = '\n' skip_header = 0
1414
);
1515
ANALYZE TABLE nation;
1616
COPY INTO orders
17-
FROM 's3://repo.databend.com/tpch100/orders/' CONNECTION =(REGION = 'us-east-2') PATTERN = 'orders.tbl.*' FILE_FORMAT =(
17+
FROM @datasets/tpch/orders/ PATTERN = 'orders.tbl.*' FILE_FORMAT =(
1818
type = 'CSV' field_delimiter = '|' record_delimiter = '\n' skip_header = 0
1919
);
2020
ANALYZE TABLE orders;
2121
COPY INTO partsupp
22-
FROM 's3://repo.databend.com/tpch100/partsupp/' CONNECTION =(REGION = 'us-east-2') PATTERN = 'partsupp.tbl.*' FILE_FORMAT =(
22+
FROM @datasets/tpch/partsupp/ PATTERN = 'partsupp.tbl.*' FILE_FORMAT =(
2323
type = 'CSV' field_delimiter = '|' record_delimiter = '\n' skip_header = 0
2424
);
2525
ANALYZE TABLE partsupp;
2626
COPY INTO part
27-
FROM 's3://repo.databend.com/tpch100/part/' CONNECTION =(REGION = 'us-east-2') PATTERN = 'part.tbl.*' FILE_FORMAT =(
27+
FROM @datasets/tpch/part/ PATTERN = 'part.tbl.*' FILE_FORMAT =(
2828
type = 'CSV' field_delimiter = '|' record_delimiter = '\n' skip_header = 0
2929
);
3030
ANALYZE TABLE part;
3131
COPY INTO region
32-
FROM 's3://repo.databend.com/tpch100/region.tbl' CONNECTION =(REGION = 'us-east-2') FILE_FORMAT =(
32+
FROM @datasets/tpch/region.tbl FILE_FORMAT =(
3333
type = 'CSV' field_delimiter = '|' record_delimiter = '\n' skip_header = 0
3434
);
3535
ANALYZE TABLE region;
3636
COPY INTO supplier
37-
FROM 's3://repo.databend.com/tpch100/supplier/' CONNECTION =(REGION = 'us-east-2') PATTERN = 'supplier.tbl.*' FILE_FORMAT =(
37+
FROM @datasets/tpch/supplier/ PATTERN = 'supplier.tbl.*' FILE_FORMAT =(
3838
type = 'CSV' field_delimiter = '|' record_delimiter = '\n' skip_header = 0
3939
);
4040
ANALYZE TABLE supplier;

0 commit comments

Comments
 (0)