Description
Issue description
- This issue is a regression.
- It is unknown if this issue is a regression.
Describe your issue in detail and steps it took to produce it.
Impact
It is unclear what a user does wrong or can do with it.
How frequently does it reproduce?
First time observed.
Installation details
Kernel Version: 5.15.0-1028-aws
Scylla version (or git commit hash): 2022.2.0-20230112.4f0f82ff2e1d
with build-id 6bb071079708f1e48d1faeee6ef3cc3ef93733af
Cluster size: 4 nodes (is4gen.4xlarge)
Scylla Nodes used in this run:
- longevity-large-partitions-4d-vp-br-db-node-b4fd1150-8 (54.246.26.242 | 10.4.0.15) (shards: 15)
- longevity-large-partitions-4d-vp-br-db-node-b4fd1150-7 (3.251.80.191 | 10.4.2.60) (shards: 15)
- longevity-large-partitions-4d-vp-br-db-node-b4fd1150-6 (3.253.32.157 | 10.4.0.237) (shards: 15)
- longevity-large-partitions-4d-vp-br-db-node-b4fd1150-5 (3.253.42.60 | 10.4.1.171) (shards: 15)
- longevity-large-partitions-4d-vp-br-db-node-b4fd1150-4 (34.243.215.150 | 10.4.1.50) (shards: 15)
- longevity-large-partitions-4d-vp-br-db-node-b4fd1150-3 (54.73.239.104 | 10.4.1.3) (shards: 15)
- longevity-large-partitions-4d-vp-br-db-node-b4fd1150-2 (52.51.73.98 | 10.4.3.254) (shards: 15)
- longevity-large-partitions-4d-vp-br-db-node-b4fd1150-1 (34.240.149.141 | 10.4.2.223) (shards: 15)
OS / Image: ami-0a018bb5789118823
(aws: eu-west-1)
Test: longevity-large-partition-4days-arm-test
Test id: b4fd1150-5888-4a5c-b35a-6f0e3fdbdad9
Test name: scylla-5.2/longevity/longevity-large-partition-4days-arm-test
Test config file(s):
Details:
Running large partitions test where we have huge latencies appeared following error:
panic: runtime error: index out of range [2] with length 0
goroutine 1 [running]:
github.com/HdrHistogram/hdrhistogram-go.(*Histogram).getCountAtIndex(...)
/go/pkg/mod/github.com/!hdr!histogram/[email protected]/hdr.go:595
github.com/HdrHistogram/hdrhistogram-go.(*iterator).nextCountAtIdx(0x10000203073?, 0x7ff01ad9d748?)
/go/pkg/mod/github.com/!hdr!histogram/[email protected]/hdr.go:662 +0xc5
github.com/HdrHistogram/hdrhistogram-go.(*iterator).next(0xc0000ef020)
/go/pkg/mod/github.com/!hdr!histogram/[email protected]/hdr.go:670 +0x25
github.com/HdrHistogram/hdrhistogram-go.(*rIterator).next(...)
/go/pkg/mod/github.com/!hdr!histogram/[email protected]/hdr.go:683
github.com/HdrHistogram/hdrhistogram-go.(*Histogram).Merge(0xf0000000e?, 0x4000000000a?)
/go/pkg/mod/github.com/!hdr!histogram/[email protected]/hdr.go:177 +0x8d
github.com/scylladb/scylla-bench/pkg/results.(*MergedResult).AddResult(0xc1cfb01bc0, {0x0, 0x0, 0x1, 0x7d0, 0x0, {0x0, 0x0, 0x0}, 0xc2f5d83300, ...})
/go/scylla-bench-0.1.16/pkg/results/merged_result.go:54 +0x1b5
github.com/scylladb/scylla-bench/pkg/results.(*TestResults).GetResultsFromThreadsAndMerge(0xc000162840)
/go/scylla-bench-0.1.16/pkg/results/thread_results.go:60 +0x89
github.com/scylladb/scylla-bench/pkg/results.(*TestResults).GetTotalResults(0xc000162840)
/go/scylla-bench-0.1.16/pkg/results/thread_results.go:82 +0xcc
main.main()
/go/scylla-bench-0.1.16/main.go:640 +0x3b1e
Probably appears having a too big loader load like in the following bug: #121
Monitoring:
- Restore Monitor Stack command:
$ hydra investigate show-monitor b4fd1150-5888-4a5c-b35a-6f0e3fdbdad9
- Restore monitor on AWS instance using Jenkins job
- Show all stored logs command:
$ hydra investigate show-logs b4fd1150-5888-4a5c-b35a-6f0e3fdbdad9