Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Cassandra-stress failed to create test table due to lack of schema agreement #9906

Closed
juliayakovlev opened this issue Jan 12, 2022 · 74 comments
Assignees
Labels
area/commitlog Issues related to the commit log. bug
Milestone

Comments

@juliayakovlev
Copy link

Installation details
Kernel version: 5.11.0-1022-aws
Scylla version (or git commit hash): 4.7.dev-0.20211230.12fa68fe6 with build-id 938d162cf1be1483853bf5b5417f5daf11e5ac24
Cluster size: 6 nodes (r5b.2xlarge)
OS (RHEL/CentOS/Ubuntu/AWS AMI): ami-0668e251eb4f0f394 (aws: eu-west-1)

Scylla running with shards number (live nodes):
longevity-100gb-4h-master-db-node-bce0924c-1 (34.252.180.109 | 10.0.3.53): 8 shards
longevity-100gb-4h-master-db-node-bce0924c-2 (34.244.58.179 | 10.0.1.246): 8 shards
longevity-100gb-4h-master-db-node-bce0924c-3 (3.250.166.213 | 10.0.0.220): 8 shards
longevity-100gb-4h-master-db-node-bce0924c-4 (54.216.84.128 | 10.0.1.149): 8 shards
longevity-100gb-4h-master-db-node-bce0924c-5 (54.75.69.84 | 10.0.2.137): 8 shards
longevity-100gb-4h-master-db-node-bce0924c-6 (34.244.236.61 | 10.0.3.61): 8 shards

Test: longevity-100gb-4h-ebs-gp3-test
Test name: longevity_test.LongevityTest.test_custom_time
Test config file(s):

Issue description

Test started with running of cassandra-stress load. There are a lot of WriteTimeoutException. But nodes are not overloaded.

In parallel during first nemesis (NoCorruptRepair) 10 keyspaces are created by running cassandra-stress commands.
Creation of drop_table_during_repair_ks_0.standard1 table failed because of luck of schema agreement (see log cassandra-stress-l0-c0-k1-c8336276-d647-4b88-90dd-34615541c1f0.log in the loader-set logs):

< t:2022-01-09 04:55:46,260 f:stress_thread.py l:168  c:sdcm.stress_thread   p:INFO  > cassandra-stress write n=400000 cl=QUORUM -port jmx=6868 -mode native cql3 -schema keyspace=drop_table_during_repair_ks_0 'replication(factor=6)' -log interval=5 -node 10.0.3.53 -errors skip-unsupported-columns

04:56:00.326 [cluster1-worker-3] DEBUG c.d.driver.core.ControlConnection - Checking for schema agreement: versions are [38810554-e5fb-3e69-924c-2eeb0cb72164, 4c6da8c9-2204-395b-9752-4937ab19e9b9]
04:56:00.526 [cluster1-worker-3] WARN  com.datastax.driver.core.Cluster - No schema agreement from live replicas after 10 s. The schema may not be up to date on some nodes.
04:56:00.526 [cluster1-worker-0] DEBUG c.d.driver.core.ControlConnection - [Control connection] Refreshing schema for drop_table_during_repair_ks_0.standard1 (TABLE)
Created keyspaces. Sleeping 1s for propagation.
Sleeping 2s...
Warming up WRITE with 50000 iterations...
Failed to connect over JMX; not collecting these stats
04:56:03.823 [Thread-70] DEBUG com.datastax.driver.core.Connection - Connection[/10.0.0.220:9042-4, inFlight=1, closed=false] Setting keyspace drop_table_during_repair_ks_0
04:56:03.832 [cluster1-nio-worker-5] DEBUG com.datastax.driver.core.Connection - Connection[/10.0.0.220:9042-4, inFlight=1, closed=false] Keyspace set to drop_table_during_repair_ks_0
04:56:03.836 [Thread-100] DEBUG com.datastax.driver.core.Connection - Connection[/10.0.2.137:9042-3, inFlight=1, closed=false] Setting keyspace drop_table_during_repair_ks_0
com.datastax.driver.core.exceptions.InvalidQueryException: unconfigured table standard1
	at com.datastax.driver.core.exceptions.InvalidQueryException.copy(InvalidQueryException.java:50)
	at com.datastax.driver.core.DriverThrowables.propagateCause(DriverThrowables.java:35)
	at com.datastax.driver.core.AbstractSession.prepare(AbstractSession.java:86)
	at org.apache.cassandra.stress.util.JavaDriverClient.prepare(JavaDriverClient.java:116)
	at org.apache.cassandra.stress.operations.predefined.CqlOperation$JavaDriverWrapper.createPreparedStatement(CqlOperation.java:318)
	at org.apache.cassandra.stress.operations.predefined.CqlOperation.run(CqlOperation.java:77)
	at org.apache.cassandra.stress.operations.predefined.CqlOperation.run(CqlOperation.java:109)04:56:03.838 [cluster1-nio-worker-4] DEBUG com.datastax.driver.core.Connection - Connection[/10.0.2.137:9042-3, inFlight=1, closed=false] Keyspace set to drop_table_during_repair_ks_0

	at org.apache.cassandra.stress.operations.predefined.CqlOperation.run(CqlOperation.java:264)
	at org.apache.cassandra.stress.StressAction$Consumer.run(StressAction.java:469)
Caused by: com.datastax.driver.core.exceptions.InvalidQueryException: unconfigured table standard1
	at com.datastax.driver.core.Responses$Error.asException(Responses.java:181)
	at com.datastax.driver.core.SessionManager$4.apply(SessionManager.java:249)
	at com.datastax.driver.core.SessionManager$4.apply(SessionManager.java:218)
	at com.google.common.util.concurrent.Futures$ChainingListenableFuture.run(Futures.java:906)
	at com.google.common.util.concurrent.Futures$1$1.run(Futures.java:635)
	at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128)
	at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628)
	at com.datastax.shaded.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30)
	at java.base/java.lang.Thread.run(Thread.java:834)
04:56:03.971 [cluster1-nio-worker-2] DEBUG c.d.driver.core.RequestHandler - [609104306-0] Error querying 10.0.0.220/10.0.0.220:9042 : com.datastax.driver.core.exceptions.DriverException: Error preparing query, got ERROR INVALID: unconfigured table standard1
04:56:03.974 [cluster1-nio-worker-3] DEBUG com.datastax.driver.core.Connection - Connection[/10.0.1.246:9042-10, inFlight=1, closed=false] Keyspace set to drop_table_during_repair_ks_0
java.lang.RuntimeException: Failed to execute warmup
	at org.apache.cassandra.stress.StressAction.warmup(StressAction.java:124)
	at org.apache.cassandra.stress.StressAction.run(StressAction.java:71)
	at org.apache.cassandra.stress.Stress.run(Stress.java:143)
	at org.apache.cassandra.stress.Stress.main(Stress.java:62)

schema_version_loading_failed and mutation_write_timeout_exception on the longevity-100gb-4h-master-db-node-bce0924c-3 node

Jan 09 04:55:47 longevity-100gb-4h-master-db-node-bce0924c-3 scylla[16275]:  [shard 4] schema_tables - Creating keyspace drop_table_during_repair_ks_0
Jan 09 04:55:47 longevity-100gb-4h-master-db-node-bce0924c-3 scylla[16275]:  [shard 4] schema_tables - Schema version changed to 4c6da8c9-2204-395b-9752-4937ab19e9b9
Jan 09 04:55:48 longevity-100gb-4h-master-db-node-bce0924c-3 scylla[16275]:  [shard 0] compaction - [Compact system_schema.keyspaces 69108490-7108-11ec-817c-81941c0da70d] Compacted 2 sstables to [/var/lib/scylla
/data/system_schema/keyspaces-abac5682dea631c5b535b3d6cffd0fb6/md-88-big-Data.db:level=0]. 11kB to 6kB (~53% of original) in 249ms = 25kB/s. ~256 total partitions merged to 8.
Jan 09 04:55:50 longevity-100gb-4h-master-db-node-bce0924c-3 scylla[16275]:  [shard 0] storage_service - fail to update schema_version for 10.0.3.61: exceptions::mutation_write_timeout_exception (Operation timed
 out for system.peers - received only 0 responses from 1 CL=ONE.)
Jan 09 04:55:50 longevity-100gb-4h-master-db-node-bce0924c-3 scylla[16275]:  [shard 0] storage_service - fail to update schema_version for 10.0.1.246: exceptions::mutation_write_timeout_exception (Operation time
d out for system.peers - received only 0 responses from 1 CL=ONE.)
Jan 09 04:55:50 longevity-100gb-4h-master-db-node-bce0924c-3 scylla[16275]:  [shard 6] migration_manager - Failed to update definitions from 10.0.1.149:6: seastar::timed_out_error (timedout)
Jan 09 04:55:51 longevity-100gb-4h-master-db-node-bce0924c-3 scylla[16275]:  [shard 0] storage_service - fail to update schema_version for 10.0.2.137: exceptions::mutation_write_timeout_exception (Operation time
d out for system.peers - received only 0 responses from 1 CL=ONE.)
Jan 09 04:56:04 longevity-100gb-4h-master-db-node-bce0924c-3 scylla[16275]:  [shard 2] storage_proxy - Failed to apply mutation from 10.0.1.149#2: std::_Nested_exception<schema_version_loading_failed> (Failed to load schema version 8533f20a-dfdb-3852-ad2a-2ff8b9757ab3): seastar::timed_out_error (timedout)
Jan 09 04:56:04 longevity-100gb-4h-master-db-node-bce0924c-3 scylla[16275]:  [shard 3] storage_proxy - Failed to apply mutation from 10.0.1.246#3: std::_Nested_exception<schema_version_loading_failed> (Failed to load schema version 8533f20a-dfdb-3852-ad2a-2ff8b9757ab3): seastar::timed_out_error (timedout)
Jan 09 04:56:04 longevity-100gb-4h-master-db-node-bce0924c-3 scylla[16275]:  [shard 3] storage_proxy - Failed to apply mutation from 10.0.1.149#3: std::_Nested_exception<schema_version_loading_failed> (Failed to load schema version 8533f20a-dfdb-3852-ad2a-2ff8b9757ab3): seastar::timed_out_error (timedout)
Jan 09 04:56:04 longevity-100gb-4h-master-db-node-bce0924c-3 scylla[16275]:  [shard 4] storage_proxy - Failed to apply mutation from 10.0.1.149#4: std::_Nested_exception<schema_version_loading_failed> (Failed to load schema version 8533f20a-dfdb-3852-ad2a-2ff8b9757ab3): seastar::timed_out_error (timedout)
Jan 09 04:56:04 longevity-100gb-4h-master-db-node-bce0924c-3 scylla[16275]:  [shard 5] storage_proxy - Failed to apply mutation from 10.0.1.149#5: std::_Nested_exception<schema_version_loading_failed> (Failed to load schema version 8533f20a-dfdb-3852-ad2a-2ff8b9757ab3): seastar::timed_out_error (timedout)
Jan 09 04:56:42 longevity-100gb-4h-master-db-node-bce0924c-3 scylla[16275]:  [shard 1] seastar - Exceptional future ignored: exceptions::mutation_write_timeout_exception (Operation timed out for system.clients - received only 0 responses from 1 CL=ONE.), backtrace: 0x4411fce 0x44124c0 0x44127c8 0x403e80b 0x12e9f1e 0x4081c64 0x4083057 0x40a1fb5 0x405427a /opt/scylladb/libreloc/libpthread.so.0+0x9298 /opt/scylladb/libreloc/libc.so.6+0x100352
                                                                               --------
                                                                               N7seastar12continuationINS_8internal22promise_base_with_typeIvEEZNS_6futureINS_10shared_ptrIN4cql318untyped_result_setEEEE14discard_resultEvEUlDpOT_E_ZNS9_14then_impl_nrvoISD_NS4_IvEEEET0_OT_EUlOS3_RSD_ONS_12future_stateIS8_EEE_S8_EE

Same failure for drop_table_during_repair_ks_9 keyspace.

Screenshot from 2022-01-12 17-24-12
Screenshot from 2022-01-12 17-23-40
Screenshot from 2022-01-12 17-22-58

Restore Monitor Stack command: $ hydra investigate show-monitor bce0924c-4be2-49a3-afdd-35cf020461b8
Restore monitor on AWS instance using Jenkins job
Show all stored logs command: $ hydra investigate show-logs bce0924c-4be2-49a3-afdd-35cf020461b8

Test id: bce0924c-4be2-49a3-afdd-35cf020461b8

Logs:
db-cluster - https://cloudius-jenkins-test.s3.amazonaws.com/bce0924c-4be2-49a3-afdd-35cf020461b8/20220109_060331/db-cluster-bce0924c.tar.gz

loader-set - https://cloudius-jenkins-test.s3.amazonaws.com/bce0924c-4be2-49a3-afdd-35cf020461b8/20220109_060331/loader-set-bce0924c.tar.gz

sct - https://cloudius-jenkins-test.s3.amazonaws.com/bce0924c-4be2-49a3-afdd-35cf020461b8/20220109_060331/sct-runner-bce0924c.tar.gz

Jenkins job URL

@roydahan roydahan added the triage/master Looking for assignee label Jan 12, 2022
@roydahan
Copy link

Whoever triaging this, please note it's a test with "slow disks" (using EBS gp3) - so the nodes are probably bounded by disk.

@slivne slivne changed the title Cassandra-stress failed to create test table due to luck of the schema agreement Cassandra-stress failed to create test table due to lack of schema agreement Jan 16, 2022
@slivne slivne self-assigned this Jan 26, 2022
@bentsi
Copy link
Contributor

bentsi commented Jan 30, 2022

happens also on 4.6

Installation details
Kernel version: 5.11.0-1022-aws
Scylla version (or git commit hash): 4.6.rc4-0.20220130.dde50536904 with build-id 1a8cb922cb8ab10a866f4dd89d1d4b63275211fb
Cluster size: 6 nodes (r5b.2xlarge)
Scylla running with shards number (live nodes):
longevity-100gb-4h-4-6-db-node-d7a36396-1 (3.236.191.31 | 10.0.1.51): 8 shards
longevity-100gb-4h-4-6-db-node-d7a36396-2 (3.230.163.199 | 10.0.3.70): 8 shards
longevity-100gb-4h-4-6-db-node-d7a36396-3 (18.209.210.255 | 10.0.2.133): 8 shards
longevity-100gb-4h-4-6-db-node-d7a36396-4 (44.195.80.96 | 10.0.0.196): 8 shards
longevity-100gb-4h-4-6-db-node-d7a36396-5 (44.200.157.199 | 10.0.0.16): 8 shards
longevity-100gb-4h-4-6-db-node-d7a36396-6 (3.238.69.86 | 10.0.3.124): 8 shards
OS (RHEL/CentOS/Ubuntu/AWS AMI): ami-0edd50b6601c2f805 (aws: us-east-1)

Test: longevity-100gb-4h-ebs-gp3-test
Test name: longevity_test.LongevityTest.test_custom_time
Test config file(s):

Issue description

====================================

run nodetool flush on node 1 before any nemesis starts

< t:2022-01-30 14:55:26,516 f:remote_base.py  l:520  c:RemoteCmdRunner      p:DEBUG > Running command "/usr/bin/nodetool  flush "...
< t:2022-01-30 14:55:26,517 f:file_logger.py  l:89   c:sdcm.sct_events.file_logger p:INFO  > 2022-01-30 14:55:26.515: (NodetoolEvent Severity.NORMAL) period_type=begin event_id=6da576f6-0288-4d78-9bb8-99f238526548: nodetool_command=flush node=longevity-100gb-4h-4-6-db-node-d7a36396-1

node 1 log has system_keyspace - update compaction history failed: exceptions::mutation_write_timeout_exception

7ad54392bcdd35a684174e047860b377/md-66-big-Data.db:level=0:origin=memtable,/var/lib/scylla/data/system/local-7ad54392bcdd35a684174e047860b377/md-58-big-Data.db:level=0:origin=compaction]
Jan 30 14:55:29 longevity-100gb-4h-4-6-db-node-d7a36396-1 scylla[10239]:  [shard 4] system_keyspace - update compaction history failed: exceptions::mutation_write_timeout_exception (Operation timed out for system.compaction_history - received only 0 responses from 1 CL=ONE.): ignored
Jan 30 14:55:30 longevity-100gb-4h-4-6-db-node-d7a36396-1 scylla[10239]:  [shard 2] compaction - [Compact system.local a9ecf630-81dc-11ec-89ad-0774a9c82ea7] Compacted 2 sstables to [/var/lib/scylla/data/system/local-7ad54392bcdd35a684174e047860b377/md-74-big-Data.db:level=0]. 18kB to 12kB (~66% of original) in 451ms = 27kB/s. ~256 total partitions merged to 1.

cassandra-stress creates keyspace

an 30 14:59:59 longevity-100gb-4h-4-6-db-node-d7a36396-1 scylla[10239]:  [shard 7] migration_manager - Create new Keyspace: KSMetaData{name=drop_table_during_repair_ks_1, strategyClass=org.apache.cassandra.locator.SimpleStrategy, strategyOptions={replication_factor=6}, cfMetaData={}, durable_writes=1, userTypes=org.apache.cassandra.config.UTMetaData@0x60700b473a38}
Jan 30 15:00:02 longevity-100gb-4h-4-6-db-node-d7a36396-1 scylla[10239]:  [shard 0] migration_manager - Requesting schema pull from 10.0.2.133:0
Jan 30 15:00:02 longevity-100gb-4h-4-6-db-node-d7a36396-1 scylla[10239]:  [shard 0] migration_manager - Pulling schema from 10.0.2.133:0
Jan 30 15:00:02 longevity-100gb-4h-4-6-db-node-d7a36396-1 scylla[10239]:  [shard 0] migration_manager - Requesting schema pull from 10.0.0.196:0
Jan 30 15:00:02 longevity-100gb-4h-4-6-db-node-d7a36396-1 scylla[10239]:  [shard 0] migration_manager - Pulling schema from 10.0.0.196:0
Jan 30 15:00:02 longevity-100gb-4h-4-6-db-node-d7a36396-1 scylla[10239]:  [shard 0] migration_manager - Requesting schema pull from 10.0.0.16:0
Jan 30 15:00:02 longevity-100gb-4h-4-6-db-node-d7a36396-1 scylla[10239]:  [shard 0] migration_manager - Pulling schema from 10.0.0.16:0
Jan 30 15:00:02 longevity-100gb-4h-4-6-db-node-d7a36396-1 scylla[10239]:  [shard 0] migration_manager - Requesting schema pull from 10.0.3.70:0
Jan 30 15:00:02 longevity-100gb-4h-4-6-db-node-d7a36396-1 scylla[10239]:  [shard 0] migration_manager - Pulling schema from 10.0.3.70:0
Jan 30 15:00:03 longevity-100gb-4h-4-6-db-node-d7a36396-1 scylla[10239]:  [shard 0] compaction - [Compact system_schema.keyspaces 4db16670-81dd-11ec-a236-0771a9c82ea7] Compacting [/var/lib/scylla/data/system_schema/keyspaces-abac5682dea631c5b535b3d6cffd0fb6/md-304-big-Data.db:level=0:origin=memtable,/var/lib/scylla/data/system_schema/keyspaces-abac5682dea631c5b535b3d6cffd0fb6/md-296-big-Data.db:level=0:origin=compaction]
Jan 30 15:00:03 longevity-100gb-4h-4-6-db-node-d7a36396-1 scylla[10239]:  [shard 7] schema_tables - Creating keyspace drop_table_during_repair_ks_1
Jan 30 15:00:03 longevity-100gb-4h-4-6-db-node-d7a36396-1 scylla[10239]:  [shard 7] schema_tables - Schema version changed to 447ae749-4351-3a2c-bf7f-4605138b63bb
Jan 30 15:00:04 longevity-100gb-4h-4-6-db-node-d7a36396-1 scylla[10239]:  [shard 0] compaction - [Compact system_schema.keyspaces 4db16670-81dd-11ec-a236-0771a9c82ea7] Compacted 2 sstables to [/var/lib/scylla/data/system_schema/keyspaces-abac5682dea631c5b535b3d6cffd0fb6/md-312-big-Data.db:level=0]. 11kB to 6kB (~53% of original) in 96ms = 65kB/s. ~256 total partitions merged to 9.
Jan 30 15:00:04 longevity-100gb-4h-4-6-db-node-d7a36396-1 scylla[10239]:  [shard 0] migration_manager - Requesting schema pull from 10.0.3.124:0
Jan 30 15:00:04 longevity-100gb-4h-4-6-db-node-d7a36396-1 scylla[10239]:  [shard 0] migration_manager - Pulling schema from 10.0.3.124:0
Jan 30 15:00:06 longevity-100gb-4h-4-6-db-node-d7a36396-1 scylla[10239]:  [shard 0] compaction - [Compact system_schema.columns 4fd4e2b0-81dd-11ec-a236-0771a9c82ea7] Compacting [/var/lib/scylla/data/system_schema/columns-24101c25a2ae3af787c1b40ee1aca33f/md-160-big-Data.db:level=0:origin=memtable,/var/lib/scylla/data/system_schema/columns-24101c25a2ae3af787c1b40ee1aca33f/md-152-big-Data.db:level=0:origin=compaction]
Jan 30 15:00:08 longevity-100gb-4h-4-6-db-node-d7a36396-1 scylla[10239]:  [shard 0] compaction - [Compact system_schema.columns 4fd4e2b0-81dd-11ec-a236-0771a9c82ea7] Compacted 2 sstables to [/var/lib/scylla/data/system_schema/columns-24101c25a2ae3af787c1b40ee1aca33f/md-168-big-Data.db:level=0]. 27kB to 18kB (~67% of original) in 96ms = 189kB/s. ~256 total partitions merged to 8.
Jan 30 15:00:08 longevity-100gb-4h-4-6-db-node-d7a36396-1 scylla[10239]:  [shard 0] schema_tables - Schema version changed to 447ae749-4351-3a2c-bf7f-4605138b63bb
Jan 30 15:00:08 longevity-100gb-4h-4-6-db-node-d7a36396-1 scylla[10239]:  [shard 0] migration_manager - Schema merge with 10.0.0.196:0 completed
Jan 30 15:00:09 longevity-100gb-4h-4-6-db-node-d7a36396-1 scylla[10239]:  [shard 0] system_keyspace - update compaction history failed: exceptions::mutation_write_timeout_exception (Operation timed out for system.compaction_history - received only 0 responses from 1 CL=ONE.): ignored
Jan 30 15:00:09 longevity-100gb-4h-4-6-db-node-d7a36396-1 scylla[10239]:  [shard 0] compaction - [Compact system_schema.tables 5191ade0-81dd-11ec-a236-0771a9c82ea7] Compacting [/var/lib/scylla/data/system_schema/tables-afddfb9dbc1e30688056eed6c302ba09/md-192-big-Data.db:level=0:origin=memtable,/var/lib/scylla/data/system_schema/tables-afddfb9dbc1e30688056eed6c302ba09/md-184-big-Data.db:level=0:origin=compaction]
Jan 30 15:00:11 longevity-100gb-4h-4-6-db-node-d7a36396-1 scylla[10239]:  [shard 0] compaction - [Compact system_schema.tables 5191ade0-81dd-11ec-a236-0771a9c82ea7] Compacted 2 sstables to [/var/lib/scylla/data/system_schema/tables-afddfb9dbc1e30688056eed6c302ba09/md-208-big-Data.db:level=0]. 23kB to 13kB (~60% of original) in 100ms = 139kB/s. ~256 total partitions merged to 8.
Jan 30 15:00:11 longevity-100gb-4h-4-6-db-node-d7a36396-1 scylla[10239]:  [shard 0] system_keyspace - update compaction history failed: exceptions::mutation_write_timeout_exception (Operation timed out for system.compaction_history - received only 0 responses from 1 CL=ONE.): ignored

starting to get seastar - Exceptional future ignored: exceptions::mutation_write_timeout_exception

Jan 30 15:00:47 longevity-100gb-4h-4-6-db-node-d7a36396-1 scylla[10239]:  [shard 0] compaction - [Compact system_schema.dropped_columns 6754f7e0-81dd-11ec-a236-0771a9c82ea7] Compacted 2 sstables to [/var/lib/scylla/data/system_schema/dropped_columns-5e7583b5f3f43af19a39b7e1d6f5f11f/md-192-big-Data.db:level=0]. 11kB to 5kB (~51% of original) in 90ms = 63kB/s. ~256 total partitions merged to 2.
Jan 30 15:00:48 longevity-100gb-4h-4-6-db-node-d7a36396-1 scylla[10239]:  [shard 0] system_keyspace - update compaction history failed: exceptions::mutation_write_timeout_exception (Operation timed out for system.compaction_history - received only 0 responses from 1 CL=ONE.): ignored
Jan 30 15:00:48 longevity-100gb-4h-4-6-db-node-d7a36396-1 scylla[10239]:  [shard 0] compaction - [Compact system_schema.scylla_tables 68ede760-81dd-11ec-a236-0771a9c82ea7] Compacting [/var/lib/scylla/data/system_schema/scylla_tables-5d912ff1f7593665b2c88042ab5103dd/md-256-big-Data.db:level=0:origin=memtable,/var/lib/scylla/data/system_schema/scylla_tables-5d912ff1f7593665b2c88042ab5103dd/md-240-big-Data.db:level=0:origin=memtable,/var/lib/scylla/data/system_schema/scylla_tables-5d912ff1f7593665b2c88042ab5103dd/md-248-big-Data.db:level=0:origin=compaction]
Jan 30 15:00:49 longevity-100gb-4h-4-6-db-node-d7a36396-1 scylla[10239]:  [shard 0] migration_manager - Pulling schema from 10.0.0.16:0
Jan 30 15:00:49 longevity-100gb-4h-4-6-db-node-d7a36396-1 scylla[10239]:  [shard 1] seastar - Exceptional future ignored: exceptions::mutation_write_timeout_exception (Operation timed out for system.local - received only 0 responses from 1 CL=ONE.), backtrace: 0x4154b1e 0x4155010 0x4155318 0x3d861d5 0x12caa7c 0x3dc4354 0x3dc5747 0x3de2d75 0x3d9c98a /opt/scylladb/libreloc/libpthread.so.0+0x9298 /opt/scylladb/libreloc/libc.so.6+0x100352
                                                                            --------
                                                                            seastar::continuation<seastar::internal::promise_base_with_type<void>, schema_registry_entry::maybe_sync(std::function<seastar::future<void> ()>)::$_3, seastar::future<void>::then_wrapped_nrvo<void, schema_registry_entry::maybe_sync(std::function<seastar::future<void> ()>)::$_3>(schema_registry_entry::maybe_sync(std::function<seastar::future<void> ()>)::$_3&&)::{lambda(seastar::internal::promise_base_with_type<void>&&, schema_registry_entry::maybe_sync(std::function<seastar::future<void> ()>)::$_3&, seastar::future_state<seastar::internal::monostate>&&)#1}, void>
Jan 30 15:00:51 longevity-100gb-4h-4-6-db-node-d7a36396-1 scylla[10239]:  [shard 0] migration_manager - Pulling schema from 10.0.0.196:0
Jan 30 15:00:51 longevity-100gb-4h-4-6-db-node-d7a36396-1 scylla[10239]:  [shard 5] seastar - Exceptional future ignored: seastar::timed_out_error (timedout), backtrace: 0x4154b1e 0x4155010 0x4155318 0x3d861d5 0x12caa7c 0x3dc4354 0x3dc5747 0x3de2d75 0x3d9c98a /opt/scylladb/libreloc/libpthread.so.0+0x9298 /opt/scylladb/libreloc/libc.so.6+0x100352
                                                                            --------
                                                                            seastar::continuation<seastar::internal::promise_base_with_type<void>, schema_registry_entry::maybe_sync(std::function<seastar::future<void> ()>)::$_3, seastar::future<void>::then_wrapped_nrvo<void, schema_registry_entry::maybe_sync(std::function<seastar::future<void> ()>)::$_3>(schema_registry_entry::maybe_sync(std::function<seastar::future<void> ()>)::$_3&&)::{lambda(seastar::internal::promise_base_with_type<void>&&, schema_registry_entry::maybe_sync(std::function<seastar::future<void> ()>)::$_3&, seastar::future_state<seastar::internal::monostate>&&)#1}, void>
Jan 30 15:00:52 longevity-100gb-4h-4-6-db-node-d7a36396-1 scylla[10239]:  [shard 0] seastar - Exceptional future ignored: seastar::timed_out_error (timedout), backtrace: 0x4154b1e 0x4155010 0x4155318 0x3d861d5 0x12caa7c 0x3dc4354 0x3dc5747 0x3dc498c 0x3d6e21e 0x3d6d596 0xf402f5 /opt/scylladb/libreloc/libc.so.6+0x27b74 0xf3d4ad
                                                                            --------
                                                                            seastar::continuation<seastar::internal::promise_base_with_type<void>, schema_registry_entry::maybe_sync(std::function<seastar::future<void> ()>)::$_3, seastar::future<void>::then_wrapped_nrvo<void, schema_registry_entry::maybe_sync(std::function<seastar::future<void> ()>)::$_3>(schema_registry_entry::maybe_sync(std::function<seastar::future<void> ()>)::$_3&&)::{lambda(seastar::internal::promise_base_with_type<void>&&, schema_registry_entry::maybe_sync(std::function<seastar::future<void> ()>)::$_3&, seastar::future_state<seastar::internal::monostate>&&)#1}, void>
Jan 30 15:00:53 longevity-100gb-4h-4-6-db-node-d7a36396-1 scylla[10239]:  [shard 0] compaction - [Compact system_schema.scylla_tables 68ede760-81dd-11ec-a236-0771a9c82ea7] Compacted 3 sstables to [/var/lib/scylla/data/system_schema/scylla_tables-5d912ff1f7593665b2c88042ab5103dd/md-264-big-Data.db:level=0]. 20kB to 8kB (~39% of original) in 405ms = 20kB/s. ~384 total partitions merged to 9.

storage_service - fail to update schema_version for

Jan 30 15:00:53 longevity-100gb-4h-4-6-db-node-d7a36396-1 scylla[10239]:  [shard 0] storage_service - fail to update schema_version for 10.0.3.124: exceptions::mutation_write_timeout_exception (Operation timed out for system.peers - received only 0 responses from 1 CL=ONE.)
Jan 30 15:00:54 longevity-100gb-4h-4-6-db-node-d7a36396-1 scylla[10239]:  [shard 0] system_keyspace - update compaction history failed: exceptions::mutation_write_timeout_exception (Operation timed out for system.compaction_history - received only 0 responses from 1 CL=ONE.): ignored
Jan 30 15:00:54 longevity-100gb-4h-4-6-db-node-d7a36396-1 scylla[10239]:  [shard 0] storage_service - Failed to pull schema from 10.0.3.124: seastar::timed_out_error (timedout)
Jan 30 15:00:54 longevity-100gb-4h-4-6-db-node-d7a36396-1 scylla[10239]:  [shard 0] compaction - [Compact system_schema.keyspaces 6c466130-81dd-11ec-a236-0771a9c82ea7] Compacting [/var/lib/scylla/data/system_schema/keyspaces-abac5682dea631c5b535b3d6cffd0fb6/md-376-big-Data.db:level=0:origin=memtable,/var/lib/scylla/data/system_schema/keyspaces-abac5682dea631c5b535b3d6cffd0fb6/md-360-big-Data.db:level=0:origin=memtable,/var/lib/scylla/data/system_schema/keyspaces-abac5682dea631c5b535b3d6cffd0fb6/md-392-big-Data.db:level=0:origin=memtable,/var/lib/scylla/data/system_schema/keyspaces-abac5682dea631c5b535b3d6cffd0fb6/md-384-big-Data.db:level=0:origin=memtable,/var/lib/scylla/data/system_schema/keyspaces-abac5682dea631c5b535b3d6cffd0fb6/md-368-big-Data.db:level=0:origin=compaction]
Jan 30 15:00:57 longevity-100gb-4h-4-6-db-node-d7a36396-1 scylla[10239]:  [shard 6] seastar - Exceptional future ignored: seastar::timed_out_error (timedout), backtrace: 0x4154b1e 0x4155010 0x4155318 0x3d861d5 0x12caa7c 0x3dc4354 0x3dc5747 0x3de2d75 0x3d9c98a /opt/scylladb/libreloc/libpthread.so.0+0x9298 /opt/scylladb/libreloc/libc.so.6+0x100352
                                                                            --------
                                                                            seastar::continuation<seastar::internal::promise_base_with_type<void>, schema_registry_entry::maybe_sync(std::function<seastar::future<void> ()>)::$_3, seastar::future<void>::then_wrapped_nrvo<void, schema_registry_entry::maybe_sync(std::function<seastar::future<void> ()>)::$_3>(schema_registry_entry::maybe_sync(std::function<seastar::future<void> ()>)::$_3&&)::{lambda(seastar::internal::promise_base_with_type<void>&&, schema_registry_entry::maybe_sync(std::function<seastar::future<void> ()>)::$_3&, seastar::future_state<seastar::internal::monostate>&&)#1}, void>
Jan 30 15:01:09 longevity-100gb-4h-4-6-db-node-d7a36396-1 scylla[10239]:  [shard 0] compaction - [Compact system_schema.keyspaces 6c466130-81dd-11ec-a236-0771a9c82ea7] Compacted 5 sstables to [/var/lib/s
Jan 30 15:01:15 longevity-100gb-4h-4-6-db-node-d7a36396-1 scylla[10239]:  [shard 4] seastar - Exceptional future ignored: exceptions::mutation_write_timeout_exception (Operation timed out for system.clie
nts - received only 0 responses from 1 CL=ONE.), backtrace: 0x4154b1e 0x4155010 0x4155318 0x3d861d5 0x1277cba 0x3dc4354 0x3dc5747 0x3de2d75 0x3d9c98a /opt/scylladb/libreloc/libpthread.so.0+0x9298 /opt/sc
ylladb/libreloc/libc.so.6+0x100352

translated:

void seastar::backtrace<seastar::current_backtrace_tasklocal()::$_3>(seastar::current_backtrace_tasklocal()::$_3&&) at ./build/release/seastar/./seastar/include/seastar/util/backtrace.hh:59
 (inlined by) seastar::current_backtrace_tasklocal() at ./build/release/seastar/./seastar/src/util/backtrace.cc:86
seastar::current_tasktrace() at ./build/release/seastar/./seastar/src/util/backtrace.cc:137
seastar::current_backtrace() at ./build/release/seastar/./seastar/src/util/backtrace.cc:170
seastar::report_failed_future(std::__exception_ptr::exception_ptr const&) at ./build/release/seastar/./seastar/src/core/future.cc:210
 (inlined by) seastar::report_failed_future(seastar::future_state_base::any&&) at ./build/release/seastar/./seastar/src/core/future.cc:218
seastar::future_state_base::any::check_failure() at ././seastar/include/seastar/core/future.hh:567
 (inlined by) seastar::future_state<seastar::shared_ptr<cql3::untyped_result_set> >::clear() at ././seastar/include/seastar/core/future.hh:609
 (inlined by) ~future_state at ././seastar/include/seastar/core/future.hh:614
 (inlined by) ~continuation_base at ././seastar/include/seastar/core/future.hh:697
 (inlined by) ~continuation_base_with_promise at ././seastar/include/seastar/core/future.hh:735
 (inlined by) _ZN7seastar12continuationINS_8internal22promise_base_with_typeIvEEZNS_6futureINS_10shared_ptrIN4cql318untyped_result_setEEEE14discard_resultEvEUlDpOT_E_ZNS9_14then_impl_nrvoISD_NS4_IvEEEET0_OT_EUlOS3_RSD_ONS_12future_stateIS8_EEE_S8_E15run_and_disposeEv at ././seastar/include/seastar/core/future.hh:771
seastar::reactor::run_tasks(seastar::reactor::task_queue&) at ./build/release/seastar/./seastar/src/core/reactor.cc:2374
 (inlined by) seastar::reactor::run_some_tasks() at ./build/release/seastar/./seastar/src/core/reactor.cc:2783
seastar::reactor::do_run() at ./build/release/seastar/./seastar/src/core/reactor.cc:2952
operator() at ./build/release/seastar/./seastar/src/core/reactor.cc:4166
 (inlined by) void std::__invoke_impl<void, seastar::smp::configure(boost::program_options::variables_map, seastar::reactor_config)::$_98&>(std::__invoke_other, seastar::smp::configure(boost::program_options::variables_map, seastar::reactor_config)::$_98&) at /usr/lib/gcc/x86_64-redhat-linux/11/../../../../include/c++/11/bits/invoke.h:61
 (inlined by) std::enable_if<is_invocable_r_v<void, seastar::smp::configure(boost::program_options::variables_map, seastar::reactor_config)::$_98&>, void>::type std::__invoke_r<void, seastar::smp::configure(boost::program_options::variables_map, seastar::reactor_config)::$_98&>(seastar::smp::configure(boost::program_options::variables_map, seastar::reactor_config)::$_98&) at /usr/lib/gcc/x86_64-redhat-linux/11/../../../../include/c++/11/bits/invoke.h:111
 (inlined by) std::_Function_handler<void (), seastar::smp::configure(boost::program_options::variables_map, seastar::reactor_config)::$_98>::_M_invoke(std::_Any_data const&) at /usr/lib/gcc/x86_64-redhat-linux/11/../../../../include/c++/11/bits/std_function.h:291
std::function<void ()>::operator()() const at /usr/lib/gcc/x86_64-redhat-linux/11/../../../../include/c++/11/bits/std_function.h:560
 (inlined by) seastar::posix_thread::start_routine(void*) at ./build/release/seastar/./seastar/src/core/posix.cc:60
?? ??:0
?? ??:0


                                                                            --------
                                                                            seastar::continuation_base<seastar::shared_ptr<cql3::untyped_result_set> >
Jan 30 15:01:15 longevity-100gb-4h-4-6-db-node-d7a36396-1 scylla[10239]:  [shard 2] seastar - Exceptional future ignored: exceptions::mutation_write_timeout_exception (Operation timed out for system.clie
nts - received only 0 responses from 1 CL=ONE.), backtrace: 0x4154b1e 0x4155010 0x4155318 0x3d861d5 0x1277cba 0x3dc4354 0x3dc5747 0x3de2d75 0x3d9c98a /opt/scylladb/libreloc/libpthread.so.0+0x9298 /opt/sc
ylladb/libreloc/libc.so.6+0x100352
                                                                            --------
                                                                            seastar::continuation_base<seastar::shared_ptr<cql3::untyped_result_set> >
Jan 30 15:01:15 longevity-100gb-4h-4-6-db-node-d7a36396-1 scylla[10239]:  [shard 3] seastar - Exceptional future ignored: exceptions::mutation_write_timeout_exception (Operation timed out for system.clie
nts - received only 0 responses from 1 CL=ONE.), backtrace: 0x4154b1e 0x4155010 0x4155318 0x3d8612b 0x1277c6e 0x3dc4354 0x3dc5747 0x3de2d75 0x3d9c98a /opt/scylladb/libreloc/libpthread.so.0+0x9298 /opt/sc
ylladb/libreloc/libc.so.6+0x100352
                                                                            --------
                                                                            N7seastar12continuationINS_8internal22promise_base_with_typeIvEEZNS_6futureINS_10shared_ptrIN4cql318untyped_result_setEEEE14dis
card_resultEvEUlDpOT_E_ZNS9_14then_impl_nrvoISD_NS4_IvEEEET0_OT_EUlOS3_RSD_ONS_12future_stateIS8_EEE_S8_EE
Jan 30 15:01:15 longevity-100gb-4h-4-6-db-node-d7a36396-1 scylla[10239]:  [shard 3] seastar - Exceptional future ignored: exceptions::mutation_write_timeout_exception (Operation timed out for system.clie
nts - received only 0 responses from 1 CL=ONE.), backtrace: 0x4154b1e 0x4155010 0x4155318 0x3d861d5 0x1277cba 0x3dc4354 0x3dc5747 0x3de2d75 0x3d9c98a /opt/scylladb/libreloc/libpthread.so.0+0x9298 /opt/sc
ylladb/libreloc/libc.so.6+0x100352
                                                                            --------
                                                                            seastar::continuation_base<seastar::shared_ptr<cql3::untyped_result_set> >
Jan 30 15:01:18 longevity-100gb-4h-4-6-db-node-d7a36396-1 scylla[10239]:  [shard 4] cql_server - exception while advertising new connection: exceptions::mutation_write_timeout_exception (Operation timed 
out for system.clients - received only 0 responses from 1 CL=ONE.)
Jan 30 15:01:18 longevity-100gb-4h-4-6-db-node-d7a36396-1 scylla[10239]:  [shard 4] seastar - Exceptional future ignored: exceptions::mutation_write_timeout_exception (Operation timed out for system.loca
l - received only 0 responses from 1 CL=ONE.), backtrace: 0x4154b1e 0x4155010 0x4155318 0x3d861d5 0x12caa7c 0x3dc4354 0x3dc5747 0x3de2d75 0x3d9c98a /opt/scylladb/libreloc/libpthread.so.0+0x9298 /opt/scyl
ladb/libreloc/libc.so.6+0x100352
                                                                            --------
                                                                            seastar::continuation<seastar::internal::promise_base_with_type<void>, schema_registry_entry::maybe_sync(std::function<seastar:
:future<void> ()>)::$_3, seastar::future<void>::then_wrapped_nrvo<void, schema_registry_entry::maybe_sync(std::function<seastar::future<void> ()>)::$_3>(schema_registry_entry::maybe_sync(std::function<se
astar::future<void> ()>)::$_3&&)::{lambda(seastar::internal::promise_base_with_type<void>&&, schema_registry_entry::maybe_sync(std::function<seastar::future<void> ()>)::$_3&, seastar::future_state<seasta
r::internal::monostate>&&)#1}, void>
Jan 30 15:01:20 longevity-100gb-4h-4-6-db-node-d7a36396-1 scylla[10239]:  [shard 4] cql_server - exception while setting protocol_version in `system.clients`: exceptions::mutation_write_timeout_exception
 (Operation timed out for system.clients - received only 0 responses from 1 CL=ONE.)
Jan 30 15:01:20 longevity-100gb-4h-4-6-db-node-d7a36396-1 scylla[10239]:  [shard 3] seastar - Exceptional future ignored: seastar::timed_out_error (timedout), backtrace: 0x4154b1e 0x4155010 0x4155318 0x3
d861d5 0x12caa7c 0x3dc4354 0x3dc5747 0x3de2d75 0x3d9c98a /opt/scylladb/libreloc/libpthread.so.0+0x9298 /opt/scylladb/libreloc/libc.so.6+0x100352
                                                                            --------
                                                                            seastar::continuation<seastar::internal::promise_base_with_type<void>, schema_registry_entry::maybe_sync(std::function<seastar:
:future<void> ()>)::$_3, seastar::future<void>::then_wrapped_nrvo<void, schema_registry_entry::maybe_sync(std::function<seastar::future<void> ()>)::$_3>(schema_registry_entry::maybe_sync(std::function<se
astar::future<void> ()>)::$_3&&)::{lambda(seastar::internal::promise_base_with_type<void>&&, schema_registry_entry::maybe_sync(std::function<seastar::future<void> ()>)::$_3&, seastar::future_state<seasta
r::internal::monostate>&&)#1}, void>
Jan 30 15:01:20 longevity-100gb-4h-4-6-db-node-d7a36396-1 scylla[10239]:  [shard 2] seastar - Exceptional future ignored: seastar::timed_out_error (timedout), backtrace: 0x4154b1e 0x4155010 0x4155318 0x3
d861d5 0x12caa7c 0x3dc4354 0x3dc5747 0x3de2d75 0x3d9c98a /opt/scylladb/libreloc/libpthread.so.0+0x9298 /opt/scylladb/libreloc/libc.so.6+0x100352
                                                                            --------
                                                                            seastar::continuation<seastar::internal::promise_base_with_type<void>, schema_registry_entry::maybe_sync(std::function<seastar:
:future<void> ()>)::$_3, seastar::future<void>::then_wrapped_nrvo<void, schema_registry_entry::maybe_sync(std::function<seastar::future<void> ()>)::$_3>(schema_registry_entry::maybe_sync(std::function<se
astar::future<void> ()>)::$_3&&)::{lambda(seastar::internal::promise_base_with_type<void>&&, schema_registry_entry::maybe_sync(std::function<seastar::future<void> ()>)::$_3&, seastar::future_state<seasta
r::internal::monostate>&&)#1}, void>

====================================

Restore Monitor Stack command: $ hydra investigate show-monitor d7a36396-5fb2-4aa9-9242-f4733f410343
Restore monitor on AWS instance using Jenkins job
Show all stored logs command: $ hydra investigate show-logs d7a36396-5fb2-4aa9-9242-f4733f410343

Test id: d7a36396-5fb2-4aa9-9242-f4733f410343

Logs:
grafana - https://cloudius-jenkins-test.s3.amazonaws.com/d7a36396-5fb2-4aa9-9242-f4733f410343/20220130_154741/grafana-screenshot-longevity-100gb-4h-ebs-gp3-test-scylla-per-server-metrics-nemesis-20220130_154920-longevity-100gb-4h-4-6-monitor-node-d7a36396-1.png
grafana - https://cloudius-jenkins-test.s3.amazonaws.com/d7a36396-5fb2-4aa9-9242-f4733f410343/20220130_154741/grafana-screenshot-overview-20220130_154741-longevity-100gb-4h-4-6-monitor-node-d7a36396-1.png
db-cluster - https://cloudius-jenkins-test.s3.amazonaws.com/d7a36396-5fb2-4aa9-9242-f4733f410343/20220130_155520/db-cluster-d7a36396.tar.gz
loader-set - https://cloudius-jenkins-test.s3.amazonaws.com/d7a36396-5fb2-4aa9-9242-f4733f410343/20220130_155520/loader-set-d7a36396.tar.gz
monitor-set - https://cloudius-jenkins-test.s3.amazonaws.com/d7a36396-5fb2-4aa9-9242-f4733f410343/20220130_155520/monitor-set-d7a36396.tar.gz
sct - https://cloudius-jenkins-test.s3.amazonaws.com/d7a36396-5fb2-4aa9-9242-f4733f410343/20220130_155520/sct-runner-d7a36396.tar.gz

Jenkins job URL

@slivne
Copy link
Contributor

slivne commented Jan 31, 2022

Bentsi shared this as well

image (14)

I am not sure this is not worse then the reported issue above - the spike is lower yet the commiltog failed writes persistent failure

@elcallio
Copy link
Contributor

I realize it is a little late to mention this, but the logs say:
commitlog - Exception in segment reservation: storage_io_error (Storage I/O error: 28: No space left on device

The disks are literally full. It is perhaps not strange it stalls...?

@slivne
Copy link
Contributor

slivne commented Feb 2, 2022

Quoting bentsi

The out of disk space happens after - the schema disagreement issues - so the question stands as to why we have that.

The cause of out of diskpace is a different nemesis that is inflicting this on purpose and trying to see how we handle that

In any case this is a new workload with alower disk so we are not blocking on this 4.6 for now

@bentsi
Copy link
Contributor

bentsi commented Feb 6, 2022

happened again on a different run

Installation details
Kernel version: 5.11.0-1022-aws
Scylla version (or git commit hash): 4.6.rc5-0.20220203.5694ec189 with build-id f5d85bf5abe6d2f9fd3487e2469ce1c34304cc14
Cluster size: 4 nodes (i3.4xlarge)
Scylla running with shards number (live nodes):
longevity-tls-1tb-7d-4-6-db-node-f23d21db-1 (54.73.108.182 | 10.0.1.223): 14 shards
longevity-tls-1tb-7d-4-6-db-node-f23d21db-2 (18.203.172.248 | 10.0.0.219): 14 shards
longevity-tls-1tb-7d-4-6-db-node-f23d21db-3 (54.195.173.211 | 10.0.3.7): 14 shards
longevity-tls-1tb-7d-4-6-db-node-f23d21db-4 (54.194.126.137 | 10.0.2.138): 14 shards
OS (RHEL/CentOS/Ubuntu/AWS AMI): ami-06df116ea1c70171c (aws: eu-west-1)

Test: longevity-1tb-7days-test
Test name: longevity_test.LongevityTest.test_custom_time
Test config file(s):

  • longevity-1TB-5days-authorization-and-tls-ssl.yaml

Issue description

====================================

node1 is reporting a schema disagreement

< t:2022-02-04 15:40:26,539 f:thread.py       l:52   c:cassandra.cluster    p:WARNING > Node 10.0.1.223:9042 is reporting a schema disagreement: {UUID('494804d0-200a-3002-bd2a-d2b1710de292'): [<DefaultEndPoint: 10.0.2.138:9042>, <DefaultEndPoint: 10.0.0.219:9042>, <DefaultEndPoint: 10.0.1.223:9042>], UUID('26d4f6fb-9c45-3b74-82d3-d3343139a7de'): [<DefaultEndPoint: 10.0.3.7:9042>]}
< t:2022-02-04 15:40:26,539 f:thread.py       l:52   c:cassandra.cluster    p:DEBUG > Skipping schema refresh due to lack of schema agreement
< t:2022-02-04 15:40:26,539 f:thread.py       l:52   c:cassandra.cluster    p:DEBUG > [control connection] Waiting for schema agreement
< t:2022-02-04 15:40:26,553 f:thread.py       l:52   c:cassandra.cluster    p:DEBUG > [control connection] Schemas mismatched, trying again
< t:2022-02-04 15:40:29,351 f:base.py         l:146  c:RemoteCmdRunner      p:ERROR > Error executing command: "echo TAG: loader_idx:0-cpu_idx:0-keyspace_idx:1; STRESS_TEST_MARKER=QWEBAP1F35Z0DTRVSGS0; cassandra-stress user profile=/tmp/cs_mv_prof
ile.yaml ops'(insert=3,read1=1,read2=1,read3=1)' cl=QUORUM duration=6800m -port jmx=6868 -mode cql3 native  user=cassandra password=cassandra -rate threads=10 -transport "truststore=/etc/scylla/ssl_conf/client/cacerts.jks truststore-password=cassandra" -node 10.0.1.223 -errors skip-unsupported-columns"; Exit status: 1
< t:2022-02-04 15:40:29,352 f:base.py         l:148  c:RemoteCmdRunner      p:DEBUG > STDOUT: - Connection[/10.0.3.7:9042-15, inFlight=1, closed=false] Setting keyspace mview
< t:2022-02-04 15:40:29,352 f:base.py         l:148  c:RemoteCmdRunner      p:DEBUG > 15:40:29.009 [cluster1-nio-worker-8] DEBUG com.datastax.driver.core.Connection - Connection[/10.0.3.7:9042-15, inFlight=1, closed=false] Keyspace set to mview
< t:2022-02-04 15:40:29,352 f:base.py         l:150  c:RemoteCmdRunner      p:DEBUG > STDERR: Failed to connect over JMX; not collecting these stats
< t:2022-02-04 15:40:29,352 f:base.py         l:150  c:RemoteCmdRunner      p:DEBUG > Failed to connect over JMX; not collecting these stats
< t:2022-02-04 15:40:29,352 f:base.py         l:150  c:RemoteCmdRunner      p:DEBUG > com.datastax.driver.core.exceptions.InvalidQueryException: unconfigured table users
< t:2022-02-04 15:40:29,352 f:base.py         l:150  c:RemoteCmdRunner      p:DEBUG >   at com.datastax.driver.core.exceptions.InvalidQueryException.copy(InvalidQueryException.java:50)
< t:2022-02-04 15:40:29,352 f:base.py         l:150  c:RemoteCmdRunner      p:DEBUG >   at com.datastax.driver.core.DriverThrowables.propagateCause(DriverThrowables.java:35)
< t:2022-02-04 15:40:29,352 f:base.py         l:150  c:RemoteCmdRunner      p:DEBUG >   at com.datastax.driver.core.AbstractSession.prepare(AbstractSession.java:86)
< t:2022-02-04 15:40:29,352 f:base.py         l:150  c:RemoteCmdRunner      p:DEBUG >   at org.apache.cassandra.stress.util.JavaDriverClient.prepare(JavaDriverClient.java:116)
< t:2022-02-04 15:40:29,352 f:base.py         l:150  c:RemoteCmdRunner      p:DEBUG >   at org.apache.cassandra.stress.StressProfile.getQuery(StressProfile.java:387)
< t:2022-02-04 15:40:29,352 f:base.py         l:150  c:RemoteCmdRunner      p:DEBUG >   at org.apache.cassandra.stress.settings.SettingsCommandUser$1.get(SettingsCommandUser.java:93)
< t:2022-02-04 15:40:29,352 f:base.py         l:150  c:RemoteCmdRunner      p:DEBUG >   at org.apache.cassandra.stress.settings.SettingsCommandUser$1.get(SettingsCommandUser.java:82)
< t:2022-02-04 15:40:29,352 f:base.py         l:150  c:RemoteCmdRunner      p:DEBUG >   at org.apache.cassandra.stress.operations.SampledOpDistributionFactory$1.get(SampledOpDistributionFactory.java:83)
< t:2022-02-04 15:40:29,352 f:base.py         l:150  c:RemoteCmdRunner      p:DEBUG >   at org.apache.cassandra.stress.StressAction$Consumer.<init>(StressAction.java:410)
< t:2022-02-04 15:40:29,352 f:base.py         l:150  c:RemoteCmdRunner      p:DEBUG >   at org.apache.cassandra.stress.StressAction.run(StressAction.java:234)
< t:2022-02-04 15:40:29,352 f:base.py         l:150  c:RemoteCmdRunner      p:DEBUG >   at org.apache.cassandra.stress.StressAction.warmup(StressAction.java:122)
< t:2022-02-04 15:40:29,352 f:base.py         l:150  c:RemoteCmdRunner      p:DEBUG >   at org.apache.cassandra.stress.StressAction.run(StressAction.java:71)
< t:2022-02-04 15:40:29,352 f:base.py         l:150  c:RemoteCmdRunner      p:DEBUG >   at org.apache.cassandra.stress.Stress.run(Stress.java:143)
< t:2022-02-04 15:40:29,352 f:base.py         l:150  c:RemoteCmdRunner      p:DEBUG >   at org.apache.cassandra.stress.Stress.main(Stress.java:62)
< t:2022-02-04 15:40:29,352 f:base.py         l:150  c:RemoteCmdRunner      p:DEBUG > Caused by: com.datastax.driver.core.exceptions.InvalidQueryException: unconfigured table users
< t:2022-02-04 15:40:29,352 f:base.py         l:150  c:RemoteCmdRunner      p:DEBUG >   at com.datastax.driver.core.Responses$Error.asException(Responses.java:181)
< t:2022-02-04 15:40:29,352 f:base.py         l:150  c:RemoteCmdRunner      p:DEBUG >   at com.datastax.driver.core.SessionManager$4.apply(SessionManager.java:253)
< t:2022-02-04 15:40:29,352 f:base.py         l:150  c:RemoteCmdRunner      p:DEBUG >   at com.datastax.driver.core.SessionManager$4.apply(SessionManager.java:218)
< t:2022-02-04 15:40:29,352 f:base.py         l:150  c:RemoteCmdRunner      p:DEBUG >   at com.google.common.util.concurrent.Futures$ChainingListenableFuture.run(Futures.java:906)
< t:2022-02-04 15:40:29,352 f:base.py         l:150  c:RemoteCmdRunner      p:DEBUG >   at com.google.common.util.concurrent.Futures$1$1.run(Futures.java:635)
< t:2022-02-04 15:40:29,352 f:base.py         l:150  c:RemoteCmdRunner      p:DEBUG >   at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128)
< t:2022-02-04 15:40:29,352 f:base.py         l:150  c:RemoteCmdRunner      p:DEBUG >   at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628)
< t:2022-02-04 15:40:29,352 f:base.py         l:150  c:RemoteCmdRunner      p:DEBUG >   at com.datastax.shaded.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30)
< t:2022-02-04 15:40:29,352 f:base.py         l:150  c:RemoteCmdRunner      p:DEBUG >   at java.base/java.lang.Thread.run(Thread.java:834)

...
< t:2022-02-04 15:40:30,593 f:thread.py       l:52   c:cassandra.cluster    p:DEBUG > [control connection] Schemas mismatched, trying again
< t:2022-02-04 15:40:30,742 f:cluster.py      l:2972 c:sdcm.cluster         p:DEBUG > Retrying read after timeout. Attempt #1
< t:2022-02-04 15:40:30,802 f:thread.py       l:52   c:cassandra.cluster    p:DEBUG > [control connection] Schemas mismatched, trying again
< t:2022-02-04 15:40:31,013 f:thread.py       l:52   c:cassandra.cluster    p:DEBUG > [control connection] Schemas mismatched, trying again
< t:2022-02-04 15:40:32,487 f:thread.py       l:52   c:cassandra.cluster    p:DEBUG > [control connection] Schemas mismatched, trying again

casssndra stress

15:40:25.293 [cluster1-worker-9] DEBUG c.d.driver.core.ControlConnection - Checking for schema agreement: versions are [494804d0-200a-3002-bd2a-d2b1710de292, 26d4f6fb-9c45-3b74-82d3-d3343139a7de]
15:40:25.493 [cluster1-worker-9] WARN  com.datastax.driver.core.Cluster - No schema agreement from live replicas after 10 s. The schema may not be up to date on some nodes.
15:40:25.493 [cluster1-worker-11] DEBUG c.d.driver.core.ControlConnection - [Control connection] Refreshing schema for mview.users_by_last_name (TABLE)
Created extra schema. Sleeping 1s for propagation.
15:40:26.509 [main] DEBUG com.datastax.driver.core.Connection - Connection[/10.0.2.138:9042-6, inFlight=1, closed=false] Setting keyspace mview
15:40:26.510 [cluster1-nio-worker-3] DEBUG com.datastax.driver.core.Connection - Connection[/10.0.2.138:9042-6, inFlight=1, closed=false] Keyspace set to mview
15:40:26.515 [cluster1-worker-12] DEBUG com.datastax.driver.core.Connection - Connection[/10.0.1.223:9042-2, inFlight=1, closed=false] Setting keyspace mview
15:40:26.515 [cluster1-worker-12] DEBUG com.datastax.driver.core.Connection - Connection[/10.0.3.7:9042-1, inFlight=1, closed=false] Setting keyspace mview
15:40:26.515 [cluster1-worker-12] DEBUG com.datastax.driver.core.Connection - Connection[/10.0.0.219:9042-3, inFlight=1, closed=false] Setting keyspace mview
15:40:26.516 [cluster1-nio-worker-26] DEBUG com.datastax.driver.core.Connection - Connection[/10.0.3.7:9042-1, inFlight=1, closed=false] Keyspace set to mview
15:40:26.517 [cluster1-nio-worker-2] DEBUG com.datastax.driver.core.Connection - Connection[/10.0.1.223:9042-2, inFlight=1, closed=false] Keyspace set to mview
15:40:26.517 [cluster1-nio-worker-24] DEBUG com.datastax.driver.core.Connection - Connection[/10.0.0.219:9042-3, inFlight=1, closed=false] Keyspace set to mview
Generating batches with [1..1] partitions and [1..1] rows (of [1..1] total rows in the partitions)

Sleeping 2s...
Warming up insert with 0 iterations...
Failed to connect over JMX; not collecting these stats
Warming up read1 with 0 iterations...
Failed to connect over JMX; not collecting these stats
15:40:29.008 [main] DEBUG com.datastax.driver.core.Connection - Connection[/10.0.3.7:9042-15, inFlight=1, closed=false] Setting keyspace mview
15:40:29.009 [cluster1-nio-worker-8] DEBUG com.datastax.driver.core.Connection - Connection[/10.0.3.7:9042-15, inFlight=1, closed=false] Keyspace set to mview
com.datastax.driver.core.exceptions.InvalidQueryException: unconfigured table users
        at com.datastax.driver.core.exceptions.InvalidQueryException.copy(InvalidQueryException.java:50)
        at com.datastax.driver.core.DriverThrowables.propagateCause(DriverThrowables.java:35)
        at com.datastax.driver.core.AbstractSession.prepare(AbstractSession.java:86)
        at org.apache.cassandra.stress.util.JavaDriverClient.prepare(JavaDriverClient.java:116)
        at org.apache.cassandra.stress.StressProfile.getQuery(StressProfile.java:387)
        at org.apache.cassandra.stress.settings.SettingsCommandUser$1.get(SettingsCommandUser.java:93)
        at org.apache.cassandra.stress.settings.SettingsCommandUser$1.get(SettingsCommandUser.java:82)
        at org.apache.cassandra.stress.operations.SampledOpDistributionFactory$1.get(SampledOpDistributionFactory.java:83)
        at org.apache.cassandra.stress.StressAction$Consumer.<init>(StressAction.java:410)
        at org.apache.cassandra.stress.StressAction.run(StressAction.java:234)
        at org.apache.cassandra.stress.StressAction.warmup(StressAction.java:122)
        at org.apache.cassandra.stress.StressAction.run(StressAction.java:71)
        at org.apache.cassandra.stress.Stress.run(Stress.java:143)
        at org.apache.cassandra.stress.Stress.main(Stress.java:62)
Caused by: com.datastax.driver.core.exceptions.InvalidQueryException: unconfigured table users
        at com.datastax.driver.core.Responses$Error.asException(Responses.java:181)
        at com.datastax.driver.core.SessionManager$4.apply(SessionManager.java:253)
        at com.datastax.driver.core.SessionManager$4.apply(SessionManager.java:218)
        at com.google.common.util.concurrent.Futures$ChainingListenableFuture.run(Futures.java:906)
        at com.google.common.util.concurrent.Futures$1$1.run(Futures.java:635)
        at java.base/java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1128)
        at java.base/java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:628)
        at com.datastax.shaded.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30)
        at java.base/java.lang.Thread.run(Thread.java:834)

====================================

Restore Monitor Stack command: $ hydra investigate show-monitor f23d21db-40c5-4e34-882e-34c10578466e
Restore monitor on AWS instance using Jenkins job
Show all stored logs command: $ hydra investigate show-logs f23d21db-40c5-4e34-882e-34c10578466e

Test id: f23d21db-40c5-4e34-882e-34c10578466e

Logs:
grafana - [https://cloudius-jenkins-test.s3.amazonaws.com/f23d21db-40c5-4e34-882e-34c10578466e/20220204_154132/grafana-screenshot-longevity-1tb-7days-test-scylla-per-server-metrics-nemesis-20220204_154349-longevity-tls-1tb-7d-4-6-monitor-node-f23d21db-1.png](https://www.google.com/url?q=https://cloudius-jenkins-test.s3.amazonaws.com/f23d21db-40c5-4e34-882e-34c10578466e/20220204_154132/grafana-screenshot-longevity-1tb-7days-test-scylla-per-server-metrics-nemesis-20220204_154349-longevity-tls-1tb-7d-4-6-monitor-node-f23d21db-1.png%255D(https://cloudius-jenkins-test.s3.amazonaws.com/f23d21db-40c5-4e34-882e-34c10578466e/20220204_154132/grafana-screenshot-longevity-1tb-7days-test-scylla-per-server-metrics-nemesis-20220204_154349-longevity-tls-1tb-7d-4-6-monitor-node-f23d21db-1.png)&source=gmail-html&ust=1644273235205000&usg=AOvVaw2thK6_T6MoSwLvOA7ontHk)
grafana - [https://cloudius-jenkins-test.s3.amazonaws.com/f23d21db-40c5-4e34-882e-34c10578466e/20220204_154132/grafana-screenshot-overview-20220204_154132-longevity-tls-1tb-7d-4-6-monitor-node-f23d21db-1.png](https://www.google.com/url?q=https://cloudius-jenkins-test.s3.amazonaws.com/f23d21db-40c5-4e34-882e-34c10578466e/20220204_154132/grafana-screenshot-overview-20220204_154132-longevity-tls-1tb-7d-4-6-monitor-node-f23d21db-1.png%255D(https://cloudius-jenkins-test.s3.amazonaws.com/f23d21db-40c5-4e34-882e-34c10578466e/20220204_154132/grafana-screenshot-overview-20220204_154132-longevity-tls-1tb-7d-4-6-monitor-node-f23d21db-1.png)&source=gmail-html&ust=1644273235205000&usg=AOvVaw0S_-nWZaOGmuGft1G14AWz)
db-cluster - [https://cloudius-jenkins-test.s3.amazonaws.com/f23d21db-40c5-4e34-882e-34c10578466e/20220204_155337/db-cluster-f23d21db.tar.gz](https://www.google.com/url?q=https://cloudius-jenkins-test.s3.amazonaws.com/f23d21db-40c5-4e34-882e-34c10578466e/20220204_155337/db-cluster-f23d21db.tar.gz%255D(https://cloudius-jenkins-test.s3.amazonaws.com/f23d21db-40c5-4e34-882e-34c10578466e/20220204_155337/db-cluster-f23d21db.tar.gz)&source=gmail-html&ust=1644273235205000&usg=AOvVaw27o86hLfoEc0KOcxH3xXP2)
loader-set - [https://cloudius-jenkins-test.s3.amazonaws.com/f23d21db-40c5-4e34-882e-34c10578466e/20220204_155337/loader-set-f23d21db.tar.gz](https://www.google.com/url?q=https://cloudius-jenkins-test.s3.amazonaws.com/f23d21db-40c5-4e34-882e-34c10578466e/20220204_155337/loader-set-f23d21db.tar.gz%255D(https://cloudius-jenkins-test.s3.amazonaws.com/f23d21db-40c5-4e34-882e-34c10578466e/20220204_155337/loader-set-f23d21db.tar.gz)&source=gmail-html&ust=1644273235205000&usg=AOvVaw33r73fAZRKMhK1xWPZhk5m)
monitor-set - [https://cloudius-jenkins-test.s3.amazonaws.com/f23d21db-40c5-4e34-882e-34c10578466e/20220204_155337/monitor-set-f23d21db.tar.gz](https://www.google.com/url?q=https://cloudius-jenkins-test.s3.amazonaws.com/f23d21db-40c5-4e34-882e-34c10578466e/20220204_155337/monitor-set-f23d21db.tar.gz%255D(https://cloudius-jenkins-test.s3.amazonaws.com/f23d21db-40c5-4e34-882e-34c10578466e/20220204_155337/monitor-set-f23d21db.tar.gz)&source=gmail-html&ust=1644273235205000&usg=AOvVaw3D--i3GNPOnbfWCdWJ4pDx)
sct - [https://cloudius-jenkins-test.s3.amazonaws.com/f23d21db-40c5-4e34-882e-34c10578466e/20220204_155337/sct-runner-f23d21db.tar.gz](https://www.google.com/url?q=https://cloudius-jenkins-test.s3.amazonaws.com/f23d21db-40c5-4e34-882e-34c10578466e/20220204_155337/sct-runner-f23d21db.tar.gz%255D(https://cloudius-jenkins-test.s3.amazonaws.com/f23d21db-40c5-4e34-882e-34c10578466e/20220204_155337/sct-runner-f23d21db.tar.gz)&source=gmail-html&ust=1644273235205000&usg=AOvVaw3NhBO2em-offTe-NO7i_V1)

Jenkins job URL

@slivne slivne added area/schema changes bug and removed triage/master Looking for assignee labels Feb 7, 2022
@slivne slivne assigned eliransin and psarna and unassigned slivne Feb 7, 2022
@slivne slivne added this to the 4.6 milestone Feb 7, 2022
@slivne
Copy link
Contributor

slivne commented Feb 7, 2022

@eliransin / @psarna sending this your way as well - I am not sure this is not a client issue as well

I checked the last report on 4.6.

[shlomi@shlomi-scylladb-com i_9906]$ grep -i "schema_tables" db-cluster-f23d21db*/*/messages.log | sort -k 2 -t ':' | grep -v gossip | cut -f2- -d':'
2022-02-04T11:57:04+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-1 !    INFO |  [shard 0] schema_tables - Schema version changed to 59adb24e-f3cd-3e02-97f0-5b395827453f
2022-02-04T11:57:16+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-1 !    INFO |  [shard 0] schema_tables - Altering system_distributed.service_levels id=b8c556bd-212d-37ad-9484-690c73a5994b version=8a951d7a-1758-346d-953f-b8cc9cbc0f96
2022-02-04T11:57:16+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-1 !    INFO |  [shard 0] schema_tables - Creating keyspace system_distributed
2022-02-04T11:57:16+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-1 !    INFO |  [shard 0] schema_tables - Creating keyspace system_distributed_everywhere
2022-02-04T11:57:16+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-1 !    INFO |  [shard 0] schema_tables - Creating keyspace system_traces
2022-02-04T11:57:16+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-1 !    INFO |  [shard 0] schema_tables - Creating system_distributed.cdc_generation_timestamps id=fdf455c4-cfec-3e00-9719-d7a45436c89d version=5777e043-a81a-3324-8803-b497c056849f
2022-02-04T11:57:16+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-1 !    INFO |  [shard 0] schema_tables - Creating system_distributed.cdc_streams_descriptions_v2 id=0bf73fd7-65b2-36b0-85e5-658131d5df36 version=58a18c32-0c5a-34ba-992f-7723dd0265c0
2022-02-04T11:57:16+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-1 !    INFO |  [shard 0] schema_tables - Creating system_distributed_everywhere.cdc_generation_descriptions_v2 id=234d2227-dd63-3d37-ac5f-c013e2ea9e6e version=f04f0def-50da-31c9-96b1-ab22ba0cc08e
2022-02-04T11:57:16+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-1 !    INFO |  [shard 0] schema_tables - Creating system_distributed.service_levels id=b8c556bd-212d-37ad-9484-690c73a5994b version=08d009cd-6bd2-3be7-8d96-48fea7669507
2022-02-04T11:57:16+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-1 !    INFO |  [shard 0] schema_tables - Creating system_distributed.view_build_status id=5582b59f-8e4e-35e1-b913-3acada51eb04 version=e61ea49b-8dbc-3ff6-a69f-896f4d38b7bb
2022-02-04T11:57:16+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-1 !    INFO |  [shard 0] schema_tables - Creating system_traces.sessions id=c5e99f16-8677-3914-b17e-960613512345 version=a4f13cbe-1b1d-38de-a5ae-7d8befcf9c32
2022-02-04T11:57:16+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-1 !    INFO |  [shard 0] schema_tables - Schema version changed to 45191497-7290-38d9-b4cf-bffddd170d5a
2022-02-04T11:57:16+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-1 !    INFO |  [shard 0] schema_tables - Schema version changed to 45da1815-acb0-3d61-b444-05f708de7fda
2022-02-04T11:57:16+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-1 !    INFO |  [shard 0] schema_tables - Schema version changed to 59adb24e-f3cd-3e02-97f0-5b395827453f
2022-02-04T11:57:16+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-1 !    INFO |  [shard 0] schema_tables - Schema version changed to 59adb24e-f3cd-3e02-97f0-5b395827453f
2022-02-04T11:57:16+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-1 !    INFO |  [shard 0] schema_tables - Schema version changed to 59adb24e-f3cd-3e02-97f0-5b395827453f
2022-02-04T11:57:16+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-1 !    INFO |  [shard 0] schema_tables - Schema version changed to 59adb24e-f3cd-3e02-97f0-5b395827453f
2022-02-04T11:57:16+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-1 !    INFO |  [shard 0] schema_tables - Schema version changed to 59adb24e-f3cd-3e02-97f0-5b395827453f
2022-02-04T11:57:16+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-1 !    INFO |  [shard 0] schema_tables - Schema version changed to 5a515551-dfe6-32bf-a90c-073b486fd049
2022-02-04T11:57:16+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-1 !    INFO |  [shard 0] schema_tables - Schema version changed to 78963c01-5196-3764-93f1-0250bf48752f
2022-02-04T11:57:16+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-1 !    INFO |  [shard 0] schema_tables - Schema version changed to 8924126a-d732-391d-b15f-a8fea173f891
2022-02-04T11:57:16+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-1 !    INFO |  [shard 0] schema_tables - Schema version changed to 8ceae2b7-c4dc-39c1-becc-b4e8d52530bd
2022-02-04T11:57:16+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-1 !    INFO |  [shard 0] schema_tables - Schema version changed to d8781fb4-1866-34eb-96c7-2a1e98f4bed3
2022-02-04T11:57:16+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-1 !    INFO |  [shard 0] schema_tables - Schema version changed to e7a11473-69f6-3b6e-b8c3-54c02a212184
2022-02-04T11:57:16+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-1 !    INFO |  [shard 0] schema_tables - Schema version changed to f399fd51-3145-3257-9810-0fa435c7075f
2022-02-04T11:57:16+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-1 !    INFO |  [shard 0] schema_tables - Schema version changed to f4023b36-d6a7-340c-a040-6f2ca244cfdb
2022-02-04T11:57:17+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-1 !    INFO |  [shard 0] schema_tables - Creating keyspace system_auth
2022-02-04T11:57:17+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-1 !    INFO |  [shard 0] schema_tables - Creating system_auth.role_attributes id=6b8c7359-a843-33f2-a1d8-5dc6a187436f version=6a0146be-3060-3c3b-b045-373e52886f9c
2022-02-04T11:57:17+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-1 !    INFO |  [shard 0] schema_tables - Creating system_auth.role_members id=0ecdaa87-f8fb-3e60-88d1-74fb36fe5c0d version=833f30c7-3dfe-3463-8e04-e38ee59ef53f
2022-02-04T11:57:17+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-1 !    INFO |  [shard 0] schema_tables - Creating system_auth.role_permissions id=3afbe79f-2194-31a7-add7-f5ab90d8ec9c version=578ac7be-dc41-3ac2-bfe3-95865fcb049d
2022-02-04T11:57:17+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-1 !    INFO |  [shard 0] schema_tables - Creating system_auth.roles id=5bc52802-de25-35ed-aeab-188eecebb090 version=187c5b14-e4a3-3983-b14c-cfbe02a1b384
2022-02-04T11:57:17+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-1 !    INFO |  [shard 0] schema_tables - Creating system_traces.events id=8826e8e9-e16a-3728-8753-3bc1fc713c25 version=e0086261-a419-39ce-a186-ef625266cbc7
2022-02-04T11:57:17+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-1 !    INFO |  [shard 0] schema_tables - Creating system_traces.node_slow_log id=bfcc4e62-5b63-3aa1-a1c3-6f5e47f3325c version=8831ea15-811a-3d11-b80c-23470fb5edfc
2022-02-04T11:57:17+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-1 !    INFO |  [shard 0] schema_tables - Creating system_traces.node_slow_log_time_idx id=f9706768-aa1e-3d87-9e5c-51a3927c2870 version=9342371b-f513-36d1-915d-f649d642e5a8
2022-02-04T11:57:17+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-1 !    INFO |  [shard 0] schema_tables - Creating system_traces.sessions_time_idx id=0ebf001c-c1d1-3693-9a63-c3d96ac53318 version=c09b7e1f-3aac-3ff7-bc0d-b8c59a0de223
2022-02-04T11:57:17+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-1 !    INFO |  [shard 0] schema_tables - Schema version changed to 262a3563-2860-3dec-9543-2dfb56297357
2022-02-04T11:57:17+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-1 !    INFO |  [shard 0] schema_tables - Schema version changed to 443dfcc8-45d6-303d-82bf-b5a806627ba7
2022-02-04T11:57:17+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-1 !    INFO |  [shard 0] schema_tables - Schema version changed to 5b8884ec-6f0c-386e-9e69-7f2f5c4664c0
2022-02-04T11:57:17+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-1 !    INFO |  [shard 0] schema_tables - Schema version changed to 6acba4f9-4e2c-3d41-a61e-bedd3a6c4711
2022-02-04T11:57:17+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-1 !    INFO |  [shard 0] schema_tables - Schema version changed to 7b8570bd-1dcb-3595-b30b-80df686fa9ab
2022-02-04T11:57:17+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-1 !    INFO |  [shard 0] schema_tables - Schema version changed to ade7d3eb-2f6f-348d-ab42-128245e0ada7
2022-02-04T11:57:17+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-1 !    INFO |  [shard 0] schema_tables - Schema version changed to e29118cf-5e23-3d39-8522-fa7417cbb846
2022-02-04T11:57:17+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-1 !    INFO |  [shard 0] schema_tables - Schema version changed to f80958f3-541f-3a0c-af2a-ad98e0b35eb7
2022-02-04T11:57:17+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-1 !    INFO |  [shard 0] schema_tables - Schema version changed to fb3cb92f-2895-3986-ba57-3bd006e960ef
2022-02-04T11:59:51+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-2 !    INFO |  [shard 0] schema_tables - Schema version changed to 59adb24e-f3cd-3e02-97f0-5b395827453f
2022-02-04T11:59:53+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-2 !    INFO |  [shard 0] schema_tables - Creating keyspace system_auth
2022-02-04T11:59:53+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-2 !    INFO |  [shard 0] schema_tables - Creating keyspace system_distributed
2022-02-04T11:59:53+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-2 !    INFO |  [shard 0] schema_tables - Creating keyspace system_distributed_everywhere
2022-02-04T11:59:53+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-2 !    INFO |  [shard 0] schema_tables - Creating keyspace system_traces
2022-02-04T11:59:53+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-2 !    INFO |  [shard 0] schema_tables - Creating system_auth.role_attributes id=6b8c7359-a843-33f2-a1d8-5dc6a187436f version=6a0146be-3060-3c3b-b045-373e52886f9c
2022-02-04T11:59:53+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-2 !    INFO |  [shard 0] schema_tables - Creating system_auth.role_members id=0ecdaa87-f8fb-3e60-88d1-74fb36fe5c0d version=833f30c7-3dfe-3463-8e04-e38ee59ef53f
2022-02-04T11:59:53+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-2 !    INFO |  [shard 0] schema_tables - Creating system_auth.role_permissions id=3afbe79f-2194-31a7-add7-f5ab90d8ec9c version=578ac7be-dc41-3ac2-bfe3-95865fcb049d
2022-02-04T11:59:53+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-2 !    INFO |  [shard 0] schema_tables - Creating system_auth.roles id=5bc52802-de25-35ed-aeab-188eecebb090 version=187c5b14-e4a3-3983-b14c-cfbe02a1b384
2022-02-04T11:59:53+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-2 !    INFO |  [shard 0] schema_tables - Creating system_distributed.cdc_generation_timestamps id=fdf455c4-cfec-3e00-9719-d7a45436c89d version=5777e043-a81a-3324-8803-b497c056849f
2022-02-04T11:59:53+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-2 !    INFO |  [shard 0] schema_tables - Creating system_distributed.cdc_streams_descriptions_v2 id=0bf73fd7-65b2-36b0-85e5-658131d5df36 version=58a18c32-0c5a-34ba-992f-7723dd0265c0
2022-02-04T11:59:53+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-2 !    INFO |  [shard 0] schema_tables - Creating system_distributed_everywhere.cdc_generation_descriptions_v2 id=234d2227-dd63-3d37-ac5f-c013e2ea9e6e version=f04f0def-50da-31c9-96b1-ab22ba0cc08e
2022-02-04T11:59:53+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-2 !    INFO |  [shard 0] schema_tables - Creating system_distributed.service_levels id=b8c556bd-212d-37ad-9484-690c73a5994b version=8a951d7a-1758-346d-953f-b8cc9cbc0f96
2022-02-04T11:59:53+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-2 !    INFO |  [shard 0] schema_tables - Creating system_distributed.view_build_status id=5582b59f-8e4e-35e1-b913-3acada51eb04 version=e61ea49b-8dbc-3ff6-a69f-896f4d38b7bb
2022-02-04T11:59:53+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-2 !    INFO |  [shard 0] schema_tables - Creating system_traces.events id=8826e8e9-e16a-3728-8753-3bc1fc713c25 version=e0086261-a419-39ce-a186-ef625266cbc7
2022-02-04T11:59:53+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-2 !    INFO |  [shard 0] schema_tables - Creating system_traces.node_slow_log id=bfcc4e62-5b63-3aa1-a1c3-6f5e47f3325c version=8831ea15-811a-3d11-b80c-23470fb5edfc
2022-02-04T11:59:53+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-2 !    INFO |  [shard 0] schema_tables - Creating system_traces.node_slow_log_time_idx id=f9706768-aa1e-3d87-9e5c-51a3927c2870 version=9342371b-f513-36d1-915d-f649d642e5a8
2022-02-04T11:59:53+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-2 !    INFO |  [shard 0] schema_tables - Creating system_traces.sessions id=c5e99f16-8677-3914-b17e-960613512345 version=a4f13cbe-1b1d-38de-a5ae-7d8befcf9c32
2022-02-04T11:59:53+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-2 !    INFO |  [shard 0] schema_tables - Creating system_traces.sessions_time_idx id=0ebf001c-c1d1-3693-9a63-c3d96ac53318 version=c09b7e1f-3aac-3ff7-bc0d-b8c59a0de223
2022-02-04T11:59:53+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-2 !    INFO |  [shard 0] schema_tables - Schema version changed to 2847308f-9b62-3db8-b92f-7bd468c78f56
2022-02-04T11:59:53+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-2 !    INFO |  [shard 0] schema_tables - Schema version changed to 2847308f-9b62-3db8-b92f-7bd468c78f56
2022-02-04T12:00:04+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-2 !    INFO |  [shard 0] schema_tables - Schema version changed to 443dfcc8-45d6-303d-82bf-b5a806627ba7
2022-02-04T12:00:04+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-2 !    INFO |  [shard 0] schema_tables - Schema version changed to 443dfcc8-45d6-303d-82bf-b5a806627ba7
2022-02-04T12:00:04+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-2 !    INFO |  [shard 0] schema_tables - Schema version changed to 443dfcc8-45d6-303d-82bf-b5a806627ba7
2022-02-04T12:00:04+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-2 !    INFO |  [shard 0] schema_tables - Schema version changed to 443dfcc8-45d6-303d-82bf-b5a806627ba7
2022-02-04T12:00:04+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-2 !    INFO |  [shard 0] schema_tables - Schema version changed to 443dfcc8-45d6-303d-82bf-b5a806627ba7
2022-02-04T12:03:45+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-3 !    INFO |  [shard 0] schema_tables - Schema version changed to 59adb24e-f3cd-3e02-97f0-5b395827453f
2022-02-04T12:03:46+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-3 !    INFO |  [shard 0] schema_tables - Creating keyspace system_auth
2022-02-04T12:03:46+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-3 !    INFO |  [shard 0] schema_tables - Creating keyspace system_distributed
2022-02-04T12:03:46+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-3 !    INFO |  [shard 0] schema_tables - Creating keyspace system_distributed_everywhere
2022-02-04T12:03:46+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-3 !    INFO |  [shard 0] schema_tables - Creating keyspace system_traces
2022-02-04T12:03:47+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-3 !    INFO |  [shard 0] schema_tables - Creating system_auth.role_attributes id=6b8c7359-a843-33f2-a1d8-5dc6a187436f version=6a0146be-3060-3c3b-b045-373e52886f9c
2022-02-04T12:03:47+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-3 !    INFO |  [shard 0] schema_tables - Creating system_auth.role_members id=0ecdaa87-f8fb-3e60-88d1-74fb36fe5c0d version=833f30c7-3dfe-3463-8e04-e38ee59ef53f
2022-02-04T12:03:47+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-3 !    INFO |  [shard 0] schema_tables - Creating system_auth.role_permissions id=3afbe79f-2194-31a7-add7-f5ab90d8ec9c version=578ac7be-dc41-3ac2-bfe3-95865fcb049d
2022-02-04T12:03:47+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-3 !    INFO |  [shard 0] schema_tables - Creating system_auth.roles id=5bc52802-de25-35ed-aeab-188eecebb090 version=187c5b14-e4a3-3983-b14c-cfbe02a1b384
2022-02-04T12:03:47+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-3 !    INFO |  [shard 0] schema_tables - Creating system_distributed.cdc_generation_timestamps id=fdf455c4-cfec-3e00-9719-d7a45436c89d version=5777e043-a81a-3324-8803-b497c056849f
2022-02-04T12:03:47+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-3 !    INFO |  [shard 0] schema_tables - Creating system_distributed.cdc_streams_descriptions_v2 id=0bf73fd7-65b2-36b0-85e5-658131d5df36 version=58a18c32-0c5a-34ba-992f-7723dd0265c0
2022-02-04T12:03:47+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-3 !    INFO |  [shard 0] schema_tables - Creating system_distributed_everywhere.cdc_generation_descriptions_v2 id=234d2227-dd63-3d37-ac5f-c013e2ea9e6e version=f04f0def-50da-31c9-96b1-ab22ba0cc08e
2022-02-04T12:03:47+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-3 !    INFO |  [shard 0] schema_tables - Creating system_distributed.service_levels id=b8c556bd-212d-37ad-9484-690c73a5994b version=8a951d7a-1758-346d-953f-b8cc9cbc0f96
2022-02-04T12:03:47+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-3 !    INFO |  [shard 0] schema_tables - Creating system_distributed.view_build_status id=5582b59f-8e4e-35e1-b913-3acada51eb04 version=e61ea49b-8dbc-3ff6-a69f-896f4d38b7bb
2022-02-04T12:03:47+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-3 !    INFO |  [shard 0] schema_tables - Creating system_traces.events id=8826e8e9-e16a-3728-8753-3bc1fc713c25 version=e0086261-a419-39ce-a186-ef625266cbc7
2022-02-04T12:03:47+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-3 !    INFO |  [shard 0] schema_tables - Creating system_traces.node_slow_log id=bfcc4e62-5b63-3aa1-a1c3-6f5e47f3325c version=8831ea15-811a-3d11-b80c-23470fb5edfc
2022-02-04T12:03:47+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-3 !    INFO |  [shard 0] schema_tables - Creating system_traces.node_slow_log_time_idx id=f9706768-aa1e-3d87-9e5c-51a3927c2870 version=9342371b-f513-36d1-915d-f649d642e5a8
2022-02-04T12:03:47+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-3 !    INFO |  [shard 0] schema_tables - Creating system_traces.sessions id=c5e99f16-8677-3914-b17e-960613512345 version=a4f13cbe-1b1d-38de-a5ae-7d8befcf9c32
2022-02-04T12:03:47+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-3 !    INFO |  [shard 0] schema_tables - Creating system_traces.sessions_time_idx id=0ebf001c-c1d1-3693-9a63-c3d96ac53318 version=c09b7e1f-3aac-3ff7-bc0d-b8c59a0de223
2022-02-04T12:03:47+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-3 !    INFO |  [shard 0] schema_tables - Schema version changed to 2847308f-9b62-3db8-b92f-7bd468c78f56
2022-02-04T12:03:47+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-3 !    INFO |  [shard 0] schema_tables - Schema version changed to 2847308f-9b62-3db8-b92f-7bd468c78f56
2022-02-04T12:03:47+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-3 !    INFO |  [shard 0] schema_tables - Schema version changed to 2847308f-9b62-3db8-b92f-7bd468c78f56
2022-02-04T12:03:47+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-3 !    INFO |  [shard 0] schema_tables - Schema version changed to 2847308f-9b62-3db8-b92f-7bd468c78f56
2022-02-04T12:03:58+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-3 !    INFO |  [shard 0] schema_tables - Schema version changed to 443dfcc8-45d6-303d-82bf-b5a806627ba7
2022-02-04T12:03:58+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-3 !    INFO |  [shard 0] schema_tables - Schema version changed to 443dfcc8-45d6-303d-82bf-b5a806627ba7
2022-02-04T12:03:58+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-3 !    INFO |  [shard 0] schema_tables - Schema version changed to 443dfcc8-45d6-303d-82bf-b5a806627ba7
2022-02-04T12:03:58+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-3 !    INFO |  [shard 0] schema_tables - Schema version changed to 443dfcc8-45d6-303d-82bf-b5a806627ba7
2022-02-04T12:03:58+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-3 !    INFO |  [shard 0] schema_tables - Schema version changed to 443dfcc8-45d6-303d-82bf-b5a806627ba7
2022-02-04T12:07:40+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-4 !    INFO |  [shard 0] schema_tables - Schema version changed to 59adb24e-f3cd-3e02-97f0-5b395827453f
2022-02-04T12:07:41+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-4 !    INFO |  [shard 0] schema_tables - Creating keyspace system_auth
2022-02-04T12:07:41+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-4 !    INFO |  [shard 0] schema_tables - Creating keyspace system_distributed
2022-02-04T12:07:41+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-4 !    INFO |  [shard 0] schema_tables - Creating keyspace system_distributed_everywhere
2022-02-04T12:07:41+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-4 !    INFO |  [shard 0] schema_tables - Creating keyspace system_traces
2022-02-04T12:07:41+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-4 !    INFO |  [shard 0] schema_tables - Creating system_auth.role_attributes id=6b8c7359-a843-33f2-a1d8-5dc6a187436f version=6a0146be-3060-3c3b-b045-373e52886f9c
2022-02-04T12:07:41+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-4 !    INFO |  [shard 0] schema_tables - Creating system_auth.role_members id=0ecdaa87-f8fb-3e60-88d1-74fb36fe5c0d version=833f30c7-3dfe-3463-8e04-e38ee59ef53f
2022-02-04T12:07:41+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-4 !    INFO |  [shard 0] schema_tables - Creating system_auth.role_permissions id=3afbe79f-2194-31a7-add7-f5ab90d8ec9c version=578ac7be-dc41-3ac2-bfe3-95865fcb049d
2022-02-04T12:07:41+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-4 !    INFO |  [shard 0] schema_tables - Creating system_auth.roles id=5bc52802-de25-35ed-aeab-188eecebb090 version=187c5b14-e4a3-3983-b14c-cfbe02a1b384
2022-02-04T12:07:41+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-4 !    INFO |  [shard 0] schema_tables - Creating system_distributed.cdc_generation_timestamps id=fdf455c4-cfec-3e00-9719-d7a45436c89d version=5777e043-a81a-3324-8803-b497c056849f
2022-02-04T12:07:41+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-4 !    INFO |  [shard 0] schema_tables - Creating system_distributed.cdc_streams_descriptions_v2 id=0bf73fd7-65b2-36b0-85e5-658131d5df36 version=58a18c32-0c5a-34ba-992f-7723dd0265c0
2022-02-04T12:07:41+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-4 !    INFO |  [shard 0] schema_tables - Creating system_distributed_everywhere.cdc_generation_descriptions_v2 id=234d2227-dd63-3d37-ac5f-c013e2ea9e6e version=f04f0def-50da-31c9-96b1-ab22ba0cc08e
2022-02-04T12:07:41+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-4 !    INFO |  [shard 0] schema_tables - Creating system_distributed.service_levels id=b8c556bd-212d-37ad-9484-690c73a5994b version=8a951d7a-1758-346d-953f-b8cc9cbc0f96
2022-02-04T12:07:41+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-4 !    INFO |  [shard 0] schema_tables - Creating system_distributed.view_build_status id=5582b59f-8e4e-35e1-b913-3acada51eb04 version=e61ea49b-8dbc-3ff6-a69f-896f4d38b7bb
2022-02-04T12:07:41+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-4 !    INFO |  [shard 0] schema_tables - Creating system_traces.events id=8826e8e9-e16a-3728-8753-3bc1fc713c25 version=e0086261-a419-39ce-a186-ef625266cbc7
2022-02-04T12:07:41+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-4 !    INFO |  [shard 0] schema_tables - Creating system_traces.node_slow_log id=bfcc4e62-5b63-3aa1-a1c3-6f5e47f3325c version=8831ea15-811a-3d11-b80c-23470fb5edfc
2022-02-04T12:07:41+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-4 !    INFO |  [shard 0] schema_tables - Creating system_traces.node_slow_log_time_idx id=f9706768-aa1e-3d87-9e5c-51a3927c2870 version=9342371b-f513-36d1-915d-f649d642e5a8
2022-02-04T12:07:41+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-4 !    INFO |  [shard 0] schema_tables - Creating system_traces.sessions id=c5e99f16-8677-3914-b17e-960613512345 version=a4f13cbe-1b1d-38de-a5ae-7d8befcf9c32
2022-02-04T12:07:41+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-4 !    INFO |  [shard 0] schema_tables - Creating system_traces.sessions_time_idx id=0ebf001c-c1d1-3693-9a63-c3d96ac53318 version=c09b7e1f-3aac-3ff7-bc0d-b8c59a0de223
2022-02-04T12:07:41+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-4 !    INFO |  [shard 0] schema_tables - Schema version changed to 2847308f-9b62-3db8-b92f-7bd468c78f56
2022-02-04T12:07:41+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-4 !    INFO |  [shard 0] schema_tables - Schema version changed to 2847308f-9b62-3db8-b92f-7bd468c78f56
2022-02-04T12:07:41+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-4 !    INFO |  [shard 0] schema_tables - Schema version changed to 2847308f-9b62-3db8-b92f-7bd468c78f56
2022-02-04T12:07:41+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-4 !    INFO |  [shard 0] schema_tables - Schema version changed to 2847308f-9b62-3db8-b92f-7bd468c78f56
2022-02-04T12:07:41+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-4 !    INFO |  [shard 0] schema_tables - Schema version changed to 2847308f-9b62-3db8-b92f-7bd468c78f56
2022-02-04T12:07:42+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-4 !    INFO |  [shard 0] schema_tables - Schema version changed to 2847308f-9b62-3db8-b92f-7bd468c78f56
2022-02-04T12:07:53+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-4 !    INFO |  [shard 0] schema_tables - Schema version changed to 443dfcc8-45d6-303d-82bf-b5a806627ba7
2022-02-04T12:07:53+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-4 !    INFO |  [shard 0] schema_tables - Schema version changed to 443dfcc8-45d6-303d-82bf-b5a806627ba7
2022-02-04T12:07:53+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-4 !    INFO |  [shard 0] schema_tables - Schema version changed to 443dfcc8-45d6-303d-82bf-b5a806627ba7
2022-02-04T12:07:53+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-4 !    INFO |  [shard 0] schema_tables - Schema version changed to 443dfcc8-45d6-303d-82bf-b5a806627ba7
2022-02-04T12:07:53+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-4 !    INFO |  [shard 0] schema_tables - Schema version changed to 443dfcc8-45d6-303d-82bf-b5a806627ba7
2022-02-04T12:10:22+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-1 !    INFO |  [shard 0] schema_tables - Altering keyspace system_auth
2022-02-04T12:10:22+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-1 !    INFO |  [shard 0] schema_tables - Schema version changed to 553c9bfa-4efc-3d29-ad9f-8673d521cb65
2022-02-04T12:10:22+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-2 !    INFO |  [shard 0] schema_tables - Altering keyspace system_auth
2022-02-04T12:10:22+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-2 !    INFO |  [shard 0] schema_tables - Schema version changed to 553c9bfa-4efc-3d29-ad9f-8673d521cb65
2022-02-04T12:10:22+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-3 !    INFO |  [shard 0] schema_tables - Altering keyspace system_auth
2022-02-04T12:10:22+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-3 !    INFO |  [shard 0] schema_tables - Schema version changed to 553c9bfa-4efc-3d29-ad9f-8673d521cb65
2022-02-04T12:10:22+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-4 !    INFO |  [shard 0] schema_tables - Altering keyspace system_auth
2022-02-04T12:10:22+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-4 !    INFO |  [shard 0] schema_tables - Schema version changed to 553c9bfa-4efc-3d29-ad9f-8673d521cb65
2022-02-04T12:18:47+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-1 !    INFO |  [shard 6] schema_tables - Creating keyspace keyspace1
2022-02-04T12:18:47+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-1 !    INFO |  [shard 6] schema_tables - Schema version changed to e45a7415-ce88-3f49-9a90-a2eb4e4b4a42
2022-02-04T12:18:47+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-2 !    INFO |  [shard 6] schema_tables - Creating keyspace keyspace1
2022-02-04T12:18:47+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-2 !    INFO |  [shard 6] schema_tables - Schema version changed to e45a7415-ce88-3f49-9a90-a2eb4e4b4a42
2022-02-04T12:18:47+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-3 !    INFO |  [shard 6] schema_tables - Creating keyspace keyspace1
2022-02-04T12:18:47+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-3 !    INFO |  [shard 6] schema_tables - Schema version changed to e45a7415-ce88-3f49-9a90-a2eb4e4b4a42
2022-02-04T12:18:47+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-4 !    INFO |  [shard 6] schema_tables - Creating keyspace keyspace1
2022-02-04T12:18:47+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-4 !    INFO |  [shard 6] schema_tables - Schema version changed to e45a7415-ce88-3f49-9a90-a2eb4e4b4a42
2022-02-04T12:18:48+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-1 !    INFO |  [shard 12] schema_tables - Creating keyspace1.standard1 id=9ad4d650-85b4-11ec-8f64-b6177dccc2a3 version=d2f23fe2-3e27-3a4c-9a9a-da5b7a3285a9
2022-02-04T12:18:48+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-1 !    INFO |  [shard 12] schema_tables - Schema version changed to 60a6a6c6-5682-3408-a673-b80c98f9143d
2022-02-04T12:18:48+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-2 !    INFO |  [shard 12] schema_tables - Creating keyspace1.standard1 id=9ad4d650-85b4-11ec-8f64-b6177dccc2a3 version=d2f23fe2-3e27-3a4c-9a9a-da5b7a3285a9
2022-02-04T12:18:48+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-2 !    INFO |  [shard 12] schema_tables - Schema version changed to 60a6a6c6-5682-3408-a673-b80c98f9143d
2022-02-04T12:18:48+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-3 !    INFO |  [shard 12] schema_tables - Creating keyspace1.standard1 id=9ad4d650-85b4-11ec-8f64-b6177dccc2a3 version=d2f23fe2-3e27-3a4c-9a9a-da5b7a3285a9
2022-02-04T12:18:48+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-3 !    INFO |  [shard 12] schema_tables - Schema version changed to 60a6a6c6-5682-3408-a673-b80c98f9143d
2022-02-04T12:18:48+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-4 !    INFO |  [shard 12] schema_tables - Creating keyspace1.standard1 id=9ad4d650-85b4-11ec-8f64-b6177dccc2a3 version=d2f23fe2-3e27-3a4c-9a9a-da5b7a3285a9
2022-02-04T12:18:48+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-4 !    INFO |  [shard 12] schema_tables - Schema version changed to 60a6a6c6-5682-3408-a673-b80c98f9143d
2022-02-04T15:39:33+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-4 !    INFO |  [shard 8] schema_tables - Creating keyspace mview
2022-02-04T15:39:34+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-2 !    INFO |  [shard 8] schema_tables - Creating keyspace mview
2022-02-04T15:39:34+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-4 !    INFO |  [shard 8] schema_tables - Schema version changed to 26d4f6fb-9c45-3b74-82d3-d3343139a7de
2022-02-04T15:39:35+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-2 !    INFO |  [shard 8] schema_tables - Schema version changed to 26d4f6fb-9c45-3b74-82d3-d3343139a7de
2022-02-04T15:39:38+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-1 !    INFO |  [shard 8] schema_tables - Creating keyspace mview
2022-02-04T15:39:38+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-1 !    INFO |  [shard 8] schema_tables - Schema version changed to 26d4f6fb-9c45-3b74-82d3-d3343139a7de
2022-02-04T15:39:50+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-1 !    INFO |  [shard 9] schema_tables - Creating mview.users id=afb8d320-85d0-11ec-a892-b842122da0a7 version=2b9fae2b-3897-39fa-9c65-d4795bce5332
2022-02-04T15:39:50+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-1 !    INFO |  [shard 9] schema_tables - Schema version changed to 3e9479d1-f937-3f4a-be28-f687e5ee60d3
2022-02-04T15:39:52+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-2 !    INFO |  [shard 9] schema_tables - Creating mview.users id=afb8d320-85d0-11ec-a892-b842122da0a7 version=2b9fae2b-3897-39fa-9c65-d4795bce5332
2022-02-04T15:39:52+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-2 !    INFO |  [shard 9] schema_tables - Schema version changed to 3e9479d1-f937-3f4a-be28-f687e5ee60d3
2022-02-04T15:39:52+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-4 !    INFO |  [shard 9] schema_tables - Creating mview.users id=afb8d320-85d0-11ec-a892-b842122da0a7 version=2b9fae2b-3897-39fa-9c65-d4795bce5332
2022-02-04T15:39:52+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-4 !    INFO |  [shard 9] schema_tables - Schema version changed to 3e9479d1-f937-3f4a-be28-f687e5ee60d3
2022-02-04T15:39:55+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-3 !    INFO |  [shard 8] schema_tables - Creating keyspace mview
2022-02-04T15:40:04+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-2 !    INFO |  [shard 10] schema_tables - Altering mview.users id=afb8d320-85d0-11ec-a892-b842122da0a7 version=ac29f6b0-3b37-3b51-81d3-0446f1b09a0b
2022-02-04T15:40:04+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-2 !    INFO |  [shard 10] schema_tables - Creating mview.users_by_first_name id=b87463d0-85d0-11ec-9e7f-b841122da0a7 version=e48233ac-ab06-34b8-ae03-26615b68d5fe
2022-02-04T15:40:04+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-2 !    INFO |  [shard 10] schema_tables - Schema version changed to 4275e034-bbd8-34e3-b062-8523a1a67eb4
2022-02-04T15:40:05+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-1 !    INFO |  [shard 10] schema_tables - Altering mview.users id=afb8d320-85d0-11ec-a892-b842122da0a7 version=ac29f6b0-3b37-3b51-81d3-0446f1b09a0b
2022-02-04T15:40:05+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-1 !    INFO |  [shard 10] schema_tables - Creating mview.users_by_first_name id=b87463d0-85d0-11ec-9e7f-b841122da0a7 version=e48233ac-ab06-34b8-ae03-26615b68d5fe
2022-02-04T15:40:05+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-1 !    INFO |  [shard 10] schema_tables - Schema version changed to 4275e034-bbd8-34e3-b062-8523a1a67eb4
2022-02-04T15:40:07+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-4 !    INFO |  [shard 10] schema_tables - Altering mview.users id=afb8d320-85d0-11ec-a892-b842122da0a7 version=ac29f6b0-3b37-3b51-81d3-0446f1b09a0b
2022-02-04T15:40:07+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-4 !    INFO |  [shard 10] schema_tables - Creating mview.users_by_first_name id=b87463d0-85d0-11ec-9e7f-b841122da0a7 version=e48233ac-ab06-34b8-ae03-26615b68d5fe
2022-02-04T15:40:07+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-4 !    INFO |  [shard 10] schema_tables - Schema version changed to 4275e034-bbd8-34e3-b062-8523a1a67eb4
2022-02-04T15:40:11+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-3 !    INFO |  [shard 8] schema_tables - Schema version changed to 26d4f6fb-9c45-3b74-82d3-d3343139a7de
2022-02-04T15:40:15+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-1 !    INFO |  [shard 4] schema_tables - Altering mview.users id=afb8d320-85d0-11ec-a892-b842122da0a7 version=35e4e8ad-2134-3075-a5c5-106a00d3055c
2022-02-04T15:40:15+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-1 !    INFO |  [shard 4] schema_tables - Creating mview.users_by_last_name id=bf307b00-85d0-11ec-9cf0-f269b348587d version=1459057a-2bde-327a-a744-8d62fd99f17f
2022-02-04T15:40:15+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-1 !    INFO |  [shard 4] schema_tables - Schema version changed to 494804d0-200a-3002-bd2a-d2b1710de292
2022-02-04T15:40:15+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-2 !    INFO |  [shard 4] schema_tables - Altering mview.users id=afb8d320-85d0-11ec-a892-b842122da0a7 version=35e4e8ad-2134-3075-a5c5-106a00d3055c
2022-02-04T15:40:15+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-2 !    INFO |  [shard 4] schema_tables - Creating mview.users_by_last_name id=bf307b00-85d0-11ec-9cf0-f269b348587d version=1459057a-2bde-327a-a744-8d62fd99f17f
2022-02-04T15:40:15+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-2 !    INFO |  [shard 4] schema_tables - Schema version changed to 494804d0-200a-3002-bd2a-d2b1710de292
2022-02-04T15:40:16+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-4 !    INFO |  [shard 4] schema_tables - Altering mview.users id=afb8d320-85d0-11ec-a892-b842122da0a7 version=35e4e8ad-2134-3075-a5c5-106a00d3055c
2022-02-04T15:40:16+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-4 !    INFO |  [shard 4] schema_tables - Creating mview.users_by_last_name id=bf307b00-85d0-11ec-9cf0-f269b348587d version=1459057a-2bde-327a-a744-8d62fd99f17f
2022-02-04T15:40:16+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-4 !    INFO |  [shard 4] schema_tables - Schema version changed to 494804d0-200a-3002-bd2a-d2b1710de292
2022-02-04T15:40:43+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-3 !    INFO |  [shard 9] schema_tables - Creating mview.users id=afb8d320-85d0-11ec-a892-b842122da0a7 version=2b9fae2b-3897-39fa-9c65-d4795bce5332
2022-02-04T15:40:43+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-3 !    INFO |  [shard 9] schema_tables - Schema version changed to 3e9479d1-f937-3f4a-be28-f687e5ee60d3
2022-02-04T15:40:54+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-3 !    INFO |  [shard 10] schema_tables - Altering mview.users id=afb8d320-85d0-11ec-a892-b842122da0a7 version=ac29f6b0-3b37-3b51-81d3-0446f1b09a0b
2022-02-04T15:40:54+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-3 !    INFO |  [shard 10] schema_tables - Creating mview.users_by_first_name id=b87463d0-85d0-11ec-9e7f-b841122da0a7 version=e48233ac-ab06-34b8-ae03-26615b68d5fe
2022-02-04T15:40:55+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-3 !    INFO |  [shard 10] schema_tables - Schema version changed to 4275e034-bbd8-34e3-b062-8523a1a67eb4
2022-02-04T15:41:09+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-3 !    INFO |  [shard 4] schema_tables - Altering mview.users id=afb8d320-85d0-11ec-a892-b842122da0a7 version=35e4e8ad-2134-3075-a5c5-106a00d3055c
2022-02-04T15:41:09+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-3 !    INFO |  [shard 4] schema_tables - Creating mview.users_by_last_name id=bf307b00-85d0-11ec-9cf0-f269b348587d version=1459057a-2bde-327a-a744-8d62fd99f17f
2022-02-04T15:41:11+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-3 !    INFO |  [shard 4] schema_tables - Schema version changed to 494804d0-200a-3002-bd2a-d2b1710de292
2022-02-04T15:41:13+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-1 !    INFO |  [shard 0] schema_tables - Schema version changed to 494804d0-200a-3002-bd2a-d2b1710de292
2022-02-04T15:41:13+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-3 !    INFO |  [shard 0] schema_tables - Schema version changed to 494804d0-200a-3002-bd2a-d2b1710de292
2022-02-04T15:41:15+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-3 !    INFO |  [shard 0] schema_tables - Schema version changed to 494804d0-200a-3002-bd2a-d2b1710de292
2022-02-04T15:41:16+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-3 !    INFO |  [shard 0] schema_tables - Schema version changed to 494804d0-200a-3002-bd2a-d2b1710de292
2022-02-04T15:41:18+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-3 !    INFO |  [shard 0] schema_tables - Schema version changed to 494804d0-200a-3002-bd2a-d2b1710de292
2022-02-04T15:41:19+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-3 !    INFO |  [shard 0] schema_tables - Schema version changed to 494804d0-200a-3002-bd2a-d2b1710de292
2022-02-04T15:41:20+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-3 !    INFO |  [shard 0] schema_tables - Schema version changed to 494804d0-200a-3002-bd2a-d2b1710de292

schema only

[shlomi@shlomi-scylladb-com i_9906]$ grep -i "schema version" db-cluster-f23d21db*/*/messages.log | sort -k 2 -t ':' | grep -v gossip | cut -f2- -d':'
2022-02-04T11:57:04+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-1 !    INFO |  [shard 0] schema_tables - Schema version changed to 59adb24e-f3cd-3e02-97f0-5b395827453f
2022-02-04T11:57:16+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-1 !    INFO |  [shard 0] schema_tables - Schema version changed to 45191497-7290-38d9-b4cf-bffddd170d5a
2022-02-04T11:57:16+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-1 !    INFO |  [shard 0] schema_tables - Schema version changed to 45da1815-acb0-3d61-b444-05f708de7fda
2022-02-04T11:57:16+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-1 !    INFO |  [shard 0] schema_tables - Schema version changed to 59adb24e-f3cd-3e02-97f0-5b395827453f
2022-02-04T11:57:16+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-1 !    INFO |  [shard 0] schema_tables - Schema version changed to 59adb24e-f3cd-3e02-97f0-5b395827453f
2022-02-04T11:57:16+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-1 !    INFO |  [shard 0] schema_tables - Schema version changed to 59adb24e-f3cd-3e02-97f0-5b395827453f
2022-02-04T11:57:16+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-1 !    INFO |  [shard 0] schema_tables - Schema version changed to 59adb24e-f3cd-3e02-97f0-5b395827453f
2022-02-04T11:57:16+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-1 !    INFO |  [shard 0] schema_tables - Schema version changed to 59adb24e-f3cd-3e02-97f0-5b395827453f
2022-02-04T11:57:16+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-1 !    INFO |  [shard 0] schema_tables - Schema version changed to 5a515551-dfe6-32bf-a90c-073b486fd049
2022-02-04T11:57:16+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-1 !    INFO |  [shard 0] schema_tables - Schema version changed to 78963c01-5196-3764-93f1-0250bf48752f
2022-02-04T11:57:16+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-1 !    INFO |  [shard 0] schema_tables - Schema version changed to 8924126a-d732-391d-b15f-a8fea173f891
2022-02-04T11:57:16+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-1 !    INFO |  [shard 0] schema_tables - Schema version changed to 8ceae2b7-c4dc-39c1-becc-b4e8d52530bd
2022-02-04T11:57:16+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-1 !    INFO |  [shard 0] schema_tables - Schema version changed to d8781fb4-1866-34eb-96c7-2a1e98f4bed3
2022-02-04T11:57:16+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-1 !    INFO |  [shard 0] schema_tables - Schema version changed to e7a11473-69f6-3b6e-b8c3-54c02a212184
2022-02-04T11:57:16+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-1 !    INFO |  [shard 0] schema_tables - Schema version changed to f399fd51-3145-3257-9810-0fa435c7075f
2022-02-04T11:57:16+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-1 !    INFO |  [shard 0] schema_tables - Schema version changed to f4023b36-d6a7-340c-a040-6f2ca244cfdb
2022-02-04T11:57:17+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-1 !    INFO |  [shard 0] schema_tables - Schema version changed to 262a3563-2860-3dec-9543-2dfb56297357
2022-02-04T11:57:17+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-1 !    INFO |  [shard 0] schema_tables - Schema version changed to 443dfcc8-45d6-303d-82bf-b5a806627ba7
2022-02-04T11:57:17+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-1 !    INFO |  [shard 0] schema_tables - Schema version changed to 5b8884ec-6f0c-386e-9e69-7f2f5c4664c0
2022-02-04T11:57:17+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-1 !    INFO |  [shard 0] schema_tables - Schema version changed to 6acba4f9-4e2c-3d41-a61e-bedd3a6c4711
2022-02-04T11:57:17+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-1 !    INFO |  [shard 0] schema_tables - Schema version changed to 7b8570bd-1dcb-3595-b30b-80df686fa9ab
2022-02-04T11:57:17+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-1 !    INFO |  [shard 0] schema_tables - Schema version changed to ade7d3eb-2f6f-348d-ab42-128245e0ada7
2022-02-04T11:57:17+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-1 !    INFO |  [shard 0] schema_tables - Schema version changed to e29118cf-5e23-3d39-8522-fa7417cbb846
2022-02-04T11:57:17+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-1 !    INFO |  [shard 0] schema_tables - Schema version changed to f80958f3-541f-3a0c-af2a-ad98e0b35eb7
2022-02-04T11:57:17+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-1 !    INFO |  [shard 0] schema_tables - Schema version changed to fb3cb92f-2895-3986-ba57-3bd006e960ef
2022-02-04T11:59:51+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-2 !    INFO |  [shard 0] schema_tables - Schema version changed to 59adb24e-f3cd-3e02-97f0-5b395827453f
2022-02-04T11:59:53+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-2 !    INFO |  [shard 0] schema_tables - Schema version changed to 2847308f-9b62-3db8-b92f-7bd468c78f56
2022-02-04T11:59:53+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-2 !    INFO |  [shard 0] schema_tables - Schema version changed to 2847308f-9b62-3db8-b92f-7bd468c78f56
2022-02-04T12:00:04+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-2 !    INFO |  [shard 0] schema_tables - Schema version changed to 443dfcc8-45d6-303d-82bf-b5a806627ba7
2022-02-04T12:00:04+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-2 !    INFO |  [shard 0] schema_tables - Schema version changed to 443dfcc8-45d6-303d-82bf-b5a806627ba7
2022-02-04T12:00:04+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-2 !    INFO |  [shard 0] schema_tables - Schema version changed to 443dfcc8-45d6-303d-82bf-b5a806627ba7
2022-02-04T12:00:04+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-2 !    INFO |  [shard 0] schema_tables - Schema version changed to 443dfcc8-45d6-303d-82bf-b5a806627ba7
2022-02-04T12:00:04+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-2 !    INFO |  [shard 0] schema_tables - Schema version changed to 443dfcc8-45d6-303d-82bf-b5a806627ba7
2022-02-04T12:03:45+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-3 !    INFO |  [shard 0] schema_tables - Schema version changed to 59adb24e-f3cd-3e02-97f0-5b395827453f
2022-02-04T12:03:47+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-3 !    INFO |  [shard 0] schema_tables - Schema version changed to 2847308f-9b62-3db8-b92f-7bd468c78f56
2022-02-04T12:03:47+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-3 !    INFO |  [shard 0] schema_tables - Schema version changed to 2847308f-9b62-3db8-b92f-7bd468c78f56
2022-02-04T12:03:47+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-3 !    INFO |  [shard 0] schema_tables - Schema version changed to 2847308f-9b62-3db8-b92f-7bd468c78f56
2022-02-04T12:03:47+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-3 !    INFO |  [shard 0] schema_tables - Schema version changed to 2847308f-9b62-3db8-b92f-7bd468c78f56
2022-02-04T12:03:58+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-3 !    INFO |  [shard 0] schema_tables - Schema version changed to 443dfcc8-45d6-303d-82bf-b5a806627ba7
2022-02-04T12:03:58+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-3 !    INFO |  [shard 0] schema_tables - Schema version changed to 443dfcc8-45d6-303d-82bf-b5a806627ba7
2022-02-04T12:03:58+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-3 !    INFO |  [shard 0] schema_tables - Schema version changed to 443dfcc8-45d6-303d-82bf-b5a806627ba7
2022-02-04T12:03:58+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-3 !    INFO |  [shard 0] schema_tables - Schema version changed to 443dfcc8-45d6-303d-82bf-b5a806627ba7
2022-02-04T12:03:58+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-3 !    INFO |  [shard 0] schema_tables - Schema version changed to 443dfcc8-45d6-303d-82bf-b5a806627ba7
2022-02-04T12:07:40+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-4 !    INFO |  [shard 0] schema_tables - Schema version changed to 59adb24e-f3cd-3e02-97f0-5b395827453f
2022-02-04T12:07:41+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-4 !    INFO |  [shard 0] schema_tables - Schema version changed to 2847308f-9b62-3db8-b92f-7bd468c78f56
2022-02-04T12:07:41+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-4 !    INFO |  [shard 0] schema_tables - Schema version changed to 2847308f-9b62-3db8-b92f-7bd468c78f56
2022-02-04T12:07:41+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-4 !    INFO |  [shard 0] schema_tables - Schema version changed to 2847308f-9b62-3db8-b92f-7bd468c78f56
2022-02-04T12:07:41+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-4 !    INFO |  [shard 0] schema_tables - Schema version changed to 2847308f-9b62-3db8-b92f-7bd468c78f56
2022-02-04T12:07:41+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-4 !    INFO |  [shard 0] schema_tables - Schema version changed to 2847308f-9b62-3db8-b92f-7bd468c78f56
2022-02-04T12:07:42+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-4 !    INFO |  [shard 0] schema_tables - Schema version changed to 2847308f-9b62-3db8-b92f-7bd468c78f56
2022-02-04T12:07:53+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-4 !    INFO |  [shard 0] schema_tables - Schema version changed to 443dfcc8-45d6-303d-82bf-b5a806627ba7
2022-02-04T12:07:53+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-4 !    INFO |  [shard 0] schema_tables - Schema version changed to 443dfcc8-45d6-303d-82bf-b5a806627ba7
2022-02-04T12:07:53+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-4 !    INFO |  [shard 0] schema_tables - Schema version changed to 443dfcc8-45d6-303d-82bf-b5a806627ba7
2022-02-04T12:07:53+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-4 !    INFO |  [shard 0] schema_tables - Schema version changed to 443dfcc8-45d6-303d-82bf-b5a806627ba7
2022-02-04T12:07:53+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-4 !    INFO |  [shard 0] schema_tables - Schema version changed to 443dfcc8-45d6-303d-82bf-b5a806627ba7
2022-02-04T12:10:22+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-1 !    INFO |  [shard 0] schema_tables - Schema version changed to 553c9bfa-4efc-3d29-ad9f-8673d521cb65
2022-02-04T12:10:22+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-2 !    INFO |  [shard 0] schema_tables - Schema version changed to 553c9bfa-4efc-3d29-ad9f-8673d521cb65
2022-02-04T12:10:22+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-3 !    INFO |  [shard 0] schema_tables - Schema version changed to 553c9bfa-4efc-3d29-ad9f-8673d521cb65
2022-02-04T12:10:22+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-4 !    INFO |  [shard 0] schema_tables - Schema version changed to 553c9bfa-4efc-3d29-ad9f-8673d521cb65
2022-02-04T12:18:47+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-1 !    INFO |  [shard 6] schema_tables - Schema version changed to e45a7415-ce88-3f49-9a90-a2eb4e4b4a42
2022-02-04T12:18:47+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-2 !    INFO |  [shard 6] schema_tables - Schema version changed to e45a7415-ce88-3f49-9a90-a2eb4e4b4a42
2022-02-04T12:18:47+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-3 !    INFO |  [shard 6] schema_tables - Schema version changed to e45a7415-ce88-3f49-9a90-a2eb4e4b4a42
2022-02-04T12:18:47+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-4 !    INFO |  [shard 6] schema_tables - Schema version changed to e45a7415-ce88-3f49-9a90-a2eb4e4b4a42
2022-02-04T12:18:48+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-1 !    INFO |  [shard 12] schema_tables - Schema version changed to 60a6a6c6-5682-3408-a673-b80c98f9143d
2022-02-04T12:18:48+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-2 !    INFO |  [shard 12] schema_tables - Schema version changed to 60a6a6c6-5682-3408-a673-b80c98f9143d
2022-02-04T12:18:48+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-3 !    INFO |  [shard 12] schema_tables - Schema version changed to 60a6a6c6-5682-3408-a673-b80c98f9143d
2022-02-04T12:18:48+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-4 !    INFO |  [shard 12] schema_tables - Schema version changed to 60a6a6c6-5682-3408-a673-b80c98f9143d
2022-02-04T15:39:34+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-4 !    INFO |  [shard 8] schema_tables - Schema version changed to 26d4f6fb-9c45-3b74-82d3-d3343139a7de
2022-02-04T15:39:35+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-2 !    INFO |  [shard 8] schema_tables - Schema version changed to 26d4f6fb-9c45-3b74-82d3-d3343139a7de
2022-02-04T15:39:38+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-1 !    INFO |  [shard 8] schema_tables - Schema version changed to 26d4f6fb-9c45-3b74-82d3-d3343139a7de
2022-02-04T15:39:50+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-1 !    INFO |  [shard 9] schema_tables - Schema version changed to 3e9479d1-f937-3f4a-be28-f687e5ee60d3
2022-02-04T15:39:52+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-2 !    INFO |  [shard 9] schema_tables - Schema version changed to 3e9479d1-f937-3f4a-be28-f687e5ee60d3
2022-02-04T15:39:52+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-4 !    INFO |  [shard 9] schema_tables - Schema version changed to 3e9479d1-f937-3f4a-be28-f687e5ee60d3
2022-02-04T15:40:04+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-2 !    INFO |  [shard 10] schema_tables - Schema version changed to 4275e034-bbd8-34e3-b062-8523a1a67eb4
2022-02-04T15:40:05+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-1 !    INFO |  [shard 10] schema_tables - Schema version changed to 4275e034-bbd8-34e3-b062-8523a1a67eb4
2022-02-04T15:40:07+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-4 !    INFO |  [shard 10] schema_tables - Schema version changed to 4275e034-bbd8-34e3-b062-8523a1a67eb4
2022-02-04T15:40:11+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-3 !    INFO |  [shard 8] schema_tables - Schema version changed to 26d4f6fb-9c45-3b74-82d3-d3343139a7de
2022-02-04T15:40:15+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-1 !    INFO |  [shard 4] schema_tables - Schema version changed to 494804d0-200a-3002-bd2a-d2b1710de292
2022-02-04T15:40:15+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-2 !    INFO |  [shard 4] schema_tables - Schema version changed to 494804d0-200a-3002-bd2a-d2b1710de292
2022-02-04T15:40:16+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-4 !    INFO |  [shard 4] schema_tables - Schema version changed to 494804d0-200a-3002-bd2a-d2b1710de292
2022-02-04T15:40:43+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-3 !    INFO |  [shard 9] schema_tables - Schema version changed to 3e9479d1-f937-3f4a-be28-f687e5ee60d3
2022-02-04T15:40:55+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-3 !    INFO |  [shard 10] schema_tables - Schema version changed to 4275e034-bbd8-34e3-b062-8523a1a67eb4
2022-02-04T15:41:11+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-3 !    INFO |  [shard 4] schema_tables - Schema version changed to 494804d0-200a-3002-bd2a-d2b1710de292
2022-02-04T15:41:13+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-1 !    INFO |  [shard 0] schema_tables - Schema version changed to 494804d0-200a-3002-bd2a-d2b1710de292
2022-02-04T15:41:13+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-3 !    INFO |  [shard 0] schema_tables - Schema version changed to 494804d0-200a-3002-bd2a-d2b1710de292
2022-02-04T15:41:15+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-3 !    INFO |  [shard 0] schema_tables - Schema version changed to 494804d0-200a-3002-bd2a-d2b1710de292
2022-02-04T15:41:16+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-3 !    INFO |  [shard 0] schema_tables - Schema version changed to 494804d0-200a-3002-bd2a-d2b1710de292
2022-02-04T15:41:18+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-3 !    INFO |  [shard 0] schema_tables - Schema version changed to 494804d0-200a-3002-bd2a-d2b1710de292
2022-02-04T15:41:19+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-3 !    INFO |  [shard 0] schema_tables - Schema version changed to 494804d0-200a-3002-bd2a-d2b1710de292
2022-02-04T15:41:20+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-3 !    INFO |  [shard 0] schema_tables - Schema version changed to 494804d0-200a-3002-bd2a-d2b1710de292

I did not have time to look but I think it does converge

2022-02-04T15:40:15+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-1 !    INFO |  [shard 4] schema_tables - Schema version changed to 494804d0-200a-3002-bd2a-d2b1710de292
2022-02-04T15:40:15+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-2 !    INFO |  [shard 4] schema_tables - Schema version changed to 494804d0-200a-3002-bd2a-d2b1710de292
2022-02-04T15:40:16+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-4 !    INFO |  [shard 4] schema_tables - Schema version changed to 494804d0-200a-3002-bd2a-d2b1710de292
2022-02-04T15:40:43+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-3 !    INFO |  [shard 9] schema_tables - Schema version changed to 3e9479d1-f937-3f4a-be28-f687e5ee60d3
2022-02-04T15:40:55+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-3 !    INFO |  [shard 10] schema_tables - Schema version changed to 4275e034-bbd8-34e3-b062-8523a1a67eb4
2022-02-04T15:41:11+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-3 !    INFO |  [shard 4] schema_tables - Schema version changed to 494804d0-200a-3002-bd2a-d2b1710de292
2022-02-04T15:41:13+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-1 !    INFO |  [shard 0] schema_tables - Schema version changed to 494804d0-200a-3002-bd2a-d2b1710de292
2022-02-04T15:41:13+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-3 !    INFO |  [shard 0] schema_tables - Schema version changed to 494804d0-200a-3002-bd2a-d2b1710de292
2022-02-04T15:41:15+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-3 !    INFO |  [shard 0] schema_tables - Schema version changed to 494804d0-200a-3002-bd2a-d2b1710de292
2022-02-04T15:41:16+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-3 !    INFO |  [shard 0] schema_tables - Schema version changed to 494804d0-200a-3002-bd2a-d2b1710de292
2022-02-04T15:41:18+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-3 !    INFO |  [shard 0] schema_tables - Schema version changed to 494804d0-200a-3002-bd2a-d2b1710de292
2022-02-04T15:41:19+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-3 !    INFO |  [shard 0] schema_tables - Schema version changed to 494804d0-200a-3002-bd2a-d2b1710de292
2022-02-04T15:41:20+00:00 longevity-tls-1tb-7d-4-6-db-node-f23d21db-3 !    INFO |  [shard 0] schema_tables - Schema version changed to 494804d0-200a-3002-bd2a-d2b1710de292

at least according to the logs - not sure if its a client issue ot something else - taking to long to converge

@slivne
Copy link
Contributor

slivne commented Feb 7, 2022

@tgrabiec can you please look at the last output and see if you can explain why schema is still pulled

@slivne slivne assigned bhalevy and unassigned psarna and eliransin Feb 7, 2022
@k0machi
Copy link
Contributor

k0machi commented Feb 8, 2022

Had what looks like same issue happen on longevity-multi-keyspaces-60h, where cassandra-stress is started in batches, creating around 1000 keyspaces (and tables) total (100 per loader). Pretty much guaranteed to receive at least one schema agreement failure there:
c-s starts here:

< t:2022-02-07 21:39:57,023 f:remote_base.py  l:520  c:RemoteCmdRunner      p:DEBUG > Running command "echo TAG: loader_idx:0-cpu_idx:0-keyspace_idx:1; STRESS_TEST_MARKER=H53DK5KUJPLVBO2VIN50; cassandra-stress write no-warmup cl=QUORUM n=4000000 -schema keyspace=keyspace599 'replication(factor=3)' -port jmx=6868 -mode cql3 native -rate threads=20 -pop seq=1..4000000 -log interval=30 -node 10.0.1.72 -errors skip-unsupported-columns"...

and fails around 10 seconds later

< t:2022-02-07 21:40:11,950 f:base.py         l:146  c:RemoteCmdRunner      p:ERROR > Error executing command: "echo TAG: loader_idx:0-cpu_idx:0-keyspace_idx:1; STRESS_TEST_MARKER=H53DK5KUJPLVBO2VIN50; cassandra-stress write no-warmup cl=QUORUM n=4000000 -schema keyspace=keyspace599 'replication(factor=3)' -port jmx=6868 -mode cql3 native -rate threads=20 -pop seq=1..4000000 -log interval=30 -node 10.0.1.72 -errors skip-unsupported-columns"; Exit status: 1
< t:2022-02-07 21:40:11,951 f:base.py         l:148  c:RemoteCmdRunner      p:DEBUG > STDOUT: 10.0.1.225:9042; Rack: 1a
< t:2022-02-07 21:40:11,951 f:base.py         l:148  c:RemoteCmdRunner      p:DEBUG > Datacenter: eu-west; Host: /10.0.0.70:9042; Rack: 1a
< t:2022-02-07 21:40:11,951 f:base.py         l:148  c:RemoteCmdRunner      p:DEBUG > Datacenter: eu-west; Host: /10.0.1.72:9042; Rack: 1a
< t:2022-02-07 21:40:11,951 f:base.py         l:148  c:RemoteCmdRunner      p:DEBUG > Datacenter: eu-west; Host: /10.0.0.106:9042; Rack: 1a
< t:2022-02-07 21:40:11,951 f:base.py         l:148  c:RemoteCmdRunner      p:DEBUG > Datacenter: eu-west; Host: /10.0.0.139:9042; Rack: 1a
< t:2022-02-07 21:40:11,951 f:base.py         l:150  c:RemoteCmdRunner      p:DEBUG > STDERR: java.lang.RuntimeException: Encountered exception creating schema
< t:2022-02-07 21:40:11,951 f:base.py         l:150  c:RemoteCmdRunner      p:DEBUG >         at org.apache.cassandra.stress.settings.SettingsSchema.createKeySpaces(SettingsSchema.java:88)
< t:2022-02-07 21:40:11,951 f:base.py         l:150  c:RemoteCmdRunner      p:DEBUG >         at org.apache.cassandra.stress.settings.StressSettings.maybeCreateKeyspaces(StressSettings.java:156)
< t:2022-02-07 21:40:11,951 f:base.py         l:150  c:RemoteCmdRunner      p:DEBUG >         at org.apache.cassandra.stress.StressAction.run(StressAction.java:57)
< t:2022-02-07 21:40:11,951 f:base.py         l:150  c:RemoteCmdRunner      p:DEBUG >         at org.apache.cassandra.stress.Stress.run(Stress.java:143)
< t:2022-02-07 21:40:11,951 f:base.py         l:150  c:RemoteCmdRunner      p:DEBUG >         at org.apache.cassandra.stress.Stress.main(Stress.java:62)
< t:2022-02-07 21:40:11,951 f:base.py         l:150  c:RemoteCmdRunner      p:DEBUG > Caused by: com.datastax.driver.core.exceptions.OperationTimedOutException: [/10.0.0.70:9042] Timed out waiting for server response
< t:2022-02-07 21:40:11,951 f:base.py         l:150  c:RemoteCmdRunner      p:DEBUG >         at com.datastax.driver.core.exceptions.OperationTimedOutException.copy(OperationTimedOutException.java:43)
< t:2022-02-07 21:40:11,951 f:base.py         l:150  c:RemoteCmdRunner      p:DEBUG >         at com.datastax.driver.core.exceptions.OperationTimedOutException.copy(OperationTimedOutException.java:25)
< t:2022-02-07 21:40:11,951 f:base.py         l:150  c:RemoteCmdRunner      p:DEBUG >         at com.datastax.driver.core.DriverThrowables.propagateCause(DriverThrowables.java:35)
< t:2022-02-07 21:40:11,951 f:base.py         l:150  c:RemoteCmdRunner      p:DEBUG >         at com.datastax.driver.core.DefaultResultSetFuture.getUninterruptibly(DefaultResultSetFuture.java:293)
< t:2022-02-07 21:40:11,951 f:base.py         l:150  c:RemoteCmdRunner      p:DEBUG >         at com.datastax.driver.core.AbstractSession.execute(AbstractSession.java:58)
< t:2022-02-07 21:40:11,951 f:base.py         l:150  c:RemoteCmdRunner      p:DEBUG >         at org.apache.cassandra.stress.util.JavaDriverClient.execute(JavaDriverClient.java:190)
< t:2022-02-07 21:40:11,951 f:base.py         l:150  c:RemoteCmdRunner      p:DEBUG >         at org.apache.cassandra.stress.settings.SettingsSchema.createKeySpaces(SettingsSchema.java:77)
< t:2022-02-07 21:40:11,951 f:base.py         l:150  c:RemoteCmdRunner      p:DEBUG >         ... 4 more
< t:2022-02-07 21:40:11,951 f:base.py         l:150  c:RemoteCmdRunner      p:DEBUG > Caused by: com.datastax.driver.core.exceptions.OperationTimedOutException: [/10.0.0.70:9042] Timed out waiting for server response
< t:2022-02-07 21:40:11,951 f:base.py         l:150  c:RemoteCmdRunner      p:DEBUG >         at com.datastax.driver.core.RequestHandler$SpeculativeExecution.onTimeout(RequestHandler.java:954)
< t:2022-02-07 21:40:11,951 f:base.py         l:150  c:RemoteCmdRunner      p:DEBUG >         at com.datastax.driver.core.Connection$ResponseHandler$1.run(Connection.java:1575)
< t:2022-02-07 21:40:11,951 f:base.py         l:150  c:RemoteCmdRunner      p:DEBUG >         at com.datastax.shaded.netty.util.HashedWheelTimer$HashedWheelTimeout.expire(HashedWheelTimer.java:663)
< t:2022-02-07 21:40:11,951 f:base.py         l:150  c:RemoteCmdRunner      p:DEBUG >         at com.datastax.shaded.netty.util.HashedWheelTimer$HashedWheelBucket.expireTimeouts(HashedWheelTimer.java:738)
< t:2022-02-07 21:40:11,951 f:base.py         l:150  c:RemoteCmdRunner      p:DEBUG >         at com.datastax.shaded.netty.util.HashedWheelTimer$Worker.run(HashedWheelTimer.java:466)
< t:2022-02-07 21:40:11,951 f:base.py         l:150  c:RemoteCmdRunner      p:DEBUG >         at com.datastax.shaded.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30)
< t:2022-02-07 21:40:11,951 f:base.py         l:150  c:RemoteCmdRunner      p:DEBUG >         at java.lang.Thread.run(Thread.java:748)

on the scylla side we can see that keyspace599 took around 16 seconds to be created and it was created previously:

Feb 07 16:09:57 longevity-1000-keyspaces-4-6-db-node-698079ab-5 scylla[18296]:  [shard 29] schema_tables - Creating keyspace keyspace599
Feb 07 16:09:58 longevity-1000-keyspaces-4-6-db-node-698079ab-5 scylla[18296]:  [shard 29] mutation_partition - Memory usage of unpaged query exceeds soft limit of 1048576 (configured via max_memory_for_unlimited_query_soft_limit)
Feb 07 16:09:58 longevity-1000-keyspaces-4-6-db-node-698079ab-5 scylla[18296]:  [shard 29] schema_tables - Schema version changed to 6573967b-1ff1-37cd-a168-49970eb78e71
Feb 07 16:10:16 longevity-1000-keyspaces-4-6-db-node-698079ab-5 scylla[18296]:  [shard 12] schema_tables - Creating keyspace599.standard1 id=700d99b0-8830-11ec-b013-b6f8185e92e9 version=000db2c4-37ee-3697-952f-5eed300a1456
Feb 07 16:10:16 longevity-1000-keyspaces-4-6-db-node-698079ab-5 scylla[18296]:  [shard 12] mutation_partition - Memory usage of unpaged query exceeds soft limit of 1048576 (configured via max_memory_for_unlimited_query_soft_limit)
Feb 07 16:10:16 longevity-1000-keyspaces-4-6-db-node-698079ab-5 scylla[18296]:  [shard 12] schema_tables - Schema version changed to 3017448c-6802-3ea6-9c06-389175e027f2

after it is created, schema is succesfully propagated:

longevity-1000-keyspaces-4-6-db-node-698079ab-6/system.log:Feb 07 16:10:16 longevity-1000-keyspaces-4-6-db-node-698079ab-6 scylla[20009]:  [shard 12] schema_tables - Schema version changed to 3017448c-6802-3ea6-9c06-389175e027f2
longevity-1000-keyspaces-4-6-db-node-698079ab-4/system.log:Feb 07 16:10:16 longevity-1000-keyspaces-4-6-db-node-698079ab-4 scylla[16686]:  [shard 12] schema_tables - Schema version changed to 3017448c-6802-3ea6-9c06-389175e027f2
longevity-1000-keyspaces-4-6-db-node-698079ab-5/system.log:Feb 07 16:10:16 longevity-1000-keyspaces-4-6-db-node-698079ab-5 scylla[18296]:  [shard 12] schema_tables - Schema version changed to 3017448c-6802-3ea6-9c06-389175e027f2
longevity-1000-keyspaces-4-6-db-node-698079ab-3/system.log:Feb 07 16:10:16 longevity-1000-keyspaces-4-6-db-node-698079ab-3 scylla[15384]:  [shard 12] schema_tables - Schema version changed to 3017448c-6802-3ea6-9c06-389175e027f2
longevity-1000-keyspaces-4-6-db-node-698079ab-2/system.log:Feb 07 16:10:16 longevity-1000-keyspaces-4-6-db-node-698079ab-2 scylla[13772]:  [shard 12] schema_tables - Schema version changed to 3017448c-6802-3ea6-9c06-389175e027f2
longevity-1000-keyspaces-4-6-db-node-698079ab-1/system.log:Feb 07 16:10:16 longevity-1000-keyspaces-4-6-db-node-698079ab-1 scylla[12741]:  [shard 12] schema_tables - Schema version changed to 3017448c-6802-3ea6-9c06-389175e027f2

then when c-s fails, this is what scylla log contains:

longevity-1000-keyspaces-4-6-db-node-698079ab-2/system.log:Feb 07 21:39:59 longevity-1000-keyspaces-4-6-db-node-698079ab-2 scylla[13772]:  [shard 9] schema_tables - Creating keyspace599.counter1 id=7fe7be00-885e-11ec-8fa6-aafca5e789bc version=3c73c187-78fc-3fcd-ac1a-56618fb29ea3
longevity-1000-keyspaces-4-6-db-node-698079ab-5/system.log:Feb 07 21:40:16 longevity-1000-keyspaces-4-6-db-node-698079ab-5 scylla[18296]:  [shard 9] schema_tables - Creating keyspace599.counter1 id=7fe7be00-885e-11ec-8fa6-aafca5e789bc version=3c73c187-78fc-3fcd-ac1a-56618fb29ea3
longevity-1000-keyspaces-4-6-db-node-698079ab-6/system.log:Feb 07 21:39:59 longevity-1000-keyspaces-4-6-db-node-698079ab-6 scylla[20009]:  [shard 9] schema_tables - Creating keyspace599.counter1 id=7fe7be00-885e-11ec-8fa6-aafca5e789bc version=3c73c187-78fc-3fcd-ac1a-56618fb29ea3
longevity-1000-keyspaces-4-6-db-node-698079ab-4/system.log:Feb 07 21:40:02 longevity-1000-keyspaces-4-6-db-node-698079ab-4 scylla[16686]:  [shard 9] schema_tables - Creating keyspace599.counter1 id=7fe7be00-885e-11ec-8fa6-aafca5e789bc version=3c73c187-78fc-3fcd-ac1a-56618fb29ea3
longevity-1000-keyspaces-4-6-db-node-698079ab-1/system.log:Feb 07 21:40:05 longevity-1000-keyspaces-4-6-db-node-698079ab-1 scylla[12741]:  [shard 9] schema_tables - Creating keyspace599.counter1 id=7fe7be00-885e-11ec-8fa6-aafca5e789bc version=3c73c187-78fc-3fcd-ac1a-56618fb29ea3
longevity-1000-keyspaces-4-6-db-node-698079ab-3/system.log:Feb 07 21:39:59 longevity-1000-keyspaces-4-6-db-node-698079ab-3 scylla[15384]:  [shard 9] schema_tables - Creating keyspace599.counter1 id=7fe7be00-885e-11ec-8fa6-aafca5e789bc version=3c73c187-78fc-3fcd-ac1a-56618fb29ea3

Maybe there's an issue with reporting already existing keyspaces to driver?

Installation details

Kernel Version: 5.11.0-1022-aws
Scylla version (or git commit hash): 4.6~rc5-20220203.5694ec189 with build-id f5d85bf5abe6d2f9fd3487e2469ce1c34304cc14
Cluster size: 6 nodes (i3.8xlarge)

Scylla running with shards number (live nodes):

  • longevity-1000-keyspaces-4-6-db-node-698079ab-1 (52.51.157.206 | 10.0.1.72) (shards: 30)
  • longevity-1000-keyspaces-4-6-db-node-698079ab-2 (3.250.98.0 | 10.0.0.139) (shards: 30)
  • longevity-1000-keyspaces-4-6-db-node-698079ab-3 (18.203.102.62 | 10.0.0.47) (shards: 30)
  • longevity-1000-keyspaces-4-6-db-node-698079ab-4 (34.241.202.153 | 10.0.1.225) (shards: 30)
  • longevity-1000-keyspaces-4-6-db-node-698079ab-5 (34.245.154.94 | 10.0.0.70) (shards: 30)
  • longevity-1000-keyspaces-4-6-db-node-698079ab-6 (54.75.72.12 | 10.0.0.106) (shards: 30)

OS / Image: ami-06df116ea1c70171c (aws: eu-west-1)

Test: longevity-multi-keyspaces-60h-test
Test id: 698079ab-3937-46d5-8b15-db94815794b2
Test name: longevity_test/longevity-multi-keyspaces-60h-test
Test config file(s):

  • longevity-multi-keyspaces.yaml

  • Restore Monitor Stack command: $ hydra investigate show-monitor 698079ab-3937-46d5-8b15-db94815794b2

  • Restore monitor on AWS instance using Jenkins job

  • Show all stored logs command: $ hydra investigate show-logs 698079ab-3937-46d5-8b15-db94815794b2

Logs:

Jenkins job URL

@tgrabiec
Copy link
Contributor

tgrabiec commented Feb 8, 2022

can you please look at the last output and see if you can explain why schema is still pulled

I looked at db-cluster-f23d21db. The schema converges, but one of the nodes, longevity-tls-1tb-7d-4-6-db-node-f23d21db-3, seems slower than other nodes. I don't see excessive number of pulls. It converges eventually. Probably schema pulls are slow because the disk is overloaded (since it is a test with a slow disk).

Probably duplicate of #1459

@slivne Let me know if I should dig deeper.

@slivne
Copy link
Contributor

slivne commented Feb 9, 2022

@roydahan - the ebs (slow disks) setup is new - can we do a run on 4.5 and see that it happens on 4.5 as well (if its simple - not sure the AMI will be able to support this).

@tgrabiec thanks - yes I need more help

@ShlomiBalalis
Copy link

The same error happened in 4h-100gb run of 4.6.rc5-0.20220203.5694ec189 with build-id e06d3596d9782c9c4bdc0e06125e5b867dd117a5 :
We ran 10 c-s threads each with a different keyspace one after the other. For the first 9 keyspaces, there was no issue. However, when the tenth thread ran, the cluster could not reach an agreement over the keyspace creation for a minute:

Table creation started at 12:56:28.523:

12:56:28.483 [cluster1-nio-worker-3] DEBUG com.datastax.driver.core.Session - Added connection pool for /10.0.14.148:9042
12:56:28.523 [cluster1-nio-worker-0] DEBUG com.datastax.driver.core.Cluster - Received event EVENT CREATED KEYSPACE drop_table_during_repair_ks_9, scheduling delivery
12:56:28.731 [cluster1-nio-worker-0] DEBUG com.datastax.driver.core.Cluster - Refreshing schema for drop_table_during_repair_ks_9
12:56:29.559 [cluster1-worker-1] DEBUG c.d.driver.core.ControlConnection - [Control connection] Refreshing schema for drop_table_during_repair_ks_9
12:56:29.587 [cluster1-worker-0] DEBUG c.d.driver.core.ControlConnection - Checking for schema agreement: versions are [f6b65efa-d633-35d2-8d13-720d7b8ba801, 87e0e368-0267-3ff3-b329-b52b287bbb1a]

longevity-100gb-4h-4-6-db-node-d05cb422-1 [3.87.140.168 | 10.0.14.238]

2022-02-07T12:56:28+00:00 longevity-100gb-4h-4-6-db-node-d05cb422-1 !    INFO |  [shard 11] migration_manager - Create new Keyspace: KSMetaData{name=drop_table_during_repair_ks_9, strategyClass=org.apache.cassandra.locator.SimpleStrategy, strategyOptions={replication_factor=6}, cfMetaData={}, durable_writes=1, userTypes=org.apache.cassandra.config.UTMetaData@0xe0b006557418}
2022-02-07T12:56:28+00:00 longevity-100gb-4h-4-6-db-node-d05cb422-1 !    INFO |  [shard 11] schema_tables - Creating keyspace drop_table_during_repair_ks_9
2022-02-07T12:56:38+00:00 longevity-100gb-4h-4-6-db-node-d05cb422-1 !    INFO |  [shard 15] schema_tables - Creating drop_table_during_repair_ks_9.standard1 id=639f7c90-8815-11ec-9eac-e80beec9914d version=cd12091f-15f4-3acd-a932-f8393fc56371

longevity-100gb-4h-4-6-db-node-d05cb422-2 [54.172.252.153 | 10.0.13.204]

2022-02-07T12:56:28+00:00 longevity-100gb-4h-4-6-db-node-d05cb422-2 !    INFO |  [shard 11] schema_tables - Creating keyspace drop_table_during_repair_ks_9
2022-02-07T12:56:39+00:00 longevity-100gb-4h-4-6-db-node-d05cb422-2 !    INFO |  [shard 15] schema_tables - Creating drop_table_during_repair_ks_9.standard1 id=639f7c90-8815-11ec-9eac-e80beec9914d version=cd12091f-15f4-3acd-a932-f8393fc56371

longevity-100gb-4h-4-6-db-node-d05cb422-3 [3.90.9.134 | 10.0.14.204]

2022-02-07T12:56:28+00:00 longevity-100gb-4h-4-6-db-node-d05cb422-3 !    INFO |  [shard 11] schema_tables - Creating keyspace drop_table_during_repair_ks_9
2022-02-07T12:56:39+00:00 longevity-100gb-4h-4-6-db-node-d05cb422-3 !    INFO |  [shard 15] schema_tables - Creating drop_table_during_repair_ks_9.standard1 id=639f7c90-8815-11ec-9eac-e80beec9914d version=cd12091f-15f4-3acd-a932-f8393fc56371

longevity-100gb-4h-4-6-db-node-d05cb422-4 [54.145.171.94 | 10.0.12.9]

2022-02-07T12:56:28+00:00 longevity-100gb-4h-4-6-db-node-d05cb422-4 !    INFO |  [shard 11] schema_tables - Creating keyspace drop_table_during_repair_ks_9
2022-02-07T12:56:38+00:00 longevity-100gb-4h-4-6-db-node-d05cb422-4 !    INFO |  [shard 15] schema_tables - Creating drop_table_during_repair_ks_9.standard1 id=639f7c90-8815-11ec-9eac-e80beec9914d version=cd12091f-15f4-3acd-a932-f8393fc56371

longevity-100gb-4h-4-6-db-node-d05cb422-5 (18.212.110.153 | 10.0.14.148)

2022-02-07T12:56:57+00:00 longevity-100gb-4h-4-6-db-node-d05cb422-5 !    INFO |  [shard 11] schema_tables - Creating keyspace drop_table_during_repair_ks_9
2022-02-07T12:57:26+00:00 longevity-100gb-4h-4-6-db-node-d05cb422-5 !    INFO |  [shard 15] schema_tables - Creating drop_table_during_repair_ks_9.standard1 id=639f7c90-8815-11ec-9eac-e80beec9914d version=cd12091f-15f4-3acd-a932-f8393fc56371

longevity-100gb-4h-4-6-db-node-d05cb422-6 (3.88.239.241 | 10.0.13.128)

2022-02-07T12:56:28+00:00 longevity-100gb-4h-4-6-db-node-d05cb422-6 !    INFO |  [shard 11] schema_tables - Creating keyspace drop_table_during_repair_ks_9
2022-02-07T12:56:38+00:00 longevity-100gb-4h-4-6-db-node-d05cb422-6 !    INFO |  [shard 15] schema_tables - Creating drop_table_during_repair_ks_9.standard1 id=639f7c90-8815-11ec-9eac-e80beec9914d version=cd12091f-15f4-3acd-a932-f8393fc56371

Because of the long time it took for the schema agreement, the c-s thread failed (but it took 4 hours for it to end)

nodes:

longevity-100gb-4h-4-6-db-node-d05cb422-1 [3.87.140.168 | 10.0.14.238]
longevity-100gb-4h-4-6-db-node-d05cb422-2 [54.172.252.153 | 10.0.13.204]
longevity-100gb-4h-4-6-db-node-d05cb422-3 [3.90.9.134 | 10.0.14.204]
longevity-100gb-4h-4-6-db-node-d05cb422-4 [54.145.171.94 | 10.0.12.9]
longevity-100gb-4h-4-6-db-node-d05cb422-5 [18.212.110.153 | 10.0.14.148]
longevity-100gb-4h-4-6-db-node-d05cb422-6 [3.88.239.241 | 10.0.13.128]

Logs:

+-----------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+
|                                                                                                         Log links for testrun with test id d05cb422-5e2e-4510-b85a-3ee529209672                                                                                                         |
+-----------------+-------------+---------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+
| Date            | Log type    | Link                                                                                                                                                                                                                                                    |
+-----------------+-------------+---------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+
| 20220207_172105 | grafana     | https://cloudius-jenkins-test.s3.amazonaws.com/d05cb422-5e2e-4510-b85a-3ee529209672/20220207_172105/grafana-screenshot-longevity-100gb-4h-arm-test-scylla-per-server-metrics-nemesis-20220207_172344-longevity-100gb-4h-4-6-monitor-node-d05cb422-1.png |
| 20220207_172105 | grafana     | https://cloudius-jenkins-test.s3.amazonaws.com/d05cb422-5e2e-4510-b85a-3ee529209672/20220207_172105/grafana-screenshot-overview-20220207_172105-longevity-100gb-4h-4-6-monitor-node-d05cb422-1.png                                                      |
| 20220207_173557 | db-cluster  | https://cloudius-jenkins-test.s3.amazonaws.com/d05cb422-5e2e-4510-b85a-3ee529209672/20220207_173557/db-cluster-d05cb422.tar.gz                                                                                                                          |
| 20220207_173557 | loader-set  | https://cloudius-jenkins-test.s3.amazonaws.com/d05cb422-5e2e-4510-b85a-3ee529209672/20220207_173557/loader-set-d05cb422.tar.gz                                                                                                                          |
| 20220207_173557 | monitor-set | https://cloudius-jenkins-test.s3.amazonaws.com/d05cb422-5e2e-4510-b85a-3ee529209672/20220207_173557/monitor-set-d05cb422.tar.gz                                                                                                                         |
| 20220207_173557 | sct         | https://cloudius-jenkins-test.s3.amazonaws.com/d05cb422-5e2e-4510-b85a-3ee529209672/20220207_173557/sct-runner-d05cb422.tar.gz                                                                                                                          |
+-----------------+-------------+---------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+

@tgrabiec
Copy link
Contributor

tgrabiec commented Mar 1, 2022

@ShlomiBalalis What kind of disk was used in d05cb422-5e2e-4510-b85a-3ee529209672 ? Are nodes uniform, or is longevity-100gb-4h-4-6-db-node-d05cb422-5 special in some way?

Looks like schema merge was delayed for about a minute on longevity-100gb-4h-4-6-db-node-d05cb422-5:

Screenshot from 2022-03-01 11-33-14

Logs of node 5:

2022-02-07T12:56:57+00:00 longevity-100gb-4h-4-6-db-node-d05cb422-5 !    INFO |  [shard 11] schema_tables - Creating keyspace drop_table_during_repair_ks_9

^^^ RPC-initiated change appears on the node, 18 sec since it was sent. Unlikely to block in the network (other messages go through), not blocked on the lock (no other merge in progress).

Could be blocked on applying mutations and/or reading schema from disk after merge.

2022-02-07T12:57:25+00:00 longevity-100gb-4h-4-6-db-node-d05cb422-5 !    INFO |  [shard 11] schema_tables - Schema version changed to f6b65efa-d633-35d2-8d13-720d7b8ba801

^^^ Takes 28 seconds to process the schema change after schema was merged. This phase includes:
  - smp messages across all shards (shards underutilize cpu, so should not be a problem)
  - read of the schema to calculate the digest
  - write to system.local to save the digest


2022-02-07T12:57:26+00:00 longevity-100gb-4h-4-6-db-node-d05cb422-5 !    INFO |  [shard 15] schema_tables - Creating drop_table_during_repair_ks_9.standard1 id=639f7c90-8815-11ec-9eac-e80beec9914d version=cd12091f-15f4-3acd-a932-f8393fc56371
2022-02-07T12:57:27+00:00 longevity-100gb-4h-4-6-db-node-d05cb422-5 !    INFO |  [shard 15] schema_tables - Schema version changed to 4efd6d13-873b-379a-8d78-3d64c9bb0d07

RPC 2 completes (in 2 sec)

Looking at that shard's metrics (ignore the "commitlog" in the titles, the graphs show total for all io queues):

Screenshot from 2022-03-01 11-31-20

I/O queues experience high queuing time during schema merge. The queue (seastar side) grows large (mainly commitlog and query classes). The queue to the disk is low (0-1 per shard, not shown above, above you can see aggregate for all shards).

There is concurrent user read load ramping up during schema change.
There is a concurrent user write load during schema change (30k requests / coordinator).

CPU is underutilized. No task quota violations.

You can see that the total bandwidth is lower during schema change (1.4 GiB/s -> 1 GiB/s), and when it grows to its previous high the queues get drained and progress is made.

Here's one hypothesis. The disk runs close to its capacity during this event. During schema change (or right before it), something makes latency of requests higher than expected by the I/O scheduler. This effectively reduces disks bandwidth. This causes overload situation which slows down the schema merge. When bandwidth goes back to normal, the situation heals and schema change completes.

One suspect are the fsyncs issued during schema merge.

\cc @xemul

@ShlomiBalalis
Copy link

@ShlomiBalalis What kind of disk was used in d05cb422-5e2e-4510-b85a-3ee529209672 ? Are nodes uniform, or is longevity-100gb-4h-4-6-db-node-d05cb422-5 special in some way?

The nodes are uniform, I believe the data disks are the default nvme that is attached with i3.4xlarge (while the root device is a seperate gp2 we add)

@DoronArazii DoronArazii removed the high label Dec 28, 2022
@DoronArazii
Copy link

@mykaul are we planning to chase after this issue in 5.2?

@mykaul
Copy link
Contributor

mykaul commented Feb 7, 2023

@mykaul are we planning to chase after this issue in 5.2?

No, and eventually we'll need to see if it reproduces with Raft.

@mykaul mykaul modified the milestones: 5.2, 5.3 Feb 7, 2023
@ShlomiBalalis
Copy link

The issue was reproduced while testing 5.2 (using raft):

Issue description

In this run, the User Cassandra Stress thread (that uses this profile) was DOA because the cluster failed to achieve schema agreement:
The stress command:

cassandra-stress user profile=/tmp/cs_mv_profile.yaml ops'(insert=3,read1=1,read2=1,read3=1)' cl=QUORUM duration=6800m -mode cql3 native -rate threads=10
===== Using optimized driver!!! =====
Connected to cluster: longevity-tls-1tb-7d-5-2-db-cluster-915f76f1, max pending requests per connection null, max connections per host 8
Datatacenter: eu-west; Host: /10.4.3.67; Rack: 1a
Datatacenter: eu-west; Host: /10.4.1.243; Rack: 1a
Datatacenter: eu-west; Host: /10.4.0.91; Rack: 1a
Datatacenter: eu-west; Host: /10.4.0.108; Rack: 1a
WARN  22:13:27,414 Not using advanced port-based shard awareness with /10.4.0.108:9042 because we're missing port-based shard awareness port on the server
WARN  22:13:27,534 Not using advanced port-based shard awareness with /10.4.3.67:9042 because we're missing port-based shard awareness port on the server
WARN  22:13:27,636 Not using advanced port-based shard awareness with /10.4.0.91:9042 because we're missing port-based shard awareness port on the server
WARN  22:13:27,717 Not using advanced port-based shard awareness with /10.4.1.243:9042 because we're missing port-based shard awareness port on the server
WARN  22:13:39,748 No schema agreement from live replicas after 10 s. The schema may not be up to date on some nodes.
WARN  22:13:39,768 Unexpected error while querying /10.4.3.67:9042 - [com.datastax.driver.core.exceptions.InvalidQueryException: Keyspace 'mview' does not exist]. Find next host to query.
com.datastax.driver.core.exceptions.OperationTimedOutException: [/10.4.1.243:9042] Timed out waiting for server response
        at com.datastax.driver.core.exceptions.OperationTimedOutException.copy(OperationTimedOutException.java:43)
        at com.datastax.driver.core.exceptions.OperationTimedOutException.copy(OperationTimedOutException.java:25)
        at com.datastax.driver.core.DriverThrowables.propagateCause(DriverThrowables.java:35)
        at com.datastax.driver.core.DefaultResultSetFuture.getUninterruptibly(DefaultResultSetFuture.java:293)
        at com.datastax.driver.core.AbstractSession.execute(AbstractSession.java:58)
        at org.apache.cassandra.stress.util.JavaDriverClient.execute(JavaDriverClient.java:190)
        at org.apache.cassandra.stress.StressProfile.maybeCreateSchema(StressProfile.java:247)
        at org.apache.cassandra.stress.StressProfile.newGenerator(StressProfile.java:735)
        at org.apache.cassandra.stress.StressProfile.printSettings(StressProfile.java:131)
        at org.apache.cassandra.stress.settings.StressSettings.printSettings(StressSettings.java:383)
        at org.apache.cassandra.stress.Stress.run(Stress.java:95)
        at org.apache.cassandra.stress.Stress.main(Stress.java:62)
Caused by: com.datastax.driver.core.exceptions.OperationTimedOutException: [/10.4.1.243:9042] Timed out waiting for server response
        at com.datastax.driver.core.RequestHandler$SpeculativeExecution.onTimeout(RequestHandler.java:1026)
        at com.datastax.driver.core.Connection$ResponseHandler$1.run(Connection.java:1925)
        at com.datastax.shaded.netty.util.HashedWheelTimer$HashedWheelTimeout.run(HashedWheelTimer.java:715)
        at com.datastax.shaded.netty.util.concurrent.ImmediateExecutor.execute(ImmediateExecutor.java:34)
        at com.datastax.shaded.netty.util.HashedWheelTimer$HashedWheelTimeout.expire(HashedWheelTimer.java:703)
        at com.datastax.shaded.netty.util.HashedWheelTimer$HashedWheelBucket.expireTimeouts(HashedWheelTimer.java:790)
        at com.datastax.shaded.netty.util.HashedWheelTimer$Worker.run(HashedWheelTimer.java:503)
        at com.datastax.shaded.netty.util.concurrent.FastThreadLocalRunnable.run(FastThreadLocalRunnable.java:30)
        at java.base/java.lang.Thread.run(Thread.java:829)

Grepping the logs of the nodes, it seems that node 1's schema version was updated nearly instantly, but nodes 2 to 4 took over 10 seconds more to do the same:
Node 1:

2023-02-09T22:13:29+00:00 longevity-tls-1tb-7d-5-2-db-node-915f76f1-1     !INFO | scylla[5498]:  [shard  0] schema_tables - Creating keyspace mview
2023-02-09T22:13:29+00:00 longevity-tls-1tb-7d-5-2-db-node-915f76f1-1     !INFO | scylla[5498]:  [shard  0] schema_tables - Schema version changed to b4b735e4-03a3-38d0-ba8c-dff44cea5a83

Node 2:

2023-02-09T22:13:40+00:00 longevity-tls-1tb-7d-5-2-db-node-915f76f1-2     !INFO | scylla[5519]:  [shard  0] schema_tables - Creating keyspace mview
2023-02-09T22:13:40+00:00 longevity-tls-1tb-7d-5-2-db-node-915f76f1-2     !INFO | scylla[5519]:  [shard  0] schema_tables - Schema version changed to b4b735e4-03a3-38d0-ba8c-dff44cea5a83

Node 3:

2023-02-09T22:13:40+00:00 longevity-tls-1tb-7d-5-2-db-node-915f76f1-3     !INFO | scylla[5568]:  [shard  0] schema_tables - Creating keyspace mview
2023-02-09T22:13:40+00:00 longevity-tls-1tb-7d-5-2-db-node-915f76f1-3     !INFO | scylla[5568]:  [shard  0] schema_tables - Schema version changed to b4b735e4-03a3-38d0-ba8c-dff44cea5a83

Node 4:

2023-02-09T22:13:37+00:00 longevity-tls-1tb-7d-5-2-db-node-915f76f1-4     !INFO | scylla[5607]:  [shard  0] schema_tables - Creating keyspace mview
2023-02-09T22:13:37+00:00 longevity-tls-1tb-7d-5-2-db-node-915f76f1-4     !INFO | scylla[5607]:  [shard  0] schema_tables - Schema version changed to b4b735e4-03a3-38d0-ba8c-dff44cea5a83

The run does use raft (consistent_cluster_management set to true)

Impact

Major.

How frequently does it reproduce?

It's the only stress thread in this run that failed with this error.

Installation details

Kernel Version: 5.15.0-1028-aws
Scylla version (or git commit hash): 5.2.0~rc1-20230207.8ff4717fd010 with build-id 78fbb2c25e9244a62f57988313388a0260084528

Cluster size: 4 nodes (i3.4xlarge)

Scylla Nodes used in this run:

  • longevity-tls-1tb-7d-5-2-db-node-915f76f1-4 (3.249.249.53 | 10.4.1.243) (shards: 14)
  • longevity-tls-1tb-7d-5-2-db-node-915f76f1-3 (34.244.54.80 | 10.4.3.67) (shards: 14)
  • longevity-tls-1tb-7d-5-2-db-node-915f76f1-2 (34.245.56.45 | 10.4.0.108) (shards: 14)
  • longevity-tls-1tb-7d-5-2-db-node-915f76f1-1 (3.248.196.46 | 10.4.0.91) (shards: 14)

OS / Image: ami-0a6094bea26a69f97 (aws: eu-west-1)

Test: longevity-1tb-7days-test
Test id: 915f76f1-3254-46de-ba69-bd654a402c0d
Test name: scylla-5.2/longevity/longevity-1tb-7days-test
Test config file(s):

Logs and commands
  • Restore Monitor Stack command: $ hydra investigate show-monitor 915f76f1-3254-46de-ba69-bd654a402c0d
  • Restore monitor on AWS instance using Jenkins job
  • Show all stored logs command: $ hydra investigate show-logs 915f76f1-3254-46de-ba69-bd654a402c0d

Logs:

Jenkins job URL

@mykaul
Copy link
Contributor

mykaul commented Feb 19, 2023

@kostja - can your team take a look at the latest failure?

@roydahan
Copy link

To simplify the scenario, what happens here is as follow:

  1. Test loads 1TB of data into the cluster.
  2. Test start stressing the cluster with "mixed" workload.
  3. Test try to start c-s user profile that has the following set of commands (schema changes):
keyspace_definition: |

  CREATE KEYSPACE mview WITH replication = {'class': 'SimpleStrategy', 'replication_factor': 3};

table: users

table_definition: |

  CREATE TABLE users (
    username text,
    first_name text,
    last_name text,
    password text,
    email text,
    last_access timeuuid,
    PRIMARY KEY(username)
  );

extra_definitions:
  - CREATE MATERIALIZED VIEW mview.users_by_first_name AS SELECT * FROM mview.users WHERE first_name IS NOT NULL and username IS NOT NULL PRIMARY KEY (first_name, username);
  - CREATE MATERIALIZED VIEW mview.users_by_last_name AS SELECT * FROM mview.users WHERE last_name IS NOT NULL and username IS NOT NULL PRIMARY KEY (last_name, username);

columnspec:
  - name: username
    size: uniform(10..30)
  - name: first_name
    size: fixed(16)
  - name: last_name
    size: uniform(1..32)
  - name: password
    size: fixed(80) # sha-512
  - name: email
    size: uniform(16..50)

@kostja
Copy link
Contributor

kostja commented Feb 20, 2023

Raft tables write to the commitlog in stronlgy consistent mode, that is, they don't return until the commit log confirms the writes. If IO subsystem is overloaded we can quite realistically get more than 10 seconds to commit a write. On top of that, the current schema refresh code (the code path that loads the committed writes into memory and prints the referenced commit message) is not optimal - essentially it reloads entire schema on each schema change. @tgrabiec recently sent a pull request to speed this up. I'm looking at it with hope.
Another pending change is to switch raft log to use a dedicated schema commit log. This should short cut possible normal write traffic interference with Raft commits and make the scheduler job of quickly committing raft mutations easier.

So I'm afraid as much as we promised correctness with raft based topology changes we haven't delivered on speed, and the above two changes should improve on that.

Meanwhile, what we can do next to further isolate the problem, is to reproduce the failure with raft_group0 and raft logs enabled. That would allow us to correlate the time a schema change happens in Raft with the time it hits the commit log and the time it is then loaded into schema cache. There is a small chance that the fixes mentioned above are not sufficient to address the issue, that seeing the raft logs would help us identify that.
@gleb-cloudius @gusev-p jfyi

@kostja kostja assigned kostja and unassigned bhalevy Feb 20, 2023
@tgrabiec
Copy link
Contributor

I can see in the logs:

database - Not using schema commit log.

So the optimization to not flush the memtables is not active, which is probably the cause.

Are you overwriting the scylla.yaml?

By default it should have this:

force_schema_commit_log: true

@mykaul
Copy link
Contributor

mykaul commented Feb 21, 2023

@juliayakovlev - please see above - and in general, would be good to know, what if any, we change from default YAML.

@juliayakovlev
Copy link
Author

I can see in the logs:

database - Not using schema commit log.

So the optimization to not flush the memtables is not active, which is probably the cause.

Are you overwriting the scylla.yaml?

By default it should have this:

force_schema_commit_log: true

@tgrabiec
The force_schema_commit_log is not in the nodes' scylla.yaml. Does it mean that default value is used?

@tgrabiec
Copy link
Contributor

Yes, the default is used, but the default is to not force the schema commit log, in order to support rollback during rolling upgrade. If default is used, you need another round of restarts after full upgrade. For new installs we put the force_schema_commit_log: true in the yaml so that rolling restart is not needed.

@roydahan
Copy link

@tgrabiec I'm a bit confused with your last comment.
Do you mean that for fresh installs the default should be "force_schema_commit_log: true"?
Is it only in master or should be also for 5.2?

@tgrabiec
Copy link
Contributor

@roydahan Yes, the scylla.yaml in the repo has "force_schema_commit_log: true" since the feature was introduced. In general, settings in scylla.yaml should be preserved unless explicitly changed.

@roydahan
Copy link

roydahan commented Mar 8, 2023

@roydahan Yes, the scylla.yaml in the repo has "force_schema_commit_log: true" since the feature was introduced. In general, settings in scylla.yaml should be preserved unless explicitly changed.

Going back to this, I think found out what happened with this attribute (force_schema_commit_log).
In SCT, someone decided to guard and allow only known attributes in scylla.yaml.
For the general case, it should be ok, all defaults values usually match db/config.cc and the scylla.yaml, hence SCT needs to know about the new attribute only when we want to explicitly change it.

However, in this case, this attribute default value is different between db/config.cc (false) and scylla.yaml (true):
https://github.com/scylladb/scylladb/pull/10333/files.

Hence, the tests didn't change this attribute and also didn't include it in our test scylla.yaml.

@tgrabiec why do we have such a case that the default in the code is different than the yaml?

@tgrabiec
Copy link
Contributor

tgrabiec commented Mar 8, 2023

First, you should know two facts about the schema commit log feature. It's not possible to enable it on the node in a hot manner, while the node is running, you need a restart. Secondly, you can't safely downgrade the node if the feature is enabled. So the default is false because when upgrading, we want to enable this feature only after all nodes are upgraded, to allow downgrades during rolling upgrade. You need a round of restarts after the upgrade. We want to avoid the need for a round of restarts in fresh clusters, hence scylla.yaml forces the feature to be enabled (there is no need to support rollback, so we can do it).

@mykaul
Copy link
Contributor

mykaul commented Apr 17, 2023

@roydahan - can you provide an update on this?

@kostja kostja modified the milestones: 5.3, 5.x Apr 18, 2023
@kostja kostja added area/commitlog Issues related to the commit log. and removed area/schema changes P2 High Priority labels Apr 18, 2023
@roydahan
Copy link

Looks like using the force_schema_commit_log solves the issue.
I left it open to make sure we don't see it in all other runs of 5.2.
Looks safe to close now.

@roydahan roydahan modified the milestones: 5.x, 5.2 Apr 18, 2023
@roydahan
Copy link

Changed MS to 5.2 since the issue is resolved with this flag in 5.2.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
area/commitlog Issues related to the commit log. bug
Projects
None yet
Development

No branches or pull requests