I260415 19:16:41.875042 1 util/log/file_sync_buffer.go:238 ⋮ [config] file created at: 2026/04/15 19:16:41 I260415 19:16:41.875045 1 util/log/file_sync_buffer.go:238 ⋮ [config] running on machine: ‹w-01KP98GZE1Z52DMFW0VS7MAV8G› I260415 19:16:41.875047 1 util/log/file_sync_buffer.go:238 ⋮ [config] binary: CockroachDB OSS v22.1.22-64-g86fdbfca06 (x86_64-linux-gnu, built 2026/03/18 01:46:58, go1.22.11) I260415 19:16:41.875050 1 util/log/file_sync_buffer.go:238 ⋮ [config] arguments: [‹cockroach› ‹start-single-node› ‹--insecure› ‹--http-addr=:0› ‹--store=path=/var/tmp/omicron_tmp/.tmp2m9Oy9/data,ballast-size=0› ‹--listen-addr› ‹[::1]:0› ‹--listening-url-file› ‹/var/tmp/omicron_tmp/.tmp2m9Oy9/listen-url› ‹--max-sql-memory› ‹256MiB›] I260415 19:16:41.875055 1 util/log/file_sync_buffer.go:238 ⋮ [config] log format (utf8=✓): crdb-v2 I260415 19:16:41.875057 1 util/log/file_sync_buffer.go:238 ⋮ [config] line format: [IWEF]yymmdd hh:mm:ss.uuuuuu goid [chan@]file:line redactionmark \[tags\] [counter] msg W260415 19:16:41.874954 1 1@cli/start.go:1180 ⋮ [n?] 1 ALL SECURITY CONTROLS HAVE BEEN DISABLED! W260415 19:16:41.874954 1 1@cli/start.go:1180 ⋮ [n?] 1 + W260415 19:16:41.874954 1 1@cli/start.go:1180 ⋮ [n?] 1 +This mode is intended for non-production testing only. W260415 19:16:41.874954 1 1@cli/start.go:1180 ⋮ [n?] 1 + W260415 19:16:41.874954 1 1@cli/start.go:1180 ⋮ [n?] 1 +In this mode: W260415 19:16:41.874954 1 1@cli/start.go:1180 ⋮ [n?] 1 +- Your cluster is open to any client that can access ‹::1›. W260415 19:16:41.874954 1 1@cli/start.go:1180 ⋮ [n?] 1 +- Intruders with access to your machine or network can observe client-server traffic. W260415 19:16:41.874954 1 1@cli/start.go:1180 ⋮ [n?] 1 +- Intruders can log in without password and read or write any data in the cluster. W260415 19:16:41.874954 1 1@cli/start.go:1180 ⋮ [n?] 1 +- Intruders can consume all your server's resources and cause unavailability. I260415 19:16:41.875109 1 1@cli/start.go:1190 ⋮ [n?] 2 To start a secure server without mandating TLS for clients, I260415 19:16:41.875109 1 1@cli/start.go:1190 ⋮ [n?] 2 +consider --accept-sql-without-tls instead. For other options, see: I260415 19:16:41.875109 1 1@cli/start.go:1190 ⋮ [n?] 2 + I260415 19:16:41.875109 1 1@cli/start.go:1190 ⋮ [n?] 2 +- ‹https://go.crdb.dev/issue-v/53404/v22.1› I260415 19:16:41.875109 1 1@cli/start.go:1190 ⋮ [n?] 2 +- https://www.cockroachlabs.com/docs/v22.1/secure-a-cluster.html I260415 19:16:41.875507 1 server/status/recorder.go:620 ⋮ [n?] 3 ‹available memory from cgroups (8.0 EiB) is unsupported, using system memory 31 GiB instead:› W260415 19:16:41.875523 1 1@cli/start.go:1106 ⋮ [n?] 4 ‹Using the default setting for --cache (128 MiB).› W260415 19:16:41.875523 1 1@cli/start.go:1106 ⋮ [n?] 4 +‹ A significantly larger value is usually needed for good performance.› W260415 19:16:41.875523 1 1@cli/start.go:1106 ⋮ [n?] 4 +‹ If you have a dedicated server a reasonable setting is --cache=.25 (7.7 GiB).› I260415 19:16:41.875615 1 server/status/recorder.go:620 ⋮ [n?] 5 ‹available memory from cgroups (8.0 EiB) is unsupported, using system memory 31 GiB instead:› I260415 19:16:41.875620 1 1@cli/start.go:1219 ⋮ [n?] 6 ‹CockroachDB OSS v22.1.22-64-g86fdbfca06 (x86_64-linux-gnu, built 2026/03/18 01:46:58, go1.22.11)› I260415 19:16:41.877505 1 server/status/recorder.go:620 ⋮ [n?] 7 ‹available memory from cgroups (8.0 EiB) is unsupported, using system memory 31 GiB instead:› I260415 19:16:41.877521 1 server/config.go:487 ⋮ [n?] 8 system total memory: 31 GiB I260415 19:16:41.877534 1 server/config.go:489 ⋮ [n?] 9 server configuration: I260415 19:16:41.877534 1 server/config.go:489 ⋮ [n?] 9 +‹max offset 500000000› I260415 19:16:41.877534 1 server/config.go:489 ⋮ [n?] 9 +‹cache size 128 MiB› I260415 19:16:41.877534 1 server/config.go:489 ⋮ [n?] 9 +‹SQL memory pool size 256 MiB› I260415 19:16:41.877534 1 server/config.go:489 ⋮ [n?] 9 +‹scan interval 10m0s› I260415 19:16:41.877534 1 server/config.go:489 ⋮ [n?] 9 +‹scan min idle time 10ms› I260415 19:16:41.877534 1 server/config.go:489 ⋮ [n?] 9 +‹scan max idle time 1s› I260415 19:16:41.877534 1 server/config.go:489 ⋮ [n?] 9 +‹event log enabled true› I260415 19:16:41.877534 1 server/config.go:489 ⋮ [n?] 9 +‹span configs enabled true› I260415 19:16:41.877586 1 1@cli/start.go:1084 ⋮ [n?] 10 using local environment variables: I260415 19:16:41.877586 1 1@cli/start.go:1084 ⋮ [n?] 10 +GOTRACEBACK=crash I260415 19:16:41.877586 1 1@cli/start.go:1084 ⋮ [n?] 10 +LANG=‹en_US.UTF-8› I260415 19:16:41.877586 1 1@cli/start.go:1084 ⋮ [n?] 10 +LC_ALL=‹en_US.UTF-8› I260415 19:16:41.877586 1 1@cli/start.go:1084 ⋮ [n?] 10 +TZ=‹UTC› I260415 19:16:41.877594 1 1@cli/start.go:1091 ⋮ [n?] 11 process identity: ‹uid 12345 euid 12345 gid 12345 egid 12345› W260415 19:16:41.877715 1 1@cli/start.go:1304 ⋮ [n?] 12 could not initialize GEOS - spatial functions may not be available: geos: error during GEOS init: geos: cannot load GEOS from dir ‹"/usr/local/lib/cockroach"›: ‹geos error: /usr/local/lib/cockroach/libgeos.so: cannot open shared object file: No such file or directory› I260415 19:16:41.877748 1 1@cli/start.go:581 ⋮ [n?] 13 ‹starting cockroach node› I260415 19:16:42.045481 13 server/config.go:665 ⋮ [n?] 14 1 storage engine‹› initialized I260415 19:16:42.045516 13 server/config.go:668 ⋮ [n?] 15 Pebble cache size: 128 MiB I260415 19:16:42.045643 13 server/config.go:668 ⋮ [n?] 16 store 0: max size 0 B, max open file limit 519288 I260415 19:16:42.045650 13 server/config.go:668 ⋮ [n?] 17 store 0: {Encrypted:false ReadOnly:false FileStoreProperties:{path=‹/var/tmp/omicron_tmp/.tmp2m9Oy9/data›, fs=nsfs, blkdev=‹nsfs›, mnt=‹/run/snapd/ns/lxd.mnt› opts=‹rw›}} I260415 19:16:42.046513 13 server/tracedumper/tracedumper.go:120 ⋮ [n?] 18 writing job trace dumps to ‹/var/tmp/omicron_tmp/.tmp2m9Oy9/data/logs/inflight_trace_dump› I260415 19:16:42.090888 13 1@server/clock_monotonicity.go:65 ⋮ [n?] 19 ‹monitoring forward clock jumps based on server.clock.forward_jump_check_enabled› I260415 19:16:42.097110 13 1@server/clock_monotonicity.go:140 ⋮ [n1] 20 Sleeping till wall time 1776280602097107713 to catches up to 1776280602590880971 to ensure monotonicity. Delta: 493.773258ms I260415 19:16:42.097149 211 1@server/server.go:1294 ⋮ [n1] 21 connecting to gossip network to verify cluster ID ‹"34e5383c-5d4f-410d-9cb5-ecab8d778412"› I260415 19:16:42.591717 13 1@cli/start.go:521 ⋮ [n1] 22 listening URL file: ‹/var/tmp/omicron_tmp/.tmp2m9Oy9/listen-url› W260415 19:16:42.605318 13 1@gossip/gossip.go:1531 ⋮ [n1] 23 no addresses found; use --join to specify a connected node I260415 19:16:42.605364 13 gossip/gossip.go:401 ⋮ [n1] 24 NodeDescriptor set to ‹node_id:1 address: attrs:<> locality:<> ServerVersion: build_tag:"v22.1.22-64-g86fdbfca06" started_at:1776280602605359132 cluster_name:"" sql_address: http_address:› I260415 19:16:42.608111 13 server/node.go:470 ⋮ [n1] 25 initialized store s1 I260415 19:16:42.608138 13 kv/kvserver/stores.go:250 ⋮ [n1] 26 read 0 node addresses from persistent storage I260415 19:16:42.608295 211 1@server/server.go:1297 ⋮ [n1] 27 node connected via gossip I260415 19:16:42.611258 13 server/node.go:549 ⋮ [n1] 28 started with engine type ‹2› I260415 19:16:42.611313 13 server/node.go:551 ⋮ [n1] 29 started with attributes [] I260415 19:16:42.611345 13 server/goroutinedumper/goroutinedumper.go:122 ⋮ [n1] 30 writing goroutine dumps to ‹/var/tmp/omicron_tmp/.tmp2m9Oy9/data/logs/goroutine_dump› I260415 19:16:42.611359 13 server/heapprofiler/heapprofiler.go:49 ⋮ [n1] 31 writing go heap profiles to ‹/var/tmp/omicron_tmp/.tmp2m9Oy9/data/logs/heap_profiler› at least every 1h0m0s I260415 19:16:42.611368 13 server/heapprofiler/cgoprofiler.go:53 ⋮ [n1] 32 to enable jmalloc profiling: "export MALLOC_CONF=prof:true" or "ln -s prof:true /etc/malloc.conf" I260415 19:16:42.611375 13 server/heapprofiler/statsprofiler.go:54 ⋮ [n1] 33 writing memory stats to ‹/var/tmp/omicron_tmp/.tmp2m9Oy9/data/logs/heap_profiler› at last every 1h0m0s I260415 19:16:42.611511 13 server/heapprofiler/activequeryprofiler.go:85 ⋮ [n1] 34 writing go query profiles to ‹/var/tmp/omicron_tmp/.tmp2m9Oy9/data/logs/heap_profiler› I260415 19:16:42.611574 13 1@server/server.go:1423 ⋮ [n1] 35 starting http server at ‹[::1]:32963› (use: ‹[::1]:32963›) I260415 19:16:42.611587 13 1@server/server.go:1430 ⋮ [n1] 36 starting grpc/postgres server at ‹[::1]:37839› I260415 19:16:42.611596 13 1@server/server.go:1431 ⋮ [n1] 37 advertising CockroachDB node at ‹[::1]:37839› I260415 19:16:42.711481 344 kv/kvclient/rangefeed/rangefeedcache/watcher.go:316 ⋮ [n1] 38 spanconfig-subscriber: established range feed cache I260415 19:16:42.721156 13 1@util/log/event_log.go:32 ⋮ [n1] 39 ={"Timestamp":1776280602721153668,"EventType":"node_restart","NodeID":1,"StartedAt":1776280602605359132,"LastUp":1776280545418577331} I260415 19:16:42.721189 13 sql/sqlliveness/slinstance/slinstance.go:313 ⋮ [n1] 40 starting SQL liveness instance I260415 19:16:42.754398 582 sql/sqlstats/persistedsqlstats/provider.go:132 ⋮ [n1] 41 starting sql-stats-worker with initial delay: 9m9.329670084s I260415 19:16:42.754474 580 sql/temporary_schema.go:514 ⋮ [n1] 42 running temporary object cleanup background job I260415 19:16:42.754516 595 kv/kvclient/rangefeed/rangefeedcache/watcher.go:316 ⋮ [n1] 43 settings-watcher: established range feed cache I260415 19:16:42.755021 580 sql/temporary_schema.go:559 ⋮ [n1] 44 found 0 temporary schemas I260415 19:16:42.755034 580 sql/temporary_schema.go:562 ⋮ [n1] 45 early exiting temporary schema cleaner as no temporary schemas were found I260415 19:16:42.755041 580 sql/temporary_schema.go:563 ⋮ [n1] 46 completed temporary object cleanup job I260415 19:16:42.755046 580 sql/temporary_schema.go:646 ⋮ [n1] 47 temporary object cleaner next scheduled to run at 2026-04-15 19:46:42.75445246 +0000 UTC I260415 19:16:42.755144 13 server/server_sql.go:1287 ⋮ [n1] 48 done ensuring all necessary startup migrations have run I260415 19:16:42.755270 539 kv/kvclient/rangefeed/rangefeedcache/watcher.go:316 ⋮ [n1] 49 tenant-settings-watcher: established range feed cache I260415 19:16:42.755453 533 kv/kvclient/rangefeed/rangefeedcache/watcher.go:316 ⋮ [n1] 50 system-config-cache: established range feed cache I260415 19:16:42.761330 535 jobs/job_scheduler.go:433 ⋮ [n1] 51 waiting 4m0s before scheduled jobs daemon start I260415 19:16:42.761701 615 kv/kvserver/replica_rangefeed.go:700 ⋮ [n1,rangefeed=‹lease›,s1,r6] 52 RangeFeed closed timestamp 1776280541.581687767,0 is behind by 1m1.180011773s I260415 19:16:42.765348 538 server/auto_upgrade.go:79 ⋮ [n1] 53 ‹no need to upgrade, cluster already at the newest version› I260415 19:16:42.766543 348 sql/sqlliveness/slstorage/slstorage.go:460 ⋮ [n1] 54 inserted sqlliveness session c4352498b0f54a588af11fef651bc9e3 I260415 19:16:42.766561 348 sql/sqlliveness/slinstance/slinstance.go:198 ⋮ [n1] 55 created new SQL liveness session c4352498b0f54a588af11fef651bc9e3 I260415 19:16:42.791339 738 sql/sqlliveness/slstorage/slstorage.go:342 ⋮ [n1,intExec=‹expire-sessions›] 56 deleted session f8f06c62e6b44f52b8b21f340a2611bc which expired at 1776280583.282343430,0 I260415 19:16:42.811794 13 1@server/server_sql.go:1350 ⋮ [n1] 57 ‹serving sql connections› I260415 19:16:42.811940 13 1@cli/start.go:1027 ⋮ [n1] 58 node startup completed: I260415 19:16:42.811940 13 1@cli/start.go:1027 ⋮ [n1] 58 +CockroachDB node starting at 2026-04-15 19:16:42.811845719 +0000 UTC (took 0.9s) I260415 19:16:42.811940 13 1@cli/start.go:1027 ⋮ [n1] 58 +build: OSS v22.1.22-64-g86fdbfca06 @ 2026/03/18 01:46:58 (go1.22.11) I260415 19:16:42.811940 13 1@cli/start.go:1027 ⋮ [n1] 58 +webui: ‹http://[::1]:32963› I260415 19:16:42.811940 13 1@cli/start.go:1027 ⋮ [n1] 58 +sql: ‹postgresql://root@[::1]:37839/defaultdb?sslmode=disable› I260415 19:16:42.811940 13 1@cli/start.go:1027 ⋮ [n1] 58 +sql (JDBC): ‹jdbc:postgresql://[::1]:37839/defaultdb?sslmode=disable&user=root› I260415 19:16:42.811940 13 1@cli/start.go:1027 ⋮ [n1] 58 +RPC client flags: ‹cockroach --host=[::1]:37839 --insecure› I260415 19:16:42.811940 13 1@cli/start.go:1027 ⋮ [n1] 58 +logs: ‹/var/tmp/omicron_tmp/.tmp2m9Oy9/data/logs› I260415 19:16:42.811940 13 1@cli/start.go:1027 ⋮ [n1] 58 +temp dir: ‹/var/tmp/omicron_tmp/.tmp2m9Oy9/data/cockroach-temp3530334282› I260415 19:16:42.811940 13 1@cli/start.go:1027 ⋮ [n1] 58 +external I/O path: ‹/var/tmp/omicron_tmp/.tmp2m9Oy9/data/extern› I260415 19:16:42.811940 13 1@cli/start.go:1027 ⋮ [n1] 58 +store[0]: ‹path=/var/tmp/omicron_tmp/.tmp2m9Oy9/data› I260415 19:16:42.811940 13 1@cli/start.go:1027 ⋮ [n1] 58 +storage engine: pebble I260415 19:16:42.811940 13 1@cli/start.go:1027 ⋮ [n1] 58 +clusterID: ‹34e5383c-5d4f-410d-9cb5-ecab8d778412› I260415 19:16:42.811940 13 1@cli/start.go:1027 ⋮ [n1] 58 +status: restarted pre-existing node I260415 19:16:42.811940 13 1@cli/start.go:1027 ⋮ [n1] 58 +nodeID: 1 I260415 19:16:42.822901 735 util/log/event_log.go:32 ⋮ [n1,client=[::1]:59298,user=root] 59 ={"Timestamp":1776280602821425946,"EventType":"set_cluster_setting","Statement":"SET CLUSTER SETTING \"kv.raft_log.disable_synchronization_unsafe\" = true","Tag":"SET CLUSTER SETTING","User":"root","SettingName":"kv.raft_log.disable_synchronization_unsafe","Value":"‹true›"} I260415 19:16:44.246388 1 1@cli/start.go:749 ⋮ [n1] 60 received signal 'terminated' I260415 19:16:44.247721 1 1@cli/start.go:846 ⋮ [n1] 61 ‹initiating graceful shutdown of server› I260415 19:16:44.247764 1343 1@server/drain.go:342 ⋮ [n1,server drain process] 62 ‹waiting for health probes to notice that the node is not ready for new sql connections› I260415 19:16:44.247775 1343 1@sql/pgwire/server.go:612 ⋮ [n1,server drain process] 63 ‹starting draining SQL connections› I260415 19:16:44.856530 157 1@gossip/gossip.go:1547 ⋮ [n1] 64 node has connected to cluster via gossip I260415 19:16:44.861829 157 kv/kvserver/stores.go:269 ⋮ [n1] 65 wrote 0 node addresses to persistent storage I260415 19:16:45.885363 826 1@sql/pgwire/conn.go:601 ⋮ [n1,client=[::1]:59304,user=root] 66 ‹closing existing connection while server is draining› I260415 19:16:46.084532 828 1@sql/pgwire/conn.go:601 ⋮ [n1,client=[::1]:59306,user=root] 67 ‹closing existing connection while server is draining› I260415 19:16:46.084532 810 1@sql/pgwire/conn.go:601 ⋮ [n1,client=[::1]:59308,user=root] 68 ‹closing existing connection while server is draining› I260415 19:16:46.084646 815 1@sql/pgwire/conn.go:601 ⋮ [n1,client=[::1]:59316,user=root] 69 ‹closing existing connection while server is draining› I260415 19:16:46.084672 462 1@sql/pgwire/conn.go:601 ⋮ [n1,client=[::1]:59310,user=root] 70 ‹closing existing connection while server is draining› I260415 19:16:46.084688 817 1@sql/pgwire/conn.go:601 ⋮ [n1,client=[::1]:59312,user=root] 71 ‹closing existing connection while server is draining› I260415 19:16:46.084696 812 1@sql/pgwire/conn.go:601 ⋮ [n1,client=[::1]:59314,user=root] 72 ‹closing existing connection while server is draining› I260415 19:16:46.234007 806 1@sql/pgwire/conn.go:601 ⋮ [n1,client=[::1]:59300,user=root] 73 ‹closing existing connection while server is draining› I260415 19:16:46.235125 823 1@sql/pgwire/conn.go:601 ⋮ [n1,client=[::1]:59302,user=root] 74 ‹closing existing connection while server is draining› I260415 19:16:46.235465 1343 sql/sqlstats/persistedsqlstats/flush.go:68 ⋮ [n1,server drain process] 75 flushing 108 stmt/txn fingerprints (194560 bytes) after 1.07µs I260415 19:16:46.497365 1343 1@server/drain.go:285 ⋮ [n1,server drain process] 76 drain remaining: 139 I260415 19:16:46.498227 1343 1@server/drain.go:287 ⋮ [n1,server drain process] 77 drain details: SQL clients: 9, descriptor leases: 129, liveness record: 1 I260415 19:16:46.699086 1343 sql/sqlstats/persistedsqlstats/flush.go:68 ⋮ [n1,server drain process] 78 flushing 2 stmt/txn fingerprints (10240 bytes) after 1.31µs W260415 19:16:46.700050 1343 sql/sqlstats/persistedsqlstats/flush.go:170 ⋮ [n1,server drain process] 79 ‹failed to flush statement statistics›: flush statement ‹4025915790256006928›'s statistics: ‹insert-stmt-stats›: cannot acquire lease when draining W260415 19:16:46.701666 1343 sql/sqlstats/persistedsqlstats/flush.go:170 ⋮ [n1,server drain process] 80 ‹failed to flush transaction statistics›: flushing transaction ‹11006041404254834895›'s statistics: ‹insert-txn-stats›: cannot acquire lease when draining I260415 19:16:46.706308 1343 1@server/drain.go:285 ⋮ [n1,server drain process] 81 drain remaining: 0 W260415 19:16:46.706610 2013 kv/kvserver/intentresolver/intent_resolver.go:793 ⋮ [-] 82 failed to gc transaction record: could not GC completed transaction anchored at ‹/Table/43/1/4/2026-04-15T19:00:00Z/"\x86(\xea\a\x9f\xe20\x16"/"$ internal-fetch-stmt-count"/1›: ‹node unavailable; try another peer› W260415 19:16:46.706704 348 sql/sqlliveness/slinstance/slinstance.go:235 ⋮ [n1] 83 ‹exiting heartbeat loop› W260415 19:16:46.706755 591 jobs/registry.go:815 ⋮ [n1] 84 canceling all adopted jobs due to stopper quiescing I260415 19:16:46.706777 586 sql/stats/automatic_stats.go:489 ⋮ [n1] 85 ‹quiescing auto stats refresher› I260415 19:16:46.716508 1 1@cli/start.go:898 ⋮ [n1] 86 server drained and shutdown completed